Education, Science, Technology, Innovation and Life
Open Access
Sign In

U-Net Handwriting Removal Method and Dataset with ResNetV2 Fusion

Download as PDF

DOI: 10.23977/jipta.2026.090102 | Downloads: 5 | Views: 48

Author(s)

Runqing Yan 1, Jianye An 1

Affiliation(s)

1 Tianjin University of Commerce, Tianjin, China

Corresponding Author

Jianye An

ABSTRACT

With the growing demand for paperless offices and digital archiving, many paper documents are scanned into image formats for management and dissemination. These images often contain handwritten annotations, signatures, or markings, which interfere with accurate understanding and automatic analysis, especially in educational scenarios where exams and assignments include extensive handwritten content. This highlights the need for effective handwritten text removal techniques. This work proposes an end-to-end handwritten text removal method based on a U-Net enhanced with ResNetV2 modules. The model leverages multi-scale feature extraction, residual learning, and skip connections to remove handwritten marks while preserving printed text and document layout. In addition, a high-quality, large-scale handwritten text removal dataset is constructed and publicly released to provide a standardized benchmark for evaluation and reproducibility. Experimental results show that the proposed approach efficiently removes handwritten traces while maintaining document structure and visual consistency, improving the usability of digital documents. The study contributes to research on handwritten text removal and provides technical support for educational resource digitization, smart learning, and document management.

KEYWORDS

ResNetV2 Fusion, U-Net, Handwriting Erase, Removal of Handwritten Text

CITE THIS PAPER

Runqing Yan, Jianye An. U-Net Handwriting Removal Method and Dataset with ResNetV2 Fusion. Journal of Image Processing Theory and Applications (2026) Vol. 9, No.1, 14-21. DOI: http://dx.doi.org/10.23977/jipta.2026.090102.

REFERENCES

[1] Nakamura T, Zhu A, Yanai K, Uchida S. Scene Text Eraser[C]. 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), Kyoto, Japan, 2017: 832-837. doi: 10.1109/ICDAR.2017.141.
[2] Isola P, Zhu JY, Zhou T, Efros AA. Image-to-image translation with conditional adversarial networks[C]. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
[3] Zhang S, Liu Y, Jin L, Huang Y, Lai S. EnsNet: Ensconce Text in the Wild[C]. Proceedings of the AAAI Conference on Artificial Intelligence, 2019; 33(01): 801-808. doi: 10.1609/aaai.v33i01.3301801.
[4] He K, Zhang X, Ren S, Sun J. Deep Residual Learning for Image Recognition[C]. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 2016: 770-778. doi: 10.1109/CVPR.2016.90.
[5] Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition[Preprint]. arXiv:1409.1556, 2014.
[6] Huang L, Fink GA, Jain R, Kise K, Zanibbi R. EnsExam: A Dataset for Handwritten Text Erasure on Examination Papers[C]. In: Fink GA, Jain R, Kise K, Zanibbi R, eds. Document Analysis and Recognition - ICDAR 2023. Lecture Notes in Computer Science, vol 14189. Cham: Springer; 2023. doi: 10.1007/978-3-031-41682-8_29.
[7] He K, Zhang X, Ren S, Sun J. Identity Mappings in Deep Residual Networks[C]. In: Leibe B, Matas J, Sebe N, Welling M, eds. Computer Vision – ECCV 2016. Lecture Notes in Computer Science, vol 9908. Cham: Springer; 2016. doi: 10.1007/978-3-319-46493-0_38.
[8] Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation[J]. Cham: Springer; 2015. doi: 10.1007/978-3-662-54345-0_3.

Downloads: 3026
Visits: 236746

Sponsors, Associates, and Links


All published work is licensed under a Creative Commons Attribution 4.0 International License.

Copyright © 2016 - 2031 Clausius Scientific Press Inc. All Rights Reserved.