Ajou University repository

Handwritten Text Segmentation via End-to-End Learning of Convolutional Neural Networks
Citations

SCOPUS

23

Citation Export

DC Field Value Language
dc.contributor.authorJo, Junho-
dc.contributor.authorKoo, Hyung Il-
dc.contributor.authorSoh, Jae Woong-
dc.contributor.authorCho, Nam Ik-
dc.date.issued2020-11-01-
dc.identifier.urihttps://dspace.ajou.ac.kr/dev/handle/2018.oak/31493-
dc.description.abstractWe present a method that separates handwritten and machine-printed components that are mixed and overlapped in documents. Many conventional methods addressed this problem by extracting connected components (CCs) and classifying the extracted CCs into two classes. They were based on the assumption that two types of components are not overlapping each other, while we are focusing on more challenging and realistic cases where the components are often overlapping each other. For this, we propose a new method that performs pixel-level classification with a convolutional neural network. Unlike conventional neural network methods, our method works in an end-to-end manner and does not require any preprocessing steps (e.g., foreground extraction, handcrafted feature extraction, and so on). For the training of our network, we develop a cross-entropy based loss function to alleviate the class imbalance problem. Regarding the training dataset, although there are some datasets of mixed printed characters and handwritten scripts, most of them do not have overlapping cases and do not provide pixel-level annotations. Hence, we also propose a data synthesis method that generates realistic pixel-level training samples having many overlappings of printed and handwritten components. Experimental results on synthetic and real images have shown the effectiveness of the proposed method. Although the proposed network has been trained only with synthetic images, it also improves the OCR rate of real documents. Specifically, the OCR rate for machine-printed texts is increased from 0.8087 to 0.9442 by removing the overlapped handwritten scribbles by our method.-
dc.description.sponsorshipThis work was supported in part by Institute for Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No.NI190004,Development of AI based Robot Technologies for Understanding Assembly Instruction and Automatic Assembly Task Planning), and in part by Hancom Inc.-
dc.language.isoeng-
dc.publisherSpringer-
dc.subject.meshClass imbalance problems-
dc.subject.meshConnected component-
dc.subject.meshConventional methods-
dc.subject.meshForeground extraction-
dc.subject.meshHandwritten components-
dc.subject.meshHandwritten texts-
dc.subject.meshNeural network method-
dc.subject.meshPre-processing step-
dc.titleHandwritten Text Segmentation via End-to-End Learning of Convolutional Neural Networks-
dc.typeArticle-
dc.citation.endPage32150-
dc.citation.startPage32137-
dc.citation.titleMultimedia Tools and Applications-
dc.citation.volume79-
dc.identifier.bibliographicCitationMultimedia Tools and Applications, Vol.79, pp.32137-32150-
dc.identifier.doi10.1007/s11042-020-09624-9-
dc.identifier.scopusid2-s2.0-85089864406-
dc.identifier.urlhttps://link.springer.com/journal/11042-
dc.subject.keywordClass imbalance problem-
dc.subject.keywordData synthesis-
dc.subject.keywordHandwritten text segmentation-
dc.subject.keywordOptical character recognition-
dc.subject.keywordText separation-
dc.description.isoafalse-
dc.subject.subareaSoftware-
dc.subject.subareaMedia Technology-
dc.subject.subareaHardware and Architecture-
dc.subject.subareaComputer Networks and Communications-
Show simple item record

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

 KOO, HYUNG IL Image
KOO, HYUNG IL구형일
Department of Electrical and Computer Engineering
Read More

Total Views & Downloads

File Download

  • There are no files associated with this item.