arXiv Analytics

Sign in

arXiv:2204.05535 [cs.CV]AbstractReferencesReviewsResources

Open-set Text Recognition via Character-Context Decoupling

Chang Liu, Chun Yang, Xu-Cheng Yin

Published 2022-04-12Version 1

The open-set text recognition task is an emerging challenge that requires an extra capability to cognize novel characters during evaluation. We argue that a major cause of the limited performance for current methods is the confounding effect of contextual information over the visual information of individual characters. Under open-set scenarios, the intractable bias in contextual information can be passed down to visual information, consequently impairing the classification performance. In this paper, a Character-Context Decoupling framework is proposed to alleviate this problem by separating contextual information and character-visual information. Contextual information can be decomposed into temporal information and linguistic information. Here, temporal information that models character order and word length is isolated with a detached temporal attention module. Linguistic information that models n-gram and other linguistic statistics is separated with a decoupled context anchor mechanism. A variety of quantitative and qualitative experiments show that our method achieves promising performance on open-set, zero-shot, and close-set text recognition datasets.

Related articles: Most relevant | Search more
arXiv:1110.2053 [cs.CV] (Published 2011-10-10, updated 2017-12-27)
Steps Towards a Theory of Visual Information: Active Perception, Signal-to-Symbol Conversion and the Interplay Between Sensing and Control
arXiv:1511.04510 [cs.CV] (Published 2015-11-14)
Semantic Object Parsing with Local-Global Long Short-Term Memory
arXiv:2203.14806 [cs.CV] (Published 2022-03-28, updated 2023-09-06)
Extraction of Visual Information to Predict Crowdfunding Success