Residential College | false |
Status | 已發表Published |
NRSTRNet: A Novel Network for Noise-Robust Scene Text Recognition | |
Yue, Hongwei1,2![]() ![]() ![]() ![]() | |
2023-01-23 | |
Source Publication | International Journal of Computational Intelligence Systems
![]() |
ISSN | 1875-6891 |
Volume | 16Issue:1Pages:5 |
Abstract | Scene text recognition (STR) has been widely applied in industrial and commercial fields. However, existing methods still face challenges when processing text images with defects such as low contrast, blur, low resolution, and insufficient illumination. These defects are common in actual situations because of diverse text backgrounds in natural scenes and limitations in shooting conditions. To address these challenges, we propose a novel network for noise-robust scene text recognition (NRSTRNet), which comprehensively suppresses the noise in the three critical steps of STR. Specifically, in the text feature extraction stage, NRSTRNet enhances the text-related features through the channel and spatial dimensions and disregards some disturbances from the non-text area, reducing the noise and redundancy in the input image. In the context encoding stage, fine-grained feature coding is proposed to effectively reduce the influence of previous noisy temporal features on current temporal features while simultaneously reducing the impact of partial noise on the overall encoding by sharing contextual feature encoding parameters. In the decoding stage, a self-attention module is added to enhance the connections between different temporal features, thereby leveraging the global information to obtain noise-resistant features. Through these approaches, NRSTRNet can enhance the local semantic information while considering the global semantic information. Experimental results show that the proposed NRSTRNet can improve the ability to characterize text images, enhance stability under the influence of noise, and achieve superior accuracy in text recognition. As a result, our model outperforms SOTA STR models on irregular text recognition benchmarks by 2% on average, and it is exceptionally robust when applied to noisy images. |
Keyword | Convolutional Block Attention Module Fine-grained Feature Scene Text Recognition Self-attention Mechanism |
DOI | 10.1007/s44196-023-00181-1 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science |
WOS Subject | Computer Science, Artificial Intelligence ; Computer Science, Interdisciplinary Applications |
WOS ID | WOS:000917396900001 |
Publisher | SPRINGERNATURE, CAMPUS, 4 CRINAN ST, LONDON N1 9XW, ENGLAND |
Scopus ID | 2-s2.0-85146755796 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | Faculty of Science and Technology DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Yue, Hongwei; Huang, Yufeng; Chen, Chuangquan |
Affiliation | 1.School of Physics and Information Engineering, Guangdong University of Education, Guangzhou, 510303, China 2.Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen, 529020, China 3.Department of Computer and Information Science, University of Macau, Macao 4.School of Common Courses, Guangzhou Panyu Polytechnic, Guangzhou, 511483, China |
Recommended Citation GB/T 7714 | Yue, Hongwei,Huang, Yufeng,Vong, Chi Man,et al. NRSTRNet: A Novel Network for Noise-Robust Scene Text Recognition[J]. International Journal of Computational Intelligence Systems, 2023, 16(1), 5. |
APA | Yue, Hongwei., Huang, Yufeng., Vong, Chi Man., Jin, Yingying., Zeng, Zhiqiang., Yu, Mingqi., & Chen, Chuangquan (2023). NRSTRNet: A Novel Network for Noise-Robust Scene Text Recognition. International Journal of Computational Intelligence Systems, 16(1), 5. |
MLA | Yue, Hongwei,et al."NRSTRNet: A Novel Network for Noise-Robust Scene Text Recognition".International Journal of Computational Intelligence Systems 16.1(2023):5. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment