Research on Text Recognition of Ancient Books Based on Improved Lightweight CRNN: Integrating MobileNetV3 and Hybrid Attention Enhanced Architecture
Abstract
sequence recognition, extracting visual features through convolutional layers, modeling sequence dependencies with recurrent layers, and
finally outputting results via transcription layers. However, for the specific task of ancient text recognition, existing models face multiple chal
lenges: ancient book images often suffer from severe paper degradation, blurred handwriting, and complex background interference; the com
plex structure of Chinese characters and the plethora of variant forms require a high level of precision in feature extraction; furthermore, the
massive demand for the digitization of ancient texts necessitates recognition systems with high inference speed to accommodate large-scale
deployment. To balance recognition accuracy and computational efficiency, this study proposes an improved lightweight CRNN architecture.
This architecture uses MobileNet V3 as the backbone network, enhances spatial feature acquisition through the integration of the Coordinate
Attention (CA) mechanism, and optimizes global relationship modeling with a "sandbox" structure composed of Bidirectional Gated Recur
rent Units (BiGRU) and self-attention encoders, achieving efficient and precise recognition of ancient texts.
Keywords
Full Text:
PDFReferences
[1] Z. Tian, W. Huang, T. He, et al. Detecting text in natural image with connectionist text proposal network[C], Proceedings of the Euro
pean Conference on Computer Vision. Springer, 2016: 56-72.
[2] Shi B, Bai X, Yao C. An end-to-end trainable neural network for image-based sequence recognition and its application to scene text
recognition[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 39(11): 2298-2304.
[3] Hou Q, Zhou D, Feng J. Coordinate attention for efficient mobile network design[C]//Proceedings of the IEEE/CVF Conference on
Computer Vision and Pattern Recognition. 2021: 13713-13722.
DOI: http://dx.doi.org/10.70711/aitr.v3i9.9023
Refbacks
- There are currently no refbacks.