[table]capposition=top
International Institute of Information Technology, Hyderabad - 500032, INDIA
%F. ̵Author ̵et ̵al.˝%First ̵names ̵are ̵abbreviated ̵in ̵the ̵running ̵head.%If ̵there ̵are ̵more ̵than ̵two ̵authors, ̵’et ̵al.’ ̵is ̵used.https://github.com/firesans/STRforIndicLanguages
11email: {sanjana.gunna,rohit.saluja}@research.iiit.ac.in, 11email: jawahar@iiit.ac.in
Transfer Learning for Scene Text Recognition in Indian Languages
Abstract
Scene text recognition in low-resource Indian languages is challenging because of complexities like multiple scripts, fonts, text size, and orientations. In this work, we investigate the power of transfer learning for all the layers of deep scene text recognition networks from English to two common Indian languages. We perform experiments on the conventional CRNN model and STAR-Net to ensure generalisability. To study the effect of change in different scripts, we initially run our experiments on synthetic word images rendered using Unicode fonts. We show that the transfer of English models to simple synthetic datasets of Indian languages is not practical. Instead, we propose to apply transfer learning techniques among Indian languages due to similarity in their n-gram distributions and visual features like the vowels and conjunct characters. We then study the transfer learning among six Indian languages with varying complexities in fonts and word length statistics. We also demonstrate that the learned features of the models transferred from other Indian languages are visually closer (and sometimes even better) to the individual model features than those transferred from English. We finally set new benchmarks for scene-text recognition on Hindi, Telugu, and Malayalam datasets from IIIT-ILST and Bangla dataset from MLT-17 by achieving 6%, 5%, 2%, and 23% gains in Word Recognition Rates (WRRs) compared to previous works. We further improve the MLT-17 Bangla results by plugging in a novel correction BiLSTM into our model. We additionally release a dataset of around 440 scene images containing 500 Gujarati and 2535 Tamil words. WRRs improve over the baselines by 8%, 4%, 5%, and 3% on the MLT-19 Hindi and Bangla datasets and the Gujarati and Tamil datasets.
Keywords:
Scene text recognition transfer learning photo OCR multi-lingual OCR Indian Languages Indic OCR Synthetic Data.





1 Introduction
Scene-text recognition or Photo-Optical Character Recognition (Photo-OCR) aims to read scene-text in natural images. It is an essential step for a wide variety of computer vision tasks and has enjoyed significant success in several commercial applications [9]. Photo-OCR has diverse applications like helping the visually impaired, data mining of street-view-like images for information used in map services, and geographic information systems [2]. Scene-text recognition conventionally involves two steps; i) Text detection and ii) Text recognition. Text detection typically consists of detecting bounding boxes of word images [4]. The text recognition stage involves reading cropped text images obtained from the text detection stage or from the bounding box annotations [13]. In this work, we focus on the task of text recognition.
The multi-lingual text in scenes is a crucial part of human communication and globalization. Despite the popularity of recognition algorithms, non-Latin language advancements have been slow. Reading scene-text in such low resource languages is a challenging research problem as it is generally unstructured and appears in diverse conditions such as scripts, fonts, sizes, and orientations. Hence a large amount of dataset is usually required to train the scene-text recognition models. Conventionally, the synthetic dataset is used to deal with the problem since a large number of fonts are available in such low resource languages [13]. The synthetic data may also serve as an exciting asset to perform controlled experiments, e.g., to study the effect of transfer learning with the change in script or language text. We investigate such effects for transfer from English to two Indian languages in this work, i.e., Hindi and Gujarati. We also explore the transferability of features among six different Indian languages. We share scene text word images obtained from over scenes in Gujarati and Tamil to demonstrate such effects. In Fig. 1, we illustrate the sample annotated images from our datasets, and IIIT-ILST and MLT datasets, and the predictions of our models. The overall methodology we follow is that we first generate the synthetic datasets in the six Indian languages. We describe the dataset generation process and motivate the work in Section 2. We then train the two deep neural networks we introduce in Section 3 on the individual language datasets. Subsequently, we apply transfer-learning on all the layers of different networks from one language to another. Finally, as discussed in Section 4, we fine-tune the networks on standard datasets and examine their performance on real scene-text images in Section 5. We finally conclude the work in Section 6. The summary of our contributions are as follows:
-
1.
We investigate the transfer learning of complete scene-text recognition models i) from English to two Indian languages and ii) among the six Indian languages, i.e., Gujarati, Hindi, Bangla, Telugu, Tamil, and Malayalam.
-
2.
We also contribute two datasets of around word images in Gujarati and word images in Tamil from a total of Indian scenes.
-
3.
We achieve gains of , , and in Word Recognition Rates (WRRs) on IIIT-ILST Hindi, Telugu, and Malayalam datasets in comparison to previous works [13, 20]. On the MLT-19 Hindi and Bangla datasets and our Gujarati and Tamil datasets, we observe the WRR gains of , , , and , respectively, over our baseline models.
-
4.
For the MLT-17 Bangla dataset, we show a striking improvement of in Character Recognition Rate (CRR) and in WRR compared to Bušta et al. [2], by applying transfer-learning from another Indian language and plugging in a novel correction RNN layer into our model.
1.1 Related Work
We now discuss datasets and associated works in the field of photo-OCR.
Works of Photo-OCR on Latin Datasets: As stated earlier, the process of Photo-OCR conventionally includes two steps: i) Text detection and ii) Text recognition. With the success of Convolutional Neural Networks (CNN) for object detection, the works have been extended to text detection, treating words or lines as the objects [28, 38, 12]. Liao et al. [10] extend such works to real-time detection in scene images. Karatzas et al. [8] and Bušta et al. [1] present more efficient and accurate methods for text detection. Towards reading scene-text, Wang et al. [31] propose an object recognition pipeline based on a ground truth lexicon. It achieves competitive performance without the need for an explicit text detection step. Shi et al. [21] propose a Convolutional Recurrent Neural Network (CRNN) architecture, which integrates feature extraction, sequence modeling, and transcription into a unified framework. The model achieves remarkable performances in both lexicon-free and lexicon-based scene-text recognition tasks. Liu et al. [11] introduce Spatial Attention Residue Network (STAR-Net) with spatial transformer-based attention mechanism to remove image distortions, residue convolutional blocks for feature extraction, and an RNN block for decoding the text. Shi et al. [23] propose a segmentation-free Attention-based method for Text Recognition (ASTER) by adopting Thin-Plate-Spline (TPS) as a rectification unit. It tackles complex distortions and reduces the difficulty of irregular text recognition. The model incorporates ResNet to improve the network’s feature representation module and employs an attention-based mechanism combined with a Recurrent Neural Network (RNN) to form the prediction module. Uber-Text is a large-scale Latin dataset that contains around images captured from US cities [37]. The images are available with line-level annotations. The French Street Name Signs (FSNS) data contains around annotated images, each with four street sign views. Such datasets, however, contain text-centric images. Reddy et al. [16] recently release RoadText-1K to introduce challenges with generic driving scenarios where the images are not text-centric. RoadText-1K includes video clips (each seconds long at fps) from the BDD dataset, annotated with English transcriptions [33].
Works of Photo-OCR on Non-Latin Datasets: Recently, there has been an increasing interest in scene-text recognition for non-Latin languages such as Chinese, Korean, Devanagari, Japanese, etc. Several datasets like RCTW ( scene images), ReCTS-25k ( signboard images), CTW ( scene images), and RRC-LSVT ( scene images) from ICDAR’19 Robust Reading Competition (RRC) exist for Chinese [24, 36, 34, 26]. Arabic datasets like ARASTEC ( images of signboards, hoardings, and advertisements) and ALIF ( text images from TV Broadcast) also exist in the scene-text recognition community [29, 32]. Korean and Japanese scene-text recognition datasets include KAIST ( images from signboards, book covers, and English and Korean characters) and DOST ( sequential images) [7, 5]. The MLT dataset available from the ICDAR’17 RRC contains scene images (around images per language) in Arabic, Bangla, Chinese, English, French, German, Italian, Japanese, and Korean [15]. The ICDAR’19 RRC builds MLT-19 over top of MLT-17 to contain scene images containing text from Arabic, Bangla, Chinese, English, French, German, Italian, Japanese, Korean, and Devanagari [14]. The RRC also provides synthetic images in these languages to assist the training. Mathew et al. [13] train the conventional encoder-decoder, where Convolutional Neural Network (CNN) encodes the word image features. An RNN decodes them to produce text on synthetic data for Indian languages. Here an additional connectionist temporal classification (CTC) layer aligns the RNN’s output to labels. The work also releases an IIIT-ILST dataset for testing that reports Word Recognition Rates (WRRs) of , , and on real images in Hindi, Telugu, and Malayalam, respectively. Bušta et al. [2] proposes a CNN (and CTC) based method for text localization, script identification, and text recognition. The model is trained and tested on languages of MLT-17 dataset. The WRRs are above for Latin and Hangul and are below for the remaining languages. The WRR reported for Bengali is . Recently, an OCR-on-the-go model and obtain the WRR of on the IIIT-ILST Hindi dataset and the Character Recognition Rate (CRR) of on a multi-lingual dataset containing videos in English, Hindi, and Marathi [20]. Around videos in these languages recorded with controlled camera movements like tilt, pan, etc., are additionally shared at https://catalist-2021.github.io/.
Transfer Learning in Photo-OCR: With the advent of deep learning in the last decade, transfer learning became an essential part of vision models for tasks such as detection and segmentation. [17, 18]. The CNN layers pre-trained from the Imagenet classification dataset are conventionally used in such models for better initialization and performance [19]. The scene-text recognition works also use the CNN layers from the models pre-trained on Imagenet dataset [21, 11, 23]. However, to our best knowledge, there are no significant efforts on transfer learning from one language to another in the field of scene-text recognition, although transfer learning seems to be naturally suitable for reading low resource languages. We investigate the possibilities of transfer learning in all the layers of deep photo-OCR models.
Language | # Images | Train | Test | , word length | # Fonts |
---|---|---|---|---|---|
English | 17.5M | 17M | 0.5M | 5.12, 2.99 | 1200 |
Gujarati | 2.5M | 2M | 0.5M | 5.95, 1.85 | 12 |
Hindi | 2.5M | 2M | 0.5M | 8.73, 3.10 | 97 |
Bangla | 2.5M | 2M | 0.5M | 8.48, 2.98 | 68 |
Tamil | 2.5M | 2M | 0.5M | 10.92, 3.75 | 158 |
Telugu | 5M | 5M | 0.5M | 9.75, 3.43 | 62 |
Malayalam | 7.5M | 7M | 0.5M | 12.29, 4.98 | 20 |
2 Datasets and Motivation
We now discuss the datasets we use and the motivation for our work.
Synthetic Datasets: As shown in Table 1, we generate , or more, word images each in Hindi, Bangla, Tamil, Telugu, and Malayalam111For Telugu and Malayalam, our models trained on word images achieved results lower than previous works, so we generate more examples equal to Mathew et al. [13]. with the methodology proposed by Mathew et al. [13]. For each Indian language, we use images for training our models and the remaining set for testing.











Sample images of our synthetic data are shown in Fig. 2. For English, we use the models pre-trained on the MJSynth and SynthText images [6, 3]. We generate synthetic images in English with over fonts for testing. As shown in Table 1, English has a lower average word length than Indian languages. We list the Indian languages in the increasing order of language complexity, with visually similar scripts placed consecutively, in Table 1. Gujarati is chosen as the entry point from English to Indian languages as it has the lowest word length among all Indian languages. Subsequently, like English, Gujarati does not have a top-connector line that connects different characters to form a word in Hindi and Bangla (refer to Fig. 1 and 2). Also, the number of Unicode fonts available in Gujarati is fewer than those available in other Indian languages. Next, we choose Hindi, as Hindi characters are similar to Gujarati characters and the average word length of Hindi is higher than Gujarati. Bangla has comparable word length statistics with Hindi and shares the property of the top-connector line with Hindi. Still, we keep it after Hindi in the list as its characters are visually dissimilar and more complicated than Gujarati and Hindi. We use less than for fonts in Hindi, Bangla, and Telugu. We list Tamil after Bangla because these languages share similar vowels’ appearance (see the glyphs above general characters in Fig. 2). Tamil and Malayalam have the highest variability in word length and visual complexity compared to other languages. Please note that we have over fonts available in Tamil.
Real Datasets: We also perform experiments on the real datasets from IIIT-ILST, MLT-17, and MLT-19 datasets (refer to Section 1.1 for these datasets). To enlarge scene-text recognition research in complex and straight forward low-resource Indian Languages, we release and annotated word images in Gujarati and Tamil. We crop the word images from annotated scene images, which we obtain by capturing and compiling Google images. We illustrate sample annotated images of different datasets in Fig. 1. Similar to MLT datasets, we annotate the Gujarati and Tamil datasets using four corner points around each word (see Tamil image at bottom-right of Fig. 1). IIIT-ILST dataset has two-point annotations leading to an issue of text from other words in the background of a cropped word image as shown in the Hindi scene at the top-middle of Fig. 1.
Motivation: As discussed earlier in Section 1.1, most of the scene-text recognition works use the pre-trained Convolutional Neural Networks (CNN) layers for improving results. We now motivate the need for transfer learning of the complete recognition models discussed in Section 1 and the models we use in Section 3 among different languages. As discussed in these sections, the Recurrent Neural Networks (RNNs) form another integral component of such reading models. Therefore, we illustrate the distribution of character-level n-grams they learn in Fig. 3222For plots on the right, we use moving average of 10, 100, 1000, 1000, 1000 for 1-grams, 2-grams, 3-grams, 4-grams, and 5-grams, respectively. for the first five languages we discussed in the previous section (we notice that the last two languages also follow the similar trend). On the left, we show the frequency distribution of top-5 n-grams, (). On the right, we show the frequency distribution of all n-grams with . We use words from each language for these plots. We consider both capital and small letters separately for English, as it is crucial for the text recognition task. Despite this, we note that top-5 n-grams are composed of small letters. The Indian languages, however, do not have small and capital letters like English. However, the total number of English letters (given that small letters are different from capitals) is of the same order as Indian languages. The x-values () for the drops in 1-gram plots (blue curves) of Fig. 3 also illustrates this. So it becomes possible to compare the distributions. Next, we note that most of the top-5 n-grams comprise vowels for all the languages. Moreover, the overall distributions are similar for all the languages. Hence, we propose that the RNN layers’ transfer among the models of different languages is worth an investigation.
It is important to note the differences between the n-grams of English and Indian languages. Many of the top-5 n-grams in English are the complete word forms, which is not the case with Indian languages owing to their richness in inflections (or fusions) [30]. Also, note that the second and the third 1-gram for Hindi and Bangla in Fig. 3 (left), known as Halanta, is a common feature of top-5 Indic n-grams. The Halanta forms an essential part of joint glyphs or aksharas (as advocated by Vinitha et al. [30]). In Figs. 1 and 2, the vowels, or portions of the joint glyphs for word images in Indian languages, often appear above the top-connector line or below the generic consonants. All this, in addition to complex glyphs in Indian languages, makes transfer learning from English to Indian languages ineffective, which is detailed in Section 5. Thus, we also investigate the transferability of features among the Indic scene-text recognition models in the subsequent sections.
3 Models
This section explains the two models we use for transfer learning in Indian languages and a plug-in module we propose for learning the correction mechanism in the recognition systems.
CRNN Model: The first model we train is Convolutional-Recurrent Neural Network (CRNN), which is the combination of CNN and RNN as shown in Fig. 4 (left). The CRNN network architecture consists of three fundamental components, i) an encoder composed of the standard VGG model [25], ii) a decoder consisting of RNN, and iii) a Connectionist Temporal Classification (CTC) layer to align the decoded sequence with ground truth. The CNN-based encoder consists of seven layers to extract feature representations from the input image. The model abandons fully connected layers for compactness and efficiency. It replaces standard squared pooling with sized rectangular pooling windows for and max-pooling layer to yield feature maps with a larger width. A two-layer Bi-directional Long Short-Term Memory (BiLSTM) model, each with a hidden size of units, then decodes the features. During the training phase, the CTC layer provides non-parameterized supervision to align the decoded predictions with the ground truth. The greedy decoding is used during the testing stage. We use the PyTorch implementation of the model by Shi et al. [22].

STAR-Net: As shown in Fig. 4 (right), the STAR-Net model consists of three components, i) a Spatial Transformer to handle image distortions, ii) a Residue Feature Extractor consisting of a residue CNN and an RNN, and iii) a CTC layer to align the predicted and ground truth sequences. The transformer consists of a spatial attention mechanism achieved via a CNN-based localization network, a sample, and an interpolator. The localizer predicts the parameters of an affine transformation. The sampler and the nearest-neighbor interpolator use the transformation to obtain a better version of the input image. The transformed image acts as the input to the Residue Feature Extractor, which includes the CNN and a single-layer BiLSTM of units. The CNN used here is based on the inception-resnet architecture, which can extract robust image features required for the task of scene-text recognition [27]. The CTC layer finally provides the non-parameterized supervision for text alignment. The overall model consists of convolutional layers and is end-to-end trainable [11].
Correction BiLSTM: After training the STAR-Net model on a real dataset, we add a correction BiLSTM layer (of size ), an end-to-end trainable module, to the end of the model (see Fig. 4 top-right). We train the complete model again on the same dataset to implicitly learn the error correction mechanism.
4 Experiments
The images, resized to , form the input of STAR-Net. The spatial transformer module, as shown in Fig. 4 (right), then outputs the image of size . The inputs to the CNN Layers of CRNN and STAR-Net are of the same size, i.e., , and the output size is . The STAR-Net localization network has four plain convolutional layers with , , , and channels. Each layer has the filter size, stride, and padding size of , , and , followed by a max-pooling layer with a stride of . Finally, a fully connected layer of size outputs the parameters which transform the input image. We train all our models on or more synthetic word images as discussed in Section 2. We use the batch size of and the ADADELTA optimizer for stochastic gradient descent (SGD) for all the experiments [35]. The number of epochs varies between to for different experiments. We test our models on 0.5 M synthetic images for each language. We use the word images from IIIT-ILST, MLT-17, and MLT-19 datasets for testing on real datasets. We fine-tune the Bangla models on training images and test them on validation images from the MLT-17 dataset to fairly compare with Bušta et al. [1]. Similarly, we fine-tune only our best Hindi model on the MLT-19 dataset and test it on the IIIT-ILST dataset to compare with OCR-on-the-go (since it is also trained on real data) [20]. To demonstrate generalizability, we also test our models on Hindi images and Bangla images available from MLT-19 datasets [14]. For Gujarati and Tamil, we use of word images to fine-tune our models and the remaining for testing.
Language | CRNN-CRR | CRNN-WRR | STAR-Net-CRR | STAR-Net-WRR | ||||
---|---|---|---|---|---|---|---|---|
English | 77.13 | 38.21 | 86.04 | 57.28 | ||||
Gujarati | 94.43 | 81.85 | 97.80 | 91.40 | ||||
Hindi | 89.83 | 73.15 | 95.78 | 83.93 | ||||
Bangla | 91.54 | 70.76 | 95.52 | 82.79 | ||||
Tamil | 82.86 | 48.19 | 95.40 | 79.90 | ||||
Telugu | 87.31 | 58.01 | 92.54 | 71.97 | ||||
Malayalam | 92.12 | 70.56 | 95.84 | 82.10 |
5 Results
In this section, we discuss the results of our experiments with i) individual models for each language, ii) the transfer learning from English to two Indian languages, and iii) the transfer learning from one Indian language to another.
Performance on Synthetic Datasets: It is essential to compare the results on synthetic datasets of different languages sharing common backgrounds, as it provides a good intuition about the difficulty in reading different scripts. In Tables 2 and 3, we present the results of our experiments with synthetic datasets. As noted in Table 2, the CRNN model achieves the Character Recognition Rates (CRRs) and Word Recognition Rates (WRRs) of i) 77.13% and 38.21% in English and ii) above and on the synthetic dataset of all the Indian languages (refer to columns 1 and 2 of Table 2). The low accuracy on the English synthetic test set is due to the presence of more than 1200 different fonts (refer Section 2). Nevertheless, using a large number of fonts in training helps in generalizing the model for real settings [6, 3]. The STAR-Net achieves remarkably better performance than CRNN on all the datasets, with the CRRs and WRRs above and for Indian languages. The reason for this is spatial attention mechanism and powerful residual layers, as discussed in Section 3. As shown in columns 3 and 5 of Table 2, the WRR of the models trained in Gujarati, Hindi, and Bangla are higher than the other three Indian languages despite common backgrounds. The experiments show that the scripts in latter languages pose a tougher reading challenge than the scripts in former languages.
Language | CRNN-CRR | CRNN-WRR | STAR-Net-CRR | STAR-Net-WRR | ||||
English Gujarati | 92.71 (94.43) | 77.06 (81.85) | 97.50 (97.80) | 90.90 (91.40) | ||||
English Hindi | 88.11 (89.83) | 70.12 (73.15) | 94.50 (95.78) | 80.90 (83.93) | ||||
Gujarati Hindi | 91.98 (89.83) | 73.12 (73.15) | 96.12 (95.78) | 84.32 (83.93) | ||||
Hindi Bangla | 91.13 (91.54) | 70.22 (70.76) | 95.66 (95.52) | 82.81 (82.79) | ||||
Bangla Tamil | 81.18 (82.86) | 44.74 (48.19) | 95.95 (95.40) | 81.73 (79.90) | ||||
Tamil Telugu | 87.20 (87.31) | 56.24 (58.01) | 93.25 (92.54) | 74.04 (71.97) | ||||
Telugu Malayalam | 90.62 (92.12) | 65.78 (70.56) | 94.67 (95.84) | 77.97 (82.10) |
We present the results of our transfer learning experiments on the synthetic datasets in Table 3. The best individual model results from Table 2 are included in parenthesis for comparison. We begin with the English models as the base because the models have trained on over fonts and word images as discussed in Section 2, and are generic. However, in the first two rows of the table, we note that transferring the layers from the model trained on the English dataset to Gujarati and Hindi is inefficient in improving the results compared to the individual models. The possible reason for the inefficiency is that Indic scripts have many different visual and slightly different n-gram characteristics from English, as discussed in Section 2. We then note that as we try to apply transfer learning among Indian languages with CRNN (rows 3-7, columns 1-2 in Table 3), only some combinations work well. However, with STAR-Net (rows 3-7, columns 3-4 in Table 3), transfer learning helps improve results on the synthetic dataset from a simple language to a complex language333We also discovered experiments on transfer learning from a tricky language to a simple one to be effective but slightly lesser than the reported results.. For Malayalam, we observe that the individual STAR-Net model is better than the one transferred from Telugu, perhaps due to high average word length (refer Section 2).
Language | Dataset | # Images | Model | CRR | WRR |
CRNN | 84.93 | 72.08 | |||
Gujarati | ours | 125 | STAR-Net | 88.55 | 74.60 |
STAR-Net EngGuj | 78.48 | 60.18 | |||
STAR-Net HinGuj | 90.82 | 76.98 | |||
Mathew et al. [13] | 75.60 | 42.90 | |||
CRNN | 78.84 | 46.56 | |||
STAR-Net | 78.72 | 46.60 | |||
Hindi | IIIT-ILST | 1150 | STAR-Net EngHin | 77.43 | 44.81 |
STAR-Net GujHin | 79.12 | 47.79 | |||
OCR-on-the-go [20] | - | 51.09 | |||
STAR-Net GujHin FT444Fine-tuned on MLT-19 dataset as discussed earlier. We fine-tune all the layers. | 83.64 | 56.77 | |||
CRNN | 86.56 | 64.97 | |||
Hindi | MLT-19 | 3766 | STAR-Net | 86.53 | 65.79 |
STAR-Net GujHin | 89.42 | 72.96 | |||
Bušta et al. [2] | 68.60 | 34.20 | |||
Bangla | MLT-17 | 673 | CRNN | 71.16 | 52.74 |
STAR-Net | 71.56 | 55.48 | |||
STAR-Net HinBan | 72.16 | 57.01 | |||
W/t Correction BiLSTM | 83.30 | 58.07 | |||
CRNN | 81.93 | 74.26 | |||
Bangla | MLT-19 | 3691 | STAR-Net | 82.80 | 77.48 |
STAR-Net HinBan | 82.91 | 78.02 | |||
CRNN | 90.17 | 70.44 | |||
Tamil | ours | 634 | STAR-Net | 89.69 | 71.54 |
STAR-Net BanTam | 89.97 | 72.95 | |||
Mathew et al. [13] | 86.20 | 57.20 | |||
Telugu | IIIT-ILST | 1211 | CRNN | 81.91 | 58.13 |
STAR-Net | 82.21 | 59.12 | |||
STAR-Net TamTel | 82.39 | 62.13 | |||
Mathew et al. [13] | 92.80 | 73.40 | |||
Malayalam | IIIT-ILST | 807 | CRNN | 84.12 | 70.36 |
STAR-Net | 91.50 | 72.73 | |||
STAR-Net TelMal | 92.70 | 75.21 |


Performance on Real Datasets: Table 4 depicts the performance of our models on the real datasets. At first, we observe that for each Indian language, the overall performance of the individual STAR-Net model is better than the individual CRNN model (except for Gujarati and Hindi, where the results are very close). Based on this and similar observations in the previous section, we present the results of transfer learning experiments on real datasets only with the STAR-Net model555We also tried transfer learning with CRNN; STAR-Net was more effective.. Next, similar to the previous section, we observe that the transfer learning from English to Gujarati and Hindi IIIT-ILST datasets (rows and in Table 4) is not as effective as individual models in these Indian languages (rows and in Table 4). Finally, we observe that the performance improves with the transfer learning from a simple language to a complex language, except for HindiGujarati, for which Hindi is the only most straightforward choice. We achieve performance better than the previous works, i.e., Bušta et al. [1], Mathew et al. [13], and OCR-on-the-go [20]. Overall, we observe the increase in WRRs by , , and on IIIT-ILST Hindi, Telugu, and Malayalam, and MLT-17 Bangla datasets compared to the previous works. On the MLT-19 Hindi and Bangla datasets, we achieve gains of and in WRR over the baseline individual CRNN models. On the datasets we release for Gujarati and Tamil, we improve the baselines by and increase in WRRs. We present the qualitative results of our baseline CRNN models as well as best transfer learning models in Fig. 1. The green and red colors represent the correct predictions and errors, respectively. “_” represents the missing character. As can be seen, most of the mistakes are single-character errors.
Since we observe the highest gain of in WRR (and in CRR) for the MLT-17 Bangla dataset (Table 4), we further try to improve these results. We plug in in the correction BiLSTM (refer Section 3) to the best model (row 18 of Table 4). The results are shown in row 19 of Table 4. As shown, the correction BiLSTM improves the CRR further by a notable margin of since the BiLSTM works on character level. We also observe the WRR gain, thereby achieving the overall WRR gain (and CRR gain) over Bušta et al. [1].
Features Visualization: In Fig. 5 for the CRNN model (top three triplets), we visualize the learned CNN layers of the individual Hindi model, the “English Hindi” model, and the “GujaratiHindi” model. The red boxes are the regions where the first four CNN layers of the model transferred from English to Hindi are different from the other two models. The feature visualization again strengthens our claim that transfer from the English reading model to any Indian language dataset is inefficient. We notice a similar trend for the Gujarati STAR-Net models, though the initial CNN layers look very similar to word images (bottom three triplets in Fig. 5). The similarity also demonstrates the better learnability of STAR-Net compared to CRNN, as observed in previous sections.
6 Conclusion
We generated or more synthetic images in six different Indian languages with varying complexities to investigate the language transfers for two scene-text recognition models. The underlying view is that the transfer of image features is standard in deep models, and the transfer of language text features is a plausible and natural choice for the reading models. However, we observe that transferring the generic English photo-OCR models (trained on over fonts) to Indian languages is inefficient. Our models transferred from one Indian language to another perform better than the previous works or the new baselines we created for individual languages. We, therefore, set the new benchmarks for scene-text recognition in low-resource Indian languages. The proposed Correction BiLSTM, when plugged into the STAR-Net model and trained end-to-end, further improves the results.
References
- [1] Bušta, M., Neumann, L., Matas, J.: Deep textspotter: An end-to-end trainable scene text localization and recognition framework. ICCV (2017)
- [2] Bušta, M., Patel, Y., Matas, J.: E2E-MLT-an Unconstrained End-to-End Method for Multi-Language Scene Text. In: Asian Conference on Computer Vision. pp. 127–143. Springer (2018)
- [3] Gupta, A., Vedaldi, A., Zisserman, A.: Synthetic Data for Text Localisation in Natural Images. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 2315–2324 (2016)
- [4] Huang, Z., Zhong, Z., Sun, L., Huo, Q.: Mask R-CNN with Pyramid Attention Network for Scene Text Detection. In: WACV. pp. 764–772. IEEE (2019)
- [5] Iwamura, M., Matsuda, T., Morimoto, N., Sato, H., Ikeda, Y., Kise, K.: Downtown Osaka Scene Text Dataset. In: ECCV. pp. 440–455. Springer (2016)
- [6] Jaderberg, M., Simonyan, K., Vedaldi, A., Zisserman, A.: Synthetic Data and Artificial Neural Networks for Natural Scene Text Recognition. In: Workshop on Deep Learning, NIPS (2014)
- [7] Jung, J., Lee, S., Cho, M.S., Kim, J.H.: Touch TT: Scene Text Extractor using Touchscreen Interface. ETRI Journal 33(1), 78–88 (2011)
- [8] Karatzas, D., Gomez-Bigorda, L., Nicolaou, A., Ghosh, S., Bagdanov, A., Iwamura, M., Matas, J., Neumann, L., Chandrasekhar, V.R., Lu, S., et al.: ICDAR 2015 Competition on Robust Reading. In: 2015 ICDAR. pp. 1156–1160. IEEE (2015)
- [9] Lee, C.Y., Osindero, S.: Recursive Recurrent Nets with Attention Modeling for OCR in the Wild. In: CVPR. pp. 2231–2239 (2016)
- [10] Liao, M., Shi, B., Bai, X., Wang, X., Liu, W.: Textboxes: A Fast Text Detector with a Single Deep Neural Network. In: AAAI. pp. 4161–4167 (2017)
- [11] Liu, W., Chen, C., Wong, K.Y.K., Su, Z., Han, J.: STAR-Net: A SpaTial Attention Residue Network for Scene Text Recognition. In: BMVC. vol. 2 (2016)
- [12] Liu, X., Liang, D., Yan, S., Chen, D., Qiao, Y., Yan, J.: FOTS: Fast Oriented Text Spotting with a Unified Network. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 5676–5685 (2018)
- [13] Mathew, M., Jain, M., Jawahar, C.: Benchmarking Scene Text Recognition in Devanagari, Telugu and Malayalam. In: ICDAR. vol. 7, pp. 42–46. IEEE (2017)
- [14] Nayef, N., Patel, Y., Busta, M., Chowdhury, P.N., Karatzas, D., Khlif, W., Matas, J., Pal, U., Burie, J.C., Liu, C.l., et al.: ICDAR2019 Robust Reading Challenge on Multi-lingual Scene Text Detection and Recognition–RRC-MLT-2019. In: 2019 International Conference on Document Analysis and Recognition (ICDAR). pp. 1582–1587. IEEE (2019)
- [15] Nayef, N., Yin, F., Bizid, I., Choi, H., Feng, Y., Karatzas, D., Luo, Z., Pal, U., Rigaud, C., Chazalon, J., et al.: Robust Reading Challenge on Multi-lingual Scene Text Detection and Script Identification – RRC-MLT. In: 14th ICDAR. vol. 1, pp. 1454–1459. IEEE (2017)
- [16] Reddy, S., Mathew, M., Gomez, L., Rusinol, M., Karatzas, D., Jawahar, C.: RoadText-1K: Text Detection & Recognition Dataset for Driving Videos. In: 2020 IEEE International Conference on Robotics and Automation (ICRA). pp. 11074–11080. IEEE (2020)
- [17] Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. arXiv preprint arXiv:1506.01497 (2015)
- [18] Romera, E., Alvarez, J.M., Bergasa, L.M., Arroyo, R.: ERFnet: Efficient Residual Factorized ConvNet for Real-Time Semantic Segmentation. IEEE Transactions on Intelligent Transportation Systems 19(1), 263–272 (2017)
- [19] Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al.: Imagenet Large Scale Visual Recognition Challenge. International journal of computer vision 115(3), 211–252 (2015)
- [20] Saluja, R., Maheshwari, A., Ramakrishnan, G., Chaudhuri, P., Carman, M.: OCR On-the-Go: Robust End-to-end Systems for Reading License Plates and Street Signs. In: 15th IAPR International Conference on Document Analysis and Recognition (ICDAR). pp. 154–159. IEEE (2019)
- [21] Shi, B., Bai, X., Yao, C.: An End-to-End Trainable Neural Network for Image-Based Sequence Recognition and I]ts Application to Scene Text Recognition. IEEE transactions on pattern analysis and machine intelligence 39(11), 2298–2304 (2016)
- [22] Shi, B., Bai, X., Yao, C.: An End-to-End Trainable Neural Network for Image-Based Sequence Recognition and Its Application to Scene Text Recognition. IEEE TPAMI 39(11), 2298–2304 (2017)
- [23] Shi, B., Yang, M., Wang, X., Lyu, P., Yao, C., Bai, X.: ASTER: An Attentional Scene Text Recognizer with Flexible Rectification. IEEE Transactions on Pattern Analysis and Machine Intelligence (2018)
- [24] Shi, B., Yao, C., Liao, M., Yang, M., Xu, P., Cui, L., Belongie, S., Lu, S., Bai, X.: ICDAR2017 Competition on Reading Chinese Text in the Wild (RCTW-17). In: 14th ICDAR. vol. 1, pp. 1429–1434. IEEE (2017)
- [25] Simonyan, K., Zisserman, A.: Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv:1409.1556 (2014)
- [26] Sun, Y., Ni, Z., Chng, C.K., Liu, Y., Luo, C., Ng, C.C., Han, J., Ding, E., Liu, J., Karatzas, D., et al.: ICDAR 2019 Competition on Large-Scale Street View Text with Partial Labeling - RRC-LSVT. In: 2019 International Conference on Document Analysis and Recognition (ICDAR). pp. 1557–1562. IEEE (2019)
- [27] Szegedy, C., Ioffe, S., Vanhoucke, V., Alemi, A.: Inception-v4, Inception-Resnet and the Impact of Residual Connections on Learning. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 31 (2017)
- [28] Tian, Z., Huang, W., He, T., He, P., Qiao, Y.: Detecting Text in Natural Image with Connectionist Text Proposal Network. In: ECCV. pp. 56–72 (2016)
- [29] Tounsi, M., Moalla, I., Alimi, A.M., Lebouregois, F.: Arabic Characters Recognition in Natural Scenes using Sparse Coding for Feature Representations. In: 13th ICDAR. pp. 1036–1040. IEEE (2015)
- [30] Vinitha, V., Jawahar, C.: Error Detection in Indic OCRs. In: 2016 12th IAPR Workshop on Document Analysis Systems (DAS). pp. 180–185. IEEE (2016)
- [31] Wang, K., Babenko, B., Belongie, S.: End-to-End Scene Text Recognition. In: ICCV. pp. 1457–1464. IEEE (2011)
- [32] Yousfi, S., Berrani, S.A., Garcia, C.: ALIF: A Dataset for Arabic Embedded Text Recognition in TV Broadcast. In: 2015 13th International Conference on Document Analysis and Recognition (ICDAR). pp. 1221–1225. IEEE (2015)
- [33] Yu, F., Xian, W., Chen, Y., Liu, F., Liao, M., Madhavan, V., Darrell, T.: BDD100K: A Diverse Driving Video Database with Scalable Annotation Tooling. arXiv preprint arXiv:1805.04687 2(5), 6 (2018)
- [34] Yuan, T., Zhu, Z., Xu, K., Li, C., Mu, T., Hu, S.: A Large Chinese Text Dataset in the Wild. Journal of Computer Science and Technology 34(3), 509–521 (2019)
- [35] Zeiler, M.: Adadelta: An adaptive learning rate method 1212 (12 2012)
- [36] Zhang, R., Zhou, Y., Jiang, Q., Song, Q., Li, N., Zhou, K., Wang, L., Wang, D., Liao, M., Yang, M., et al.: ICDAR 2019 Robust Reading Challenge on Reading Chinese Text on Signboard. In: 2019 International Conference on Document Analysis and Recognition (ICDAR). pp. 1577–1581. IEEE (2019)
- [37] Zhang, Y., Gueguen, L., Zharkov, I., Zhang, P., Seifert, K., Kadlec, B.: Uber-text: A Large-Scale Dataset for Optical Character Recognition from Street-Level Imagery. In: SUNw: Scene Understanding Workshop-CVPR. vol. 2017 (2017)
- [38] Zhou, X., Yao, C., Wen, H., Wang, Y., Zhou, S., He, W., Liang, J.: East: An Efficient and Accurate Scene Text Detector. In: CVPR. pp. 2642–2651 (2017)