Conference PaperPDF Available

Handwritten word recognition based on structural characteristics and lexical support

Authors:

Abstract and Figures

In this paper a handwritten recognition algorithm based on structural characteristics, histograms and profiles, is presented. The well-known horizontal and vertical histograms are used, in combination with the newly introduced radial histogram, out-in radial and in-out radial profiles for representing 32 × 32 matrices of characters, as 280-dimension vectors. The recognition process has been supported by a lexical component based on dynamic acyclic FSAs (Finite-State-Automata).
Content may be subject to copyright.
Handwritten Word Recognition based on Structural Characteristics and Lexical
Support
E. Kavallieratou, K. Sgarbas, N. Fakotakis and G. Kokkinakis
Wire Communications Lab.
University of Patras, 26500 Patras
ergina@wcl.ee.upatras.gr
Abstract
In this paper a handwritten recognition algorithm
based on structural characteristics, histograms and
profiles, is presented. The well-known horizontal and
vertical histograms are used, in combination with the
newly introduced radial histogram, out-in radial and in-
out radial profiles for representing 32x32 matrices of
characters, as 280-dimension vectors.
The recognition process has been supported by a
lexical component based on dynamic acyclic FSAs
(Finite-State-Automata).
1.Introduction
In this paper a handwritten character recognition
approach is presented. The proposed technique has been
developed as the last module of an integrated document
analysis system, shown in fig. 1 [1-6].
In general, the character recognition procedure consists
of two steps: (i) feature extraction where each character is
represented as a feature vector and (ii) classification of
the vectors into a number of classes. Govindan [7]
distinguishes two categories of features: the structural and
the statistical features, while Bunke [8] estimates that the
structural approach is closer to the human way of
recognition. In this paper, we propose a structural
approach for feature extraction. Thus, a 280-dimension
vector is extracted for each character, consisting of
histograms and profiles. One new histogram and two new
profiles are introduced. The k-means algorithm is, then,
used for classification.
The proposed technique, described in section 2, is fast
and simple, while the experimental results, illustrated in
section 4, are quite promising. An accompanying lexical
component is described in section 3. Finally, in section 5
our suggestions and plans for further improving are
presented.
Figure 1. An integrated OCR system.
Skew Angle
Estimation and
Correction
Printed -
Handwritten
Text
Discrimination
Line
segmentation
Slant
Correction
Word
Segmentation
Character
Segmentation
Character
Recognition
Document
Image
Text
Character
Classes
Segmentation
Rules
P
REPROCESS
I
NG
Lexical
Support
Text
Proceedings of the Seventh International Conference on Document Analysis and Recognition (ICDAR 2003)
0-7695-1960-1/03 $17.00 © 2003 IEEE
(a) (b) (c) (d) (e) (f)
Figure 2. a) Character matrix, b) Horizontal histogram, c) Vertical histogram, d) Radial histogram,
e) Radial out-in profile and f) Radial in-out profile.
2. Algorithm description
As can been seen in fig. 1, after the preprocessing
stage, isolated character images are produced which are
used as input to the character recognition module. Each
character is, then, represented as a 280-dimension vector,
consisting of histograms and profiles. In more detail, each
character is normalized in a 32x32 matrix. The horizontal
histogram, the vertical histogram, the radial histogram,
the out-in profile and the in-out radial profile are, then,
calculated. The first two histograms are well- known and
have been used extensively in optical character
recognition research while the remaining features are used
for first time. The formal definition of these features is
follows.
Consider that the value of the element in the m-th row
and n-th column of the character matrix is given by a
function f:
mn
anmf =),(
where α
mn
takes binary values (i.e., 0 for white pixels
and 1 for black pixels). The horizontal histogram H
h
of
the character matrix is the sum of black pixels in each row
(i.e., 32 features):
=
n
h
nmfmH ),()(
Similarly, the vertical histogram H
v
of the character
matrix is the sum of black pixels in each column (i.e., 32
features):
=
m
v
nmfnH ),()(
We define the value of the radial histogram H
r
at an
angle φ as the sum of black pixels on a rad that starts from
the center of the character matrix (i.e., the element in the
16
th
row and the 16
th
column) and ends up at the border of
the matrix, forming an angle φ with the horizontal axis.
The radial histogram values are calculated with a step of 5
degrees (i.e., 72 features):
)72,0[,*5
),cos16,sin16()(
16
1
=
+=
=
kk
iifH
r
φ
φφφ
ι
Additionally, we define the value of the out-in radial
profile P
oi
at an angle φ as the position of the first black
pixel found on the rad that starts from the periphery and
goes to the center of the character matrix forming an
angle φ with the horizontal axis. The out-in radial profile
values are calculated with a step of 5 degrees (i.e., 72
features):
)72,0[,*5
,
1)cos16,sin16(&
0)cos16,sin16(:
)(
1
16
=
+
+
=
=
kk
IIf
iifI
P
I
i
oi
φ
φφ
φφ
φ
Similarly, we define the value of the in-out radial
profile P
io
at an angle φ as the position of the first black
pixel on the rad that starts from the center and goes to the
periphery of the character forming an angle φ with the
horizontal axis. As above, the in-out radial profile values
are calculated with a step of 5 degrees (i.e., 72 features):
)72,0[,*5
,
1)cos16,sin16(&
0)cos16,sin16(:
)(
1
0
=
+
+
=
=
kk
JJf
iifJ
P
J
i
io
φ
φφ
φφ
φ
An illustrated example is given in fig. 2, where the
previously defined histograms and profiles (i.e., fig. 2b,
2c, 2d, and 2e) for a handwritten character (i.e., fig 2a) are
shown.
Thus, a 280-dimension vector is extracted for each
character image. A classification model is, then, produced
by applying the k-means algorithm to the training data.
The preprocessing of the document images (see fig. 1)
is likely to cause the undesirable segmentation of a
handwritten character image into two character images.
Taking this into account, during the classification of
unseen cases, a feature vector is extracted for each
0
5
10
15
20
25
30
0
5
10
15
20
25
30
0
5
10
15
20
25
30
0
5
10
15
20
25
30
0
10
20
30
40
50
60
70
0
5
10
15
Proceedings of the Seventh International Conference on Document Analysis and Recognition (ICDAR 2003)
0-7695-1960-1/03 $17.00 © 2003 IEEE
character image as well as for each pair of successive
character images. These vectors are classified into the
character class from which the Euclidean distance is
minimized.
3. Lexical support
The recognition process has been supported by a
lexical component based on dynamic acyclic FSAs
(Finite-State-Automata).
Acyclic FSAs provide a very efficient data structure
for lexicon representation and fast string matching, with a
great variety of applications [9-11]. They constitute very
compact representations of lexicons, since common
prefixes and suffixes are represented by the same
transitions (see Fig.3).
The FSAs consist of states and transitions between
states. Each transition has a label. The words are stored as
directed paths on the graph. They can be retrieved by
traversing the graph from an initial state (source) to a
terminal state (sink), collecting the labels of the
transitions encountered. In this way, traversing the graphs
of Fig.1 from the source () to the sink (~) we retrieve
the strings dance, darts, dart, start and smart. Storing
lexicons in FSAs is efficient because common transitions
are used to represent identical suffixes and prefixes, and
because the stρing or substring search can be performed
in various flexible ways (including content-addressable
and approximate matching).
There are two types of FSAs. The graph of Fig.1a is
called deterministic (DFA) because no transitions exist
that have the same labels and leave the same state. This
property results to a very efficient search function (linear
to the size of the string). Graphs that do not have this
property, like the one of Fig.1b, are called non-
deterministic automata (NFA). NFAs require less space
than DFAs to store the same lexicon, but they are a little
slower to search.
Recent research [12-21] has produced a class of
algorithms that are able to efficiently update acyclic
FSAs, even maintaining their minimality. These
algorithms can dynamically add and remove strings from
acyclic FSAs, without performing a minimization
procedure in the whole graph. They produce minimal
DFAs and near-minimal NFAs (called compact [21])
since the minimization of NFAs is still an open problem.
The constructed lexicon used in the current system
consists of 230,000 Greek words (average word length:
9.5 characters; alphabet size: 36). Figure 4 shows the
space and time performance for the process of building
the lexicon in both forms (DFA and NFA) using a pair of
dynamic algorithms of the aforementioned class
(described in detail in [19-21]). The number of states,
transitions and the construction time were measured. The
thick lines refer to the NFA; the thin lines refer to the
corresponding minimal DFA. The test was performed on
a 200 MHz PC.
Figures 4a, 4b and 4c display respectively the number
of states, the number of transitions and the construction
time of the automaton, in respect to the size of the lexicon
(number of words). The same results are also shown in
logarithmic scales in Figs.4d, 4e and 4f, respectively. The
results agree with the theoretic calculations [19,20,22],
indicating linear growth of transitions, less than linear
growth of states and an O(n) time performance for the
insertion process.
Given a word pattern, the lexicon is searched using an
improved A* search algorithm, which is able to retrieve a
list of N-best words that match the input pattern.
Character insertions, deletions and substitutions are
modelled as distance penalties according to a confusion
matrix that represents the possible alterations. Thus all
possible errors are expressed as scores in the process of
the A* search and each entry in the output list of N-best
words is accredited an overall matching score as well.
4. Experimental results
In order to evaluate our technique, we performed
experiments using the NIST [23] database for English
characters as well as the GRUHD [24] database for Greek
characters. In both cases, the recognition system was
trained using 2000 samples and 128 classes for each
symbol and was tested on 500 samples for each symbol.
The training and the test set were completely disjoint.
Thus, the writers used in testing were completely different
from the ones used in training.
In more detail, the recognition system was first trained
based on the NIST database, for each one of the following
categories separately: digits, uppercase characters and
lowercase characters. Then, it was tested on unseen cases
of the corresponding categories, taken from the same
database. The accuracy rate in each case is shown in table
d
a
n
c
e
*
*
s
t
r
s
m
t
a
r
t
Deterministic Finite-State
Automaton (DFA)
(a)
Non-Deterministic
Finite-State Automaton (NFA)
(b)
d
a
n
c
e
s
t
r
s
m
t
a
r
t
d
Figure 3. The same lexicon in DFA and
NFA.
Proceedings of the Seventh International Conference on Document Analysis and Recognition (ICDAR 2003)
0-7695-1960-1/03 $17.00 © 2003 IEEE
1. Since the output of the proposed character recogniser
could be further improved by using lexicons, the
recognition accuracy when the second and the third
choices are taken into account are also given.
Next, the algorithm was trained and tested using the
Modern Greek database of unconstrained writing
(GRUHD). Since, the forms of GRUHD are very similar
to those of the NIST database, it was possible to train our
system separately for digits, uppercase and lowercase
characters as well. The results on unseen cases of the
corresponding categories are shown on table 2. In contrast
to the NIST database, the GRUHD database contains
unconstrained handwriting, so the accuracy results are
somewhat lower in the latter case. In figures 5 and 6 the
relation of the recognition accuracy with the training set
size and the number of classes per symbol, respectively, is
illustrated. In both cases the given results concern the
characters of the NIST database.
The system trained as before, using NIST database,
was applied to about 280 forms of the IAM-DB, database
of unconstrained writing. The accuracy here, depending
on the form varied from 71%-75,8%, an accuracy rate
much lower than the one succeeded on NIST database.
Two are the reasons that contributed to that decrease. The
NIST is a database of hand-printed characters that are
well segmented and much easier in recognition. On the
other hand, in IAM-DB both lowercase and uppercase
characters had to be recognized at the same time. By
providing the system with lexical information, the
accuracy results are 86-87%.
Table 1: The accuracy rates for the NIST.
1
st
Choice
2
nd
Choice
3
rd
Choice
Digits
98.8% 99.91 100%
Uppercase
characters
93.85% 96.54 98.86%
Lowercase
characters
91.4% 94.50% 98.85%
Mixed characters
82.79 89.27% 96.85%
Table 2: Experimental results for the GRUHD.
1
st
Choice
2
nd
Choice
3
rd
Choice
Digits
94% 97.42% 99,54%
Uppercase
characters
86.03% 96.40% 98.96%
Lowercase
characters
81% 90.36% 96.60%
Mixed characters
72.8% 80.04% 88.83%
4. Conclusion
Figure 4. Space-time performance for dynamic DFA (thin lines) and NFA (thick lines).
Proceedings of the Seventh International Conference on Document Analysis and Recognition (ICDAR 2003)
0-7695-1960-1/03 $17.00 © 2003 IEEE
In this paper a technique for handwritten character
recognition is presented. The proposed technique is
focuses on the extraction of the features that best describe
a handwritten character introducing one new histogram
(i.e., radial) and two new profiles (i.e., in-out and out-in).
These features together with the well-known horizontal
and vertical histograms form a reliable representation of a
handwritten character.
The described approach has been tested on two
different databases with recognition accuracy varying
from 72.8% to 98.8% depending on the difficulty of the
database and the character category. On word recognition
level, the system has been supported by a lexical
component based on dynamic acyclic FSAs (Finite-State-
Automata).
5. References
[1] E.Kavallieratou, N.Fakotakis and G.Kokkinakis, ``Skew
Estimation using Cohen's class distributions'', Pattern
Recognition Letters 20, 1999, pp. 1305-1311.
[2] E.Kavallieratou, N.Fakotakis, and G.Kokkinakis, ``Slant
Estimation Algorithm for OCR Systems'', Pattern Recognition,
v.34, n.12, 2001, pp.2515-2522.
[3] M.Maragoudakis, E.Kavallieratou, N.Fakotakis and
G.Kokkinakis, ``How Conditional Independence Assumption
Affects Handwritten Character Segmentaion'', In Proc. ICDAR,
2001.
[4] E.Kavallieratou, D.C.Balcan, M.F.Popa, and N.Fakotakis,
``Handwritten Text Localization in Skewed Documents'', In
Proc. ICIP, 2001, pp.1102-1105.
[5] E. Kavallieratou, E. Stamatatos, N. Fakotakis, and G.
Kokkinakis, ``Handwritten Character Segmentation Using
Transformation-Based Learning'', In Proc. ICPR, 2000, pp.634-
637.
[6] Ε.Kavallieratou, N.Fakotakis, G.Kokkinakis, ``New
algorithms for skewing correction and slant removal on word-
level'', In Proc. ICECS, 1999, pp.1159-1162.
[7] V.K.Govindan, A.P.Shivaprasad, ``Character recognition --
A review'', Pattern Recognition, v. 23, No 7, 1990, pp. 671-683.
[8] H.Bunke, A.Sanfeliu, Syntactic and structural Pattern
Recognition, Theory and Applications, World Scientific,
Singapore.
[9] M. Crochemore and R. Verin. 1997. Direct Construction of
Compact Directed Acyclic Word Graphs. 8
th
Annual
Symposium, CPM 97, Aarhus, Denmark, 116-129.
[10] R. Lacouture and R. De Mori. 1991. Lexical Tree
Compression. EuroSpeech ’91, 2
nd
European Conference on
Speech Communications and Techniques, Genova, Italy, 581-
584.
[11] K. Sgarbas, N. Fakotakis and G. Kokkinakis. 2000. A
Straightforward Approach to Morphological Analysis and
Synthesis. Proc. COMLEX 2000, Workshop on Computational
Lexicography and Multimedia Dictionaries, Kato Achaia,
Greece, 31-34.
[12] J. Aoe, K. Morimoto and M. Hase. 1993. An Algorithm for
Compressing Common Suffixes Used in Trie Structures.
Systems and Computers in Japan, 24(12):31-42.
[13] M. Ciura and S. Deorowicz. 1999. Experimental Study of
Finite Automata Storing Static Lexicons. Report BW-453/RAu-
2/99 (Also at http://www-zo.iinf.polsl.gliwice.pl/~sdeor/pub
.htm).
[14] J. Daciuk, S. Mihov, B. Watson and R. Watson. 2000.
Incremental Construction of Minimal Acyclic Finite State
Automata. Computational Linguistics, 26(1):3-16.
[15] J. Daciuk, R.E. Watson and B.W. Watson. 1998.
Incremental Construction of Acyclic Finite-State Automata and
Transducers. Proceedings of Finite State Methods in Natural
Language Processing, Bilkent University, Ankara, Turkey.
[16] S. Mihov. 1998. Direct Construction of Minimal Acyclic
Finite States Automata. Annuaire de l' Universite de Sofia 'St.
Kl. Ohridski', Faculte de Mathematique et Informatique, Sofia,
Bulgaria, 92(2).
[17] K. Park, J. Aoe, K. Morimoto and M. Shishibori. 1994. An
Algorithm for Dynamic Processing of DAWG's. International
Journal of Computer Mathematics, Gordon and Breach
Publishers SA, OPA Amsterdam BV, 54:155-173.
[18] D. Revuz. 2000. Dynamic Acyclic Minimal Automaton.
CIAA 2000, 5
th
International Conference on Implementation and
Application of Automata, London, Canada pp.226-232.
[19] K. Sgarbas, N. Fakotakis and G. Kokkinakis. 1995. Two
Algorithms for Incremental Construction of Directed Acyclic
Word Graphs. International Journal on Artificial Intelligence
Tools, World Scientific, 4(3):369-381.
[20] K. Sgarbas, N. Fakotakis and G. Kokkinakis. 2000. Optimal
Insertion in Deterministic DAWGs. Technical Report
WCL/SLT#000524, Wire Communications Lab., Dept. of
Electrical and Computer Engineering, University of Patras,
Greece (Also at http://slt.wcl.ee.upatras.gr/sgarbas/PublAbsEN
.htm).
[21] K. Sgarbas, N. Fakotakis and G. Kokkinakis. 2001.
Incremental Construction of Compact Acyclic NFAs. Proc.
ACL-2001, 39
th
Annual Meeting of the ACL, Toulouse, France,
474-481.
[22] A. Blumer, D. Haussler, and A. Ehrenfeucht. 1989.
Average Sizes of Suffix Trees and DAWGs. Discrete Applied
Mathematics, 24:37-45.
[23] R. Wilkinson, J. Geist, S. Janet, P. Grother, C. Burges, R.
Creecy, B. Hammond, J. Hull, N. Larsen, T. Vogl, and C.
Wilson, The first census optical character recognition systems
conf. #NISTIR 4912. The U.S Bureau of Census and the
National Institute of Standards and Technology. Gaithersburg,
MD, 1992.
[24] E.Kavallieratou, N.Liolios, E.Koutsogeorgos, N.Fakotakis,
G.Kokkinakis, ``The GRUHD database of Modern Greek
Unconstrained Handwriting'', In Proc. ICDAR ,2001.
[25] G.Frosini, B.Lazzerini, A.Maggiore, F.Marcelloni, ``Fuzzy
classification based system for handwritten character
recognition'', In Proc. 2 nd Intern. Conf. On Knowledge-based
Intelligent Electronic Systems (KES'98), pp. 61-65, 1998.
Proceedings of the Seventh International Conference on Document Analysis and Recognition (ICDAR 2003)
0-7695-1960-1/03 $17.00 © 2003 IEEE
... In [36], the authors have proposed character recognition system which uses 280 dimension feature vector which uses histogram (radial) and profiles (in-out and out-in) and achieved varying accuracy of 72.8% to 98.8% depending on the database difficulty and character category. Alex Graves et al. [37] have introduced a novel approach for both online and offline handwritten character recognition using Recurrent Neural Network and compared the result using HMM based system. ...
Article
Full-text available
Character recognition comes into picture when various patterns of handwritten or optical characters are to be recognized digitally. Many researchers have proposed different approaches for character recognition in different languages. In this paper, we have reviewed several techniques of character recognition. The main important phases of character recognition include pre-processing, segmentation, feature extraction and classification. Various feature extraction techniques and classification techniques have been surveyed in this paper and an attempt is made here to draw a conclusion regarding HCR techniques from the literature survey.
... Vertical and horizontal histograms were used, in combination with the radial histogram, a 280-features vector was built from the representation of 32x32 image matrices for the characters using out-in radial and in-out radial profiles. The recognition process has been supported by a lexical component based on dynamic acyclic FSAs (Finite-State-Automata). [6] Using combination of multiple feature extraction methods and an ensemble classifier system is can increase the model accuracy. In [7] research paper Six feature set using different approaches were extracted depending on zoning, projections, edges, concavities and gradient to increase diversity to the feature extraction phase of the model. ...
Research
Full-text available
Handwritten Digit Recognition has been proposed using different techniques that were implemented over the available datasets. Although existing systems reached high recognition accuracy, more efforts regarding speed and memory allocation is required. In this research, we experiment the impact of image resolution reduction on recognition accuracy for handwritten digits. A set of features were extracted, include histogram of pixels for horizontal, vertical, diagonal and inversed diagonal orientations. Feature vector constructed by joining these features. Then, support vector machine is applied for classification. Rescaled handwritten digit images were experimented against recognition accuracy, speed and memory. MNIST database of handwritten digits is utilized for implementation. Results showed that the reduction of the size for the features vector due to image rescaling to quarter of the original size had only about 1% accuracy degradation impact.
... There are many different kinds of classification methods which have been proposed in different researches [7][8][9]. Every approach has its own advantages and disadvantages, so it is difficult to select a general and efficient classification approach [10]. Many researchers proposed different systems or approaches in order to improve the efficiency of the recognition process. ...
Article
Full-text available
Handwritten text recognition is one of the most valuable recognition systems because of the unique characteristics of each person's handwriting. Thus, recognition systems need to be more adaptable to recognize same or different characters with different characteristics. On the other hand, one of the most challenging tasks in handwritten text recognition problems is recognizing ancient documents which include several noise within them. While digitizing these documents, these noise appear in different types which effects any recognition system. So, digitizing ancient documents, applying proper pre-processing techniques and performing effective classifier are the main steps of efficient recognition system. In this paper, a complete Ethiopian ancient Geez character recognition system using deep convolutional neural network is proposed in order to recognize twenty-six base characters of this alphabet. The proposed system obtained an accuracy of 99.39% with a model loss of 0.044 which demonstrates its efficiency.
... E. Kavallieratou proposed structural characteristics and lexieal support approach by using Structural Analysis Algorithm for Handwritten Word Recognition based on Structural Characteristics and Lexical Support and It uses 280-dimension feature-Vector. The 72 .8% to 98.8% range of accuracy rate was achieved [55]. Z. Shaaban proposed the Texts based approach by using Neural Networks for Machine-Printed Arabic Texts based. ...
Conference Paper
Full-text available
OCR is the most active, interesting evaluation invention of text cum character processing recognition and pattern based image recognition. In present life OCR has been successfully using in finance, legal, banking, health care and home need appliances. The OCR consists the different levels of processing methods like as Image Pre Acquisition, Classification, Post-Acquisition, Pre-Level processing, Segmented Processing, Post-Level processing, Feature Extraction. The many researchers are proposed various levels of different methodologies and approaches in different versions of languages with help of modern and traditional technologies. This paper expressed the detail study and analysis of various character recognition methods and approaches: in details like as flow and type of approached methodology was used, type of algorithm has built with support of technology has implemented background of the proposed methodology and invention best outcomes flow for the each methodology. This paper and also expressed the main objectives and ideology of various OCR algorithms, like as neural networks algorithm, structural algorithm, support vector algorithm, statistical algorithm, template matching algorithm along with how they classified, identified, rule formed, inferred for recognition of characters and symbols.
... Kavallieratou et al. developed an integrated document analysis system, which has one module of the handwritten word recognition algorithm based on structural characteristics and lexical support [8]. In general, their grapheme recognition procedure consists of two steps. ...
Article
Full-text available
The article presents the improvements in the preprocessing part of the deciphering method (shortly preprocessing algorithm) for historical inscriptions of unknown origin. Glyphs used in historical inscriptions changed through time; therefore, various versions of the same script may contain different glyphs for each grapheme. The purpose of the preprocessing algorithm is reducing the running time of the deciphering process by filtering out the less probable interpretations of the examined inscription. However, the first version of the preprocessing algorithm leads incorrect outcome or no result in the output in certain cases. Therefore, its improved version was developed to find the most similar words in the dictionary by relaying the search conditions more accurately, but still computationally effectively. Moreover, a sophisticated similarity metric used to determine the possible meaning of the unknown inscription is introduced. The results of the evaluations are also detailed.
Chapter
Amharic is the working language in the Federal Democratic Republic of Ethiopia. The Amharic alphabet has a large number of symbols and there is a close resemblance among shapes of the different symbols available in the language which challenged the task of machine-based optical character recognition systems in the language. The absence of a standardized labeled dataset for the Amharic language created additional barriers for different researchers. Our aim in this paper is to design a deep convolutional neural network based architecture that could extract features and classify Amharic characters with significant confidence of accuracy that could be utilized for real-world applications. A total of 90,000 characters are prepared for training the proposed architecture and an additional of 25,000 characters are reserved for testing purpose. Due to the occurrence of a large number of symbols and a close resemblance in the shapes of the different characters available in the language, a relatively complex convolutional Neural Network is utilized to capture those features and categorize them into the correct characters. Dropout layers are utilized to avoid overfitting. The character recognition system proposed in this paper achieved an accuracy of 99.27% on the testing dataset which is a significant improvement for the Amharic language. The implementation was done using Tensorflow on Keras neural network layers and Opencv in python to pre-process image data which enables us to make the system readily available for software developers as an API.KeywordsAmharic recognitionCharacter recognitionDeep convolutional neural networksDeep learning application
Article
This study presents a novel ensemble classifier-based off-line handwritten word recognition system following a holistic approach. Here each handwritten word is recognised using two handcrafted features, namely (i) Arnold transform-based feature that addresses local directional feature which depends on the stroke orientation distribution of cursive word and (ii) oriented curvature-based feature which is basically the histogram of curvelet index and one machine generated feature using deep convolution neural network (DCNN). In this study, a new architecture of DCNN is proposed for handwritten word recognition. These features are recognised by three classifiers separately. Finally, the decision of three classifiers is combined to predict the ultimate word class level. To fuse the decision of individual classifiers, the authors have explored three strategies: (i) vote for strongest decision, (ii) vote for majority decision and (iii) vote for the sum of the decisions. The proposed handwritten word recognition system is tested on three handwritten word databases: (i) CENPARMI database, (ii) IAM database and (iii) ISIHWD database. The performance of the proposed system is promising and comparable to state-of-the-art handwriting recognition systems.
Chapter
This paper discusses a region based deterministic finite state automata for recognition of isolated online Teeline shorthand alphabet. The strokes are drawn in fixed template of size M × N that is M rows and N columns on a tablet or handheld devices, which are further divided into 13 logical regions using the concept of matrices. Features are extracted from each of these regions and are labeled 1 or 0, based on minimum number of pixels covered in that region and are stored in the feature vector. These feature vectors form input to deterministic finite state automata which recognizes the characters drawn. In order to test 54,600 online handwritten isolated Teeline character samples were used as there is no database of online Teeline characters is available. The characters are written five times by twenty writers. It is observed that average recognition rate is 93.57% for novice writers, 96.73% for familiar writers and 98.84% for expert writers, who practiced writing for a week. The method is simple and does not need segmentation of the strokes. Preprocessing is done at the level of regions and only those regions where the stroke passes.
Article
The main aim of this work is to compare Hindu and Arabic digits with respect to a machine’s ability to recognize them. This comparison is done on the raw representation (images) of the digits and on their features extracted using two feature selection methods. Three learning algorithms with different inductive biases were used in the comparison performed using the raw representation; two of them were also used to compare the digits using their extracted features. All classifiers gave better results for Hindu digits in both cases; when raw representation was used and when the selected features where used. The experiments also show that Hindu digits can be classified with better accuracy, higher confidence and using fewer features than Arabic digits. These results indicate that hand-written Hindu digits are actually easier to recognize than hand-written Arabic digits. The machine learning methods used in this work are instance based learning (the kNN algorithm), Naïve Bayesian and neural networks. The feature extraction methods we used were Fourier transformation and histograms.
Article
This paper proposes an algorithm that can construct a dynamically DAWG (directed acyclic word graph) by merging the common suffixes of the try, thus solving the problem of the increased number of transitions in the try. In the proposed method, the transition merge for the common suffixes in the DAWG is released locally. After modifying the transitions by additional or deletion of keys, the transitions are merged locally. By this scheme, the dynamic update of the DAWG can be realized in a reasonable time. The proposed algorithm is evaluated theoretically and the data structure for the implementation is discussed. Experimental results are presented for various key sets. It is demonstrated experimentally that the number of transitions in a DWAG is approximately 50 to 70% than for a try in the case of a key of several thousand to fifty thousand, and the keys can be updated in a reasonable time of the key set is less than ten thousand.
Conference Paper
In this paper we present the GRUHD database of Greek characters, text, digits, and other symbols in unconstrained handwriting mode. The database consists of 1,760 forms that contain 667,583 handwritten symbols and 102,692 words in total, written by 1,000 writers, 500 men and equal number of women. Special attention was paid in gathering data from writers of different age and educational level. The GRUHD database is accompanied by the GRUHD software that facilitates its installation and use and enables the user to extract and process the data from the forms selectively, depending on the application. The various types of possible installations make it appropriate for the training and validation of character recognition, character segmentation and text-dependent writer identification systems.
Article
In this paper we present two algorithms for building lexicons in Directed Acyclic Word-Graphs (DAWGs). The two algorithms, one for deterministic and the other for non-deterministic DAWGs, can be used instead of the traditional subset construction method. Although the proposed algorithms do not produce the optimal DAWG (i.e., the one with the minimum number of states), they are simple, fast and able to build the DAWG incrementally, as new words are added to the lexicon. Thus, building large lexicons in a DAWG structure becomes an easy task, even for a modest computer.
Article
An abstract is not available.
Article
SCOPE: Algorithms, Information storage and retrieval. A trie is a search tree obtained by merging the common suffixes of the key set. It has the advantage that all keys as prefixes of an input string can be retrieved with high speed. When the size of the key set is enlarged, however, a problem arises, as the number of transitions increases, so too does the need for a large storage capacity. This paper proposes an algorithm that dynamically constructs DAWGs (Directed Acyclic Word Graphs) for the handling of dynamic key sets. It also solves the problem of the increasing number of transitions in the trie structure. The proposed method constructs a DAWG through the local separation of common suffixes for updating a key and, after finishing updating a key, the local transition merge of common suffixes. The proposed algorithm is theoretically evaluated and the data structure for the implementation is discussed. Experimental results show that the number of transitions in the DAWG is reduced by approx. 50 to 70% compared of that of the trie, for key sets of several thousands of fifty thousand elements, also the updating of the keys can be executed in a practical time for sets of less than ten thousand keys.