Abstract
Word searching or keyword spotting is an important research problem in the domain of document image processing. The solution to the said problem for handwritten documents is more challenging than for printed ones. In this work, a two-stage word searching schema is introduced. In the first stage, all the irrelevant words with respect to a search word are filtered out from the document page image. This is carried out using a zonal feature vector, called pre-selection feature vector, along with a rule-based binary classification method. In the next step, a holistic word recognition paradigm is used to confirm a pre-selected word as search word. To accomplish this, a modified histogram of oriented gradients-based feature descriptor is combined with a topological feature vector. This method is experimented on a QUWI English database, which is freely available through the International Conference on Document Analysis and Recognition 2015 competition entitled “Writer Identification and Gender Classification.” This technique not only provides good retrieval performance in terms of recall, precision, and F-measure scores, but it also outperforms some state-of-the-art methods.
1 Introduction
Handwriting is an age-old and standard way of communication. Before the invention of modern technologies for communication, like printing press, typewriter, and email, communication was predominantly made via handwriting. Even in this era of advanced technology, a large number of people still prefer the ancient way of communication, which leads to an exponential growth of handwritten documents. Due to mismanagement, the quality of the document degrades with each passing day, which implies that preservation of documents is a pressing need. In addition, manual searching for some important information from a huge repository is time consuming and at times also error prone. Sometimes, due to lack of proper maintenance, some important document(s) may be misplaced or lost. All these issues are posing genuine problems for proper storage and maintenance of handwritten document images. Apart from these facts, advanced office automation demands mechanized storage, manipulation, and retrieval of documents in electronic format, i.e. handwritten documents need to be managed properly. The undebatable solution of this is to convert document images into an electronic form and then process the same with an optical character recognition (OCR) engine [4]. However, current handwritten OCR engines work poorly for large lexicon sizes [7]. Therefore, the alternative solution is to keep the documents in well-indexed digital form with appropriate tagging. One of the possible ways of tagging is keyword-based document indexing, which is an enriched research problem. It has a large number of real-life applications [13, 15, 25].
The pre-requisite for any keyword-based document indexing is identifying the keywords from documents (also known as word/keyword spotting in the literature). A word searching methodology tries to locate the occurrence(s) of some pre-defined words inside a document image. It is worthy to mention that searching a word in a handwritten document [35] poses more challenges than searching the word from its printed counterpart [23]. One of the key factors for this is the variation in writing styles, not only among different writers but also for the same writer. Even some typical problems related to handwriting, like skew, slant, overlapping, and/or joining of words/characters, add more complexity.
1.1 Categorization of Word Searching Methods
A typical way of categorizing word searching techniques is based on how query words are fed into the system. Depending on this, existing techniques are classified into two categories, namely query-by-example (QBE) [13, 15] and query-by-string (QBS) [11, 13]. In the former category, a word image is provided to the system and the system returns occurrence(s) of that given word in the document image. Methods of this approach have mostly relied on image matching techniques. On the other hand, the techniques that follow the QBS approach consider an arbitrary word as a string to be searched from a document image. Therefore, methods of this approach, in general, try to follow some word recognition model. Word searching techniques can also be categorized as segmentation based [38] and segmentation free [27]. Segmentation-based word spotting techniques have relied on pre-segmented text lines [12, 22, 38] or words [16, 22], which may be achieved by using text-line and/or word-level ground truth (GT) images. The other category of techniques [26, 27] tries to locate the words to be searched in document images without spending time for page segmentation. Use of word samples, collected in offline mode for preparing the training module, is found in recognition-based approaches [1]. On the other hand, recognition-free approaches [20, 21] try to spot search words using some matching techniques. The present work is an instance of a recognition-based approach that uses a holistic word recognition technique.
1.2 Literature Survey
From the literature survey, it is observed that, to date, many researchers have applied several distance-based methods [8, 22, 29, 36] for searching a word from a document image following the QBE way. In Ref. [22], in order to handle the word searching problem, gradient angle and its magnitude were extracted from word images to search query words using a disc-based matching schema. The work in Ref. [29] reports a word searching mechanism that uses scale-invariant feature transform (SIFT) features from word images, and search is confirmed by cosine and Euclidean distance measures. Whereas the work described in Ref. [8] extracted Gabor feature and then used Euclidean distance-based similarity check for word searching in the QBE way. Pixel value information was used in Ref. [36], where word searching in the QBE way was carried out by using a Bray-Curtis dissimilarity measure.
In another category of methods [20, 21, 30], several string matching approaches have been used for finding a solution to the said problem. The work described in Ref. [20] introduced a flexible sequence matching technique for comparing query word and target word. In this work, eight feature values, extracted from each column of the image, were used. The same set of feature values were used in Ref. [21] for investigating the effectiveness of some conventional time series matching techniques for word searching. A multi-angular feature descriptor was used in Ref. [30], where word images are represented by a variable length feature vector. Dynamic time warping (DTW) was used for matching the words.
Graph similarity-based methods [24, 37, 38] are also found in the literature for word spotting. Authors in Refs. [37, 38] used a graph edit distance-based similarity score for word spotting. In Ref. [38], each word image was represented as a sequence of skeleton-based graphs using context-labeled vertices for connected components (CCs). A similar approach is found in Ref. [37]. The only difference is the way the calculation of the graph edit distance is done, i.e. in Ref. [37] bipartite graph matching schema was used, while in Ref. [38] DTW alignment method was used. In Ref. [24], attributed graphs have been constructed using the graphemes that were extracted from words by a part-based approach. It employed an edit distance-based similarity measure.
Classical word recognition models were used in Refs. [11, 25, 36] for performing word spotting in QBS. A statistical framework for the word spotting problem, introduced in Ref. [25], has explored the use of two types of hidden Markov models (HMMs), namely continuous HMM and semi-continuous HMM. Geometrical features were extracted in Ref. [11] for performing word spotting. HMM-based character-level trained model was applied for finding the similarity score in this work. In Ref. [36], a pyramidal histogram of characters feature vector was extracted for performing word searching. The Bray-Curtis dissimilarity measure between the words was measured for performing word similarity tasks.
Researchers have also attempted to devise techniques [22; 26; 27; 28] where they segment a word into characters or character sub-parts. In Ref. [28], a patch-based segmentation-free framework was introduced. The framework used a bag-of-features (BoF)-based HMMs (BoF-HMM) model, whereas the authors of Ref. [27] used a SIFT feature descriptor for recognition purpose in their BoF-HMM model. An approach similar to that of Ref. [27] was proposed in Ref. [26]. In this case, the model was developed to search any arbitrary word and does not require any pre-segmentation of document pages. A semi-supervised handwritten word recognition-based technique using bi-directional long short-term (BLSTM) neural network (NN) model was presented in Ref. [22].
A holistic word recognition paradigm was used in Refs. [1, 14] for retrieval of search words from historical Arabic handwritten manuscripts. In Ref. [1], several structural and statistical features, extracted from connected parts of word images, are fed to a multi-layer perceptron (MLP)-based classifier for preparing a learning module. In this work, it was shown that features extracted from connected parts of the word perform well in comparison with features extracted from the entire word. The work reported in Ref. [14] introduced a hierarchical classifier, comprising support vector machine and regularized discriminant analysis for searching a word. A gradient-based feature vector is used for discriminating the words in feature space.
1.3 Word Recognition Methodologies
Word recognition is a process of realizing word images as a machine-editable form. Researches on word recognition, as found in the literature, follow either an analytical approach [4, 11, 27] or a holistic approach [6, 7, 18, 31]. In the analytical approach, a word image is represented as a collection of units known as characters and/or character sub-parts. These units are first recognized, and then the recognized units are compiled to form a machine-encoded word. It is certainly the best way of word recognition; however, it suffers from several segmentation ambiguities [7].
Consequently, researchers often use lexicon-driven character (or character-like shapes)-based model for word recognition, such as HMM [11, 25] and BLSTM-based NN [22]. These approaches allow us to obtain good performance when the learning set is well representative of the writing/font styles that are found in dataset. However, the main drawback of such approaches is the usual requirement of an enormous set of transcribed text line/word images for training. Such requirement could be costly and often needs manual effort. At times, this approach is designed focusing on some particular language. Apart from these, it takes a huge time to develop the learning module.
On contrary, in the holistic approach, a word is considered as a single and indivisible unit. This approach is, in general, segmentation free; hence, it can avoid the problems that occur while using a segmentation-based or analytical approach. As the features are extracted from the entire word image, this approach provides acceptable recognition accuracy irrespective of script [31]. Based on these facts, it can be concluded that a holistic approach can be a suitable choice for word matching in recognition-based word spotting schema, as the number of query words, in general, are pre-defined and smaller in size.
1.4 Motivation
Literature survey reveals that mainly two different approaches, viz. recognition free [22, 36] and recognition based [12, 28], are followed by researchers for providing solution to the word matching problem. The first category of works extracts geometrical/shape context characteristics of word image (e.g. [24, 29, 30, 36]) and then applies some similarity measure for word matching. Although this mechanism is fast, it generally retrieves more irrelevant words with respect to the actual search word. On the contrary, matching techniques as applied in recognition-based methods (e.g. [1, 11, 14, 27]) have tried to identify a query word in the document images by recognizing all the words present therein. Therefore, these mechanisms not only suffer from a high time requirement but also retrieve more irrelevant words.
Considering the above facts, in the present work, a two-stage word searching technique is designed. This work employs a pre-selection schema that uses the benefits of the recognition-free word spotting approach and a final matching schema that relies on the recognition-based model. Here, a holistic paradigm is used for matching a query word with the target word. The experiment has been conducted on a handwritten English document page database, called “QUWI” database [2], which was also used in the writer identification competition in the International Conference on Document Analysis and Recognition (ICDAR) 2015 [10].
2 Present Work
Any word searching mechanism, in general, comprises two major steps, namely (i) page segmentation (text lines and/or words are extracted from document images) and (ii) word matching schema (i.e. confirming a target word present in a document image as the word to be searched). A number of works, found in the literature, have used GT images containing either pre-segmented text line [12] or words [22] for localizing words within a document page. Such use of GT images helps in avoiding several inherent errors occurring due to page segmentation [17, 32, 34]. Inspired by this fact, here, document-level GT images containing pre-segmented words have been prepared using GTgen 1.1 [33] software. The details of document-level GT image preparation are described in Section 2.2.
A two-stage approach, consisting of pre-selection of target words as words to be searched and confirmation of pre-selected word(s) as search word, has been introduced for searching a word from document page images. A feature vector (say 𝔉1) of length 3 has been extracted from all the words present in a document image to filter out irrelevant words with respect to the word to be searched (refer to Section 2.3.1 for a detailed description). To confirm the remaining word(s) in the document page as expected search word(s), a holistic word recognition approach has been used. For this, a feature vector (say 𝔉2), comprising topological features and a modified histogram of oriented gradients (HOG) feature descriptor, is extracted from each word image, which is then classified using an MLP classifier. The word searching module is described in Section 2.3, while the overall working procedure of the proposed word searching technique is depicted in Figure 1.

Block Diagram of the Present Word Searching Technique.
2.1 Data Preparation
The pre-requisite for developing a word searching algorithm is to build a collection of handwritten document images written by different persons. In the present work, the QUWI database [2], which is partially public through different competitions, has been preferred. Such preference is significant because of its number and diversity (e.g. nationality, age, background, etc.) of the writers and variations in writing materials (i.e. colors and thicknesses of pen/pencil): 300 handwritten document images for each of the scripts in English and Arabic, written by 300 different writers [≈14.75% and ≈29.5% of entire document images, with respect to number of categorical (script) document images and writers, respectively] are uploaded for public use through the ICDAR 2015 competition on multi-script writer identification and gender classification [10]. These document pages contain the same text with 117 words (in most of the cases). It is to be noted that some of these document images contain fewer number of words (the minimum number of words, as observed, is 99). Moreover, wrongly spelled words, use of abbreviated forms, and different spellings from the original words definitely have added huge complexities while considering searching some words therein. The publicly available document images written in English are considered here. The document pages are divided in a 1:5 ratio. The first set is used for evaluation of the present word searching algorithm, while the rest of the document page images have been used for confirming values of different parameters required to design the searching algorithm.
2.2 GT Image Preparation
In the present work, instead of applying any word extraction technique, GT images representing ideal word boundaries are prepared. A block diagram representing the various stages of GT image generation for each document page is depicted in Figure 2. At first, document images are binarized using Otsu’s thresholding technique. Next, a morphological close operator using a 5 × 5 mask is applied on binarized document images. Finally, noise-free document images are generated by deleting CCs having ≤5 data pixels (see Figure 3A,B).

Block Diagram of GT Image Preparation.

Word Instances (A) Before and (B) After Deletion of Small Components.
To prepare a GT page image, all the words in it are colored using GTgen1.1 [33], a GT-generating tool. This tool can set uniform color for all data pixels (appearing as CCs or disconnected components) within a selected region. This feature helps in coloring words in a document page manually and takes less time than MS paint tool. An instance of GT images is shown in Figure 4.

An Instance of GT Image Where Words are Colored Differently.
Here, neighboring words belonging to same text line, or words belonging to successive text line (below or above), are painted with different colors.
2.3 Word Searching
It has already been mentioned that, here, a two-stage approach is designed for searching a given word in a document page. In the first stage, irrelevant words (with respect to a search word) from a document image are filtered out. The remaining words present in a document page are termed as pre-selected words with respect to given query words. Such selection process is shown in Figure 5A. In the second stage, the pre-selected words are confirmed as the given query image (see Figure 5B). This process is carried out using a holistic word recognition method. The details of these stages are given in the following subsections.

Schematic Diagrams.
(A) For pre-selecting probable candidate words in a document page that are relevant to a given search word and (B) for detecting instance(s) of the search word from its pre-selected candidate word set. Meaning of Notations: RC, Recognition confidence; TH, a predefined threshold value which is set here as 0.75. Other symbols are already described in Figure 1.
2.3.1 Pre-selection of Probable Candidate Words for a Particular Search Word
Words in a document, in general, have varying numbers of characters. In this scenario, searching a word directly in a document image increases the chances of mismatch as well as enhances the computation cost. Sometimes, it leads to selection of unnecessary words as search word. That is why, in the present work, words having a different number of characters than the number of characters in the search word are not considered during the search process. Rather, a rule-based method is followed for the pre-selection of probable candidate words.

Partitioning of a Word Image into Zones.
2.3.1.1 𝕱1 Extraction
Let a binarized word image be represented as
However, identification R2 and R3 is not straightforward. For identifying these lines, first, the number of horizontal transition points between data and non-data pixels or vice versa along each row of B is calculated. Such number along a horizontal row i (𝒯i, i ∈ [1, H]) is calculated by
The mean of all such transition point counts (μTP) of B is estimated by
Now, R2 and R3 are calculated as follows:
Finally, the said 𝔉1(= (f1, f2, f3)) is estimated by Eqs. (7)–(9):
In Eqs. (8) and (9), C represents CC. Now, let θ(.) and ∅(.) be functions that represent the belongingness of a CC in upper/lower zone, respectively, which are defined by
2.3.1.2 Decision Rule Formulation
To filter out words that are not relevant with respect to a given search word, here, a decision rule has been designed. For this, first, decision boundaries (lower and upper bounds) for each of the extracted feature values (i.e. f1, f2, and f3) are estimated. These decision boundaries are set by considering the mean (μfi, i = 1, 2, 3) and standard deviation (σfi, i = 1, 2, 3) of feature values extracted from manually chosen N samples for a search word. Let ℒfi and 𝒰fi be the lower and upper bounds of feature value fi (i = 1, 2, 3), respectively, which are defined by
Finally, a word is pre-classified as a probable candidate for the given search word by the decision rule as depicted in Figure 7.

Diagrammatic Representation of the Selection Process of a Word Image as a Probable Candidate Word Using a Rule-Based Mechanism.
2.3.2 Confirming a Pre-selected Candidate Word as Search Word
For recognition of the keywords, a feature vector (i.e. 𝔉2), comprising topological [18] and texture-based [3, 5] features, is extracted from both search and target word images.
2.3.2.1 Topological Feature
Topological features are extracted either from the entire word image (for extracting global information) or from sub-images (for extracting local information). For calculations of features from sub-images, word images are divided into an m × n grid.
Area: Areas covered by the different patterns (here, query and target words) vary a lot because of the presence of different numbers of characters and also for shape variations of these characters therein. The area of B is calculated by
Values of R1 and R4 are calculated using Eqs (1) and (2), respectively. Values of other parameters, i.e. C1 and C2, are estimated by
Area-based feature values are extracted from sub-images only. Therefore, the total number of area features (ℒA) extracted from an image can be represented by
Pixel density (PD): PD represents the number of data pixels per unit area. It is calculated by
PD is extracted from sub-images only. Therefore, a feature vector of dimension m*n is generated, i.e.
Here, ℒPD is the length of pixel density feature vector, extracted from an image.
Aspect ratio (AR): The AR of an image is defined as ratio of width to height that generally differs when word images contain different number of characters. So, this feature is also considered here and the feature value is calculated as
AR is extracted from the original image and sub-images. Therefore, the total number of ARs (ℒAR) extracted from a word image is represented by
Longest run: The longest run feature is considered here as another topological feature. This feature consists of two feature values, viz. lengths of the longest run along X and Y directions, respectively. Let B has Ni(≥1) number of runs (occurrence of continuous data pixels) along ith row and HLij represents the length of jth run along ith row. The horizontal longest run (HLR) is calculated as
The vertical longest run (VRL) feature is computed by identifying runs in a column-wise manner. Let B has Ni(≥1) number of runs along ith column and VLij represents the length of jth runs along ith column. Now, the VLR of B is measured as
These feature values are extracted from actual word image and sub-image structures. Therefore, a feature vector of length 2 ∗ (m ∗ n + 1) is generated, which means the total number of longest run features (ℒLR) is defined by
Centroid: The centroid feature is based on the center of gravity of an image. This position varies depending on the shape of the pattern. Centroid coordinates (CX, CY) are calculated as
These two features are extracted (in a global way) from an entire image and from upper and lower parts of an image, which are separated by principal and non-principal diagonals (see Figure 8A,B). Therefore, the number of features generated from an image is 2 ∗ (1 + 2 + 2) = 10. Also, all these features are again extracted from all the sub-images for obtaining local information. Therefore, the total number of centroid features (ℒC) can be formulated as

Showing Hypothetically Generated Sub-images by Segmenting an Image Along (A) Principal and (B) Non-principal Diagonals.
Projection length: Projection lengths of word image on the principal axes, i.e. on X and Y axes, are considered here as feature values. Let ℛ and 𝒞 be sets of row index and column index, respectively, that contain at least 1 data pixel and are defined as
Now the horizontal projection length (HPL) and vertical projection length (VPL) are defined as
These two features are extracted from the entire image and from upper and lower parts of the image, which are separated by principal and non-principal diagonals (see Figure 8A,B). Therefore, using this projection length-based concept, the total number of global features generated from such images is 2 ∗ (1 + 2 + 2) = 10, i.e. the total number of projection length features (ℒPL) can be written as
By accumulating all of the above-mentioned features, the length of the topological feature becomes
2.3.2.2 Texture-Based Feature
HOG [9] is a texture-based feature descriptor that was proven to be useful in many pattern recognition problems. It firstly computes gradient information of a cell (a primitive sub-block) in different directions and, secondly, normalization of cell information is done for each block to describe a pattern. The orientation measurement on each cell is performed by dividing the total orientation angle (0°–360° or 0°–180°) into different ranges (commonly it is 8, 9, 16, etc.). Each of these angle ranges is called a “bin.” It needs a size-normalized image to obtain the fixed length of a feature vector. Such resizing not only strives to compromise with resolution but also destroys the AR of images, which is a discriminative feature while considering the word recognition problem.
The HOG feature, as described in Ref. [19], is extracted into b bins from an image by considering c × c cell size and 2c × 2c block size. Before extracting this feature, images are padded with zeros to convert their dimension in multiples of c. Next, the mean and standard deviation of the extracted information in each of the b bins are considered here as feature values. This implies that a feature vector of length 2 ∗ b is generated for every image. This modification not only helps to generate a feature vector of equal length by preserving the actual size of the image but is also capable of generating feature vectors of lesser dimension. Here, to get local information of a word image, these HOG feature values are again extracted from sub-images, generated by diving an image into a p × q grid. Therefore, the total length of HOG feature vector (ℒH) is formulated as
Therefore, the length of 𝔉2 (ℒ) is calculated as below:
Inspired by the work in Ref. [19], the value of b is set as 9. Also, considering the work as reported in Ref. [18], values of m and n are kept the same, i.e. m = n. Therefore, Eq. (35) can be written as
The values of other parameters (i.e. m, p, and q) are decided experimentally.
2.3.2.3 Confirmation Process
For identifying the search word from pre-selected candidate words, 𝔉2 is extracted from the manually chosen N samples of all the search words. Next, these features are fed to an MLP-based classifier, which returns a learned module. Next, all the pre-selected candidate words for a given search word are passed through this learned module. The system provides a recognition confidence (say, RC) for all of these words. Finally, a pre-selected candidate word is confirmed as the said search word if RC ≥ TH. TH is a pre-defined threshold value, which is here chosen as 0.75.
3 Results and Discussion
In the present work, a process of word searching from handwritten document images is reported. For this purpose, a two-stage approach, discussed earlier, is introduced. In the following subsection, issues related to the experimental outcome are described.
3.1 Selection of Search Word Set
It has already been mentioned that the QUWI database has been chosen for carrying out the experiment. For evaluation of the designed word searching technique, a set of 15 words is selected from document images as query words (see Table 1). The total number of occurrences of these words in a document page is 25 (around 21% of the total words in an ideal document page). The choice of such words to be searched from documents is mainly based on multiple occurrences of words in the document pages. Along with this, word pairs that are almost indistinguishable in terms of shape (e.g. “today” and “today’s”), stemming words (e.g. “Asia” and “Asian” or “migrant” and “migrants”), derived forms (e.g. “large” and “largest”), and arbitrary words (e.g. “international” and “nations”) are also included here in searching word set. Use of uppercase or lowercase letters at the beginning (in a few cases in between) of the same search word is also found in the document pages. All these cases add more complexities while searching words from handwritten document images.
Instances of Search Words.
| SW ID | Sample Image | WF | SW ID | Sample Image | WF | SW ID | Sample Image | WF |
|---|---|---|---|---|---|---|---|---|
| SW01 | ![]() |
2 | SW06 | ![]() |
2 | SW11 | ![]() |
2 |
| SW02 | ![]() |
2 | SW07 | ![]() |
1 | SW12 | ![]() |
5 |
| SW03 | ![]() |
1 | SW08 | ![]() |
1 | SW13 | ![]() |
1 |
| SW04 | ![]() |
2 | SW09 | ![]() |
1 | SW14 | ![]() |
1 |
| SW05 | ![]() |
2 | SW10 | ![]() |
1 | SW15 | ![]() |
1 |
WF represents the number of occurrences of the corresponding word in a document page, and SW ID stands for search word index.
3.2 Training Sample Preparation
As mentioned earlier, 250 document pages out of 300 publicly available pages of the QUWI database are kept aside for confirming different parameters that are required to build the present word searching method. A total of 250 word images, per search word, are randomly extracted from 250 document pages that are used for the evaluation of the present technique. Therefore, a database containing 15 ∗ 250 = 3750 word images is prepared. This word set helps in selecting optimal parameter values, used for selection of the probable candidate search words and finalizing the search word from those probable candidate search words.
3.3 Parameters for Detecting Pre-selected Candidate Words
The feature values f1, f2, and f3 [see Eqs. (7)–(9)] are first extracted from each of the word images, and then the bounds of the same [see Eqs. (12)–(13)] are also calculated. These bound values for each of the query word images are reported in Table 2. Using these values and selection process, described in Figure 7, the irrelevant target words with respect to a search word are removed successfully. It has been observed that around 20% of the words in a document page are passed onto the next stage.
Search Word-Wise Lower and Upper Bound Values of f1, f2, and f3 of 𝔉1.
| SW ID | 𝓛fi |
Ufi |
||||
|---|---|---|---|---|---|---|
| i = 1 | i = 2 | i = 3 | i = 1 | i = 2 | i = 3 | |
| 01 | 17.05 | 1 | 0 | 22.65 | 3 | 0 |
| 02 | 9.43 | 1 | 0 | 11.84 | 3 | 0 |
| 03 | 12.07 | 1 | 0 | 16.04 | 3 | 0 |
| 04 | 13.45 | 0 | 0 | 17.77 | 2 | 2 |
| 05 | 22.17 | 2 | 0 | 30.10 | 4 | 2 |
| 06 | 22.56 | 2 | 0 | 32.20 | 6 | 2 |
| 07 | 26.62 | 4 | 0 | 36.22 | 6 | 0 |
| 08 | 10.76 | 0 | 0 | 14.55 | 2 | 2 |
| 09 | 14.08 | 1 | 0 | 18.47 | 3 | 2 |
| 10 | 16.20 | 1 | 0 | 22.52 | 3 | 2 |
| 11 | 18.03 | 1 | 0 | 24.27 | 3 | 2 |
| 12 | 15.94 | 3 | 0 | 20.90 | 5 | 0 |
| 13 | 16.78 | 2 | 0 | 21.95 | 4 | 0 |
| 14 | 11.70 | 1 | 1 | 16.54 | 3 | 1 |
| 15 | 13.10 | 2 | 0 | 18.03 | 4 | 2 |
SW ID indicates search word index.
3.4 Parameter Optimization for Holistic Word Recognition
Selection of optimal parameter values for the holistic word recognition [i.e. as m, p, and q in Eq. (36)] is carried out here. This selection is performed based on recognition accuracy. For this purpose, a five-fold cross-validation schema is carried out using the said isolated word samples. Here, values of m and p are varied between 2 and 4. Such choice preserves the nature of zonal variation inside a word image. Whereas, the value of q is varied between 4 and 7. It is done to include more local information of the word images. Therefore, this experimental structure leads to 3 × 3 × 4 = 36 number of experiments. All such experiments are conducted and the best recognition accuracies in five-fold cross validations are found to vary between 90.17% and 96.21% (see Table 3). In Table 3, the results of 10 such experiments are summarized that include the boundary values of the parameters. From this table, it is clear that the best recognition accuracy is achieved when values of m, p, and q are chosen as 2, 3, and 6, respectively. The learned module that provides the best recognition accuracy is used for confirming a pre-selected candidate word as search word (refer to Section 2.3.2.3).
Top 10 Recognition Results with Different Sets of Parameter Values Applied in a Five-Fold Cross-validation Mechanism.
| Experiment No. | Parameter Value |
Feature Dimension | No. of Neurons in the Hidden Layer | Recognition Accuracy (in %) | ||
|---|---|---|---|---|---|---|
| m | p | q | ||||
| 1 | 2 | 2 | 4 | 245 | 60 | 91.54 |
| 2 | 2 | 2 | 5 | 281 | 60 | 90.74 |
| 6 | 2 | 3 | 5 | 371 | 65 | 93.77 |
| 7 | 2 | 3 | 6 | 425 | 65 | 96.21 |
| 8 | 2 | 3 | 7 | 479 | 70 | 93.74 |
| 19 | 3 | 3 | 6 | 500 | 70 | 93.17 |
| 20 | 3 | 3 | 7 | 554 | 70 | 93.47 |
| 30 | 4 | 3 | 5 | 555 | 70 | 94.12 |
| 31 | 4 | 3 | 6 | 605 | 75 | 94.21 |
| 36 | 4 | 4 | 7 | 785 | 75 | 90.21 |
3.5 Evaluation Metrics
The assessment of the present word searching mechanism is carried out in terms of recall, precision, and F-measure scores [18]. Let 𝒲S and 𝒲R represent the set of valid instances of a search word inside a document page image and the set of word(s) retrieved after applying the present method on said document image, respectively. Recall and precision and F-measure scores can be formulated as
Let,
In Eqs. (40)–(42), N = 50, which is the number of document pages considered for evaluation. Search word-wise recall, precision, and F-measure scores are summarized in Table 4. From this table, it is clear that the present work has achieved reasonably good retrieval accuracy. This table also reveals that the best recall is achieved while searching the word “million” and overall satisfactory recall value (lowest recall value =0.68) indicates better retrieving capability of the present work. Whereas, lower precision values (≤0.5) are observed for the search words “Asia,” “Asian,” “today,” and “Nations.” The reason behind such lower precision rate is the similar 𝔉1 information with other words in the document page, which, in turn, retrieves more words in the pre-selection stage.
Search Word-Wise Recall, Precision, and F-measure Values of Present Word Searching Technique.
| SW ID (i) | Recall (𝓡i) | Precision (𝓟i) | F-measure (𝓕i) |
|---|---|---|---|
| 1 | 0.8462 | 0.7586 | 0.8000 |
| 2 | 0.8846 | 0.3108 | 0.4600 |
| 3 | 0.9167 | 0.3929 | 0.5500 |
| 4 | 0.7200 | 0.9205 | 0.8000 |
| 5 | 0.6800 | 0.8095 | 0.7391 |
| 6 | 0.7692 | 0.7692 | 0.7692 |
| 7 | 0.8462 | 0.8462 | 0.8462 |
| 8 | 0.8462 | 0.9167 | 0.8800 |
| 9 | 0.8333 | 0.9091 | 0.8696 |
| 10 | 0.9167 | 0.7333 | 0.8148 |
| 11 | 0.7600 | 0.9830 | 0.8636 |
| 12 | 0.9194 | 0.9344 | 0.9268 |
| 13 | 0.9000 | 0.3600 | 0.5143 |
| 14 | 0.9091 | 0.3125 | 0.4651 |
| 15 | 0.6923 | 0.9000 | 0.7826 |
| Average | 0.8349 | 0.7235 | 0.7388 |
SW ID stands for search word index. Bold numbers indicate the best scores.
3.6 Comparison with State-of-the-Art Methods
The performance of the present technique is compared with some state-of-the-art methods [36, 37]. Among these, the method reported in Ref. [36] is recognition based and another two methods [30, 37] are recognition-free word searching methods. It is to be noted that the authors of Ref. [37] compared several variations of time series matching techniques. However, here, the popularly used DTW method is used. The average performances of the said state-of-the-art methods along with the present one are reported in Table 5. The search word-wise performances of these techniques are shown in Figure 9A–C. From the results, it is clear that the present technique performs better than the other methods.
Comparison of the Present Method with Some State-of-the-Art Word Searching Techniques.
| Methods with Year of Publication | Feature Extracted from | Average |
||
|---|---|---|---|---|
| Recall | Precision | F-measure | ||
| M1: Al Aghbari and Brook [1], 2009 | Entire word | 0.7090 | 0.4994 | 0.5638 |
| M2: Al Aghbari and Brook [1], 2009 | Connected parts of word image | 0.6777 | 0.5460 | 0.5694 |
| M3: Mondal et al. [20], 2016 | Each column of word image | 0.7045 | 0.0504 | 0.0901 |
| M4: Mondal et al. [21], 2018 | Each column of word image | 0.5721 | 0.06324 | 0.1057 |
| M5: Present study | 0.8349 | 0.7235 | 0.7388 | |
Bold numbers indicate the best scores.

Search Word-Wise Evaluation Metrics of Different Methods. Refer to Table 5 for the notations M1, M2, … , M5. (A) Recall, (B) Precision, and (C) F-measure.
4 Conclusion
A two-stage approach toward solving the problem of word spotting in a handwritten document page is reported here. In the pre-selection stage, irrelevant words in terms of number of characters present in the word image are removed. In the confirmation stage, pre-selected candidate words from a document image are passed through a holistic word recognition-based system. Based on recognition, confidence words are tagged as search words. The experimental results indicate that the proposed two-stage approach yields satisfactory retrieval performance. In spite of this success, there is still some room for improvement. First and foremost, a high-dimensional feature vector (dimension as reported in the best case recognition result is 425) is used for word recognition purpose. Therefore, application of a feature selection algorithm (either wrapper or filter) would be a good choice to improve the retrieval performance of the proposed approach. Also, some more context-sensitive features in the first stage of the current work could be applied in the future.
Acknowledgments
This work was partially supported by the CMATER research laboratory of the Computer Science and Engineering Department, Jadavpur University, India, and PURSE-II and UPE-II Jadavpur University projects. Dr. Ram Sarkar is partially funded by a DST grant (EMR/2016/007213).
Bibliography
[1] Z. Al Aghbari and S. Brook, HAH manuscripts: a holistic paradigm for classifying and retrieving historical Arabic handwritten documents, Expert Syst. Appl. 36 (2009), 10942–10951.10.1016/j.eswa.2009.02.024Search in Google Scholar
[2] S. Al Maadeed, W. Ayouby, A. Hassaïne and J. M. Aljaam, QUWI: An Arabic and English handwriting dataset for offline writer identification, in: Proceedings of International Conference on Frontiers in Handwriting Recognition, pp. 746–751, IEEE, 2012.10.1109/ICFHR.2012.256Search in Google Scholar
[3] S. Barua, S. Malakar, S. Bhowmik, R. Sarkar and M. Nasipuri, Bangla handwritten city name recognition using gradient-based feature, in: Proceedings of International Conference on Frontiers in Intelligent Computing: Theory and Applications, pp. 343–352, Springer, 2017.10.1007/978-981-10-3153-3_34Search in Google Scholar
[4] S. Basu, N. Das, R. Sarkar, M. Kundu, M. Nasipuri and D. K. Basu, A hierarchical approach to recognition of handwritten Bangla characters, Pattern Recogn. 42 (2009), 1467–1484.10.1016/j.patcog.2009.01.008Search in Google Scholar
[5] S. Bhowmik, M. G. Roushan, R. Sarkar, M. Nasipuri, S. Polley and S. Malakar, Handwritten Bangla word recognition using HOG descriptor, in: Proceedings of Fourth International Conference of Emerging Applications of Information Technology, pp. 193–197, IEEE, 2014.10.1109/EAIT.2014.43Search in Google Scholar
[6] S. Bhowmik, S. Polley, M. G. Roushan, S. Malakar, R. Sarkar and M. Nasipuri, A holistic word recognition technique for handwritten Bangla words, Int. J. Appl. Pattern Recogn. 2 (2015), 142–159.10.1504/IJAPR.2015.069539Search in Google Scholar
[7] S. Bhowmik, S. Malakar, R. Sarkar, S. Basu, M. Kundu and M. Nasipuri, Off-line Bangla handwritten word recognition: a holistic approach, Neural Comput. Appl. (2018), 1–16 (in press).10.1007/s00521-018-3389-1Search in Google Scholar
[8] S. Cao and V. Govindaraju, Template-free word spotting in low-quality manuscripts, in: Proceedings of International Conference on Advances in Pattern Recognition, pp. 45–53, 2007.10.1142/9789812772381_0021Search in Google Scholar
[9] N. Dalal and B. Triggs, Histograms of oriented gradients for human detection, in: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1, pp. 886–893, IEEE, 2005.10.1109/CVPR.2005.177Search in Google Scholar
[10] C. Djeddi, S. Al-Maadeed, A. Gattal, I. Siddiqi, L. Souici-Meslati and H. El Abed, ICDAR2015 competition on multi-script writer identification and gender classification using ‘QUWI’ database, in: Proceedings of International Conference on Document Analysis and Recognition, pp. 1191–1195, IEEE, 2015.10.1109/ICDAR.2015.7333949Search in Google Scholar
[11] A. Fischer, A. Keller, V. Frinken and H. Bunke, Lexicon-free handwritten word spotting using character HMMs, Pattern Recogn. Lett. 33 (2012), 934–942.10.1016/j.patrec.2011.09.009Search in Google Scholar
[12] V. Frinken, A. Fischer, M. Baumgartner and H. Bunke, Keyword spotting for self-training of BLSTM NN-based handwriting recognition systems, Pattern Recogn. 47 (2014), 1073–1082.10.1016/j.patcog.2013.06.030Search in Google Scholar
[13] A. P. Giotis, G. Sfikas, B. Gatos and C. Nikou, A survey of document image word spotting techniques, Pattern Recogn. 68 (2017), 310–332.10.1016/j.patcog.2017.02.023Search in Google Scholar
[14] M. Khayyat, L. Lam and C. Y. Suen, Learning-based word spotting system for Arabic handwritten documents, Pattern Recogn. 47 (2014), 1021–1030.10.1016/j.patcog.2013.08.014Search in Google Scholar
[15] K. Khurshid, C. Faure and N. Vincent, A novel approach for word spotting using merge-split edit distance, in: Computer Analysis of Images and Patterns, pp. 213–220, Springer, 2009.10.1007/978-3-642-03767-2_26Search in Google Scholar
[16] Y. Liang, M. Fairhurst and R. Guest, A synthesised word approach to word retrieval in handwritten documents, Pattern Recogn. 45 (2012), 4225–4236.10.1016/j.patcog.2012.05.024Search in Google Scholar
[17] S. Malakar, P. Ghosh, R. Sarkar, N. Das, S. Basu and M. Nasipuri, An improved offline handwritten character segmentation algorithm for Bangla script, in: Proceedings of Indian International Conference on Artificial Intelligence, pp. 71–90, 2011.Search in Google Scholar
[18] S. Malakar, P. Sharma, P. K. Singh, M. Das, R. Sarkar and M. Nasipuri, A holistic approach for handwritten Hindi word recognition, Int. J. Comput. Vis. Image Process. 7 (2017), 59–78.10.4018/IJCVIP.2017010104Search in Google Scholar
[19] S. Mallick, Histogram of oriented gradients, Available at http://www.learnopencv.com/histogram-of-oriented-gradients/, accessed 31 March, 2018.Search in Google Scholar
[20] T. Mondal, N. Ragot, J. Y. Ramel and U. Pal, Flexible sequence matching technique: an effective learning-free approach for word spotting, Pattern Recogn. 60 (2016), 596–612.10.1016/j.patcog.2016.05.011Search in Google Scholar
[21] T. Mondal, N. Ragot, J. Y. Ramel and U. Pal, Comparative study of conventional time series matching techniques for word spotting, Pattern Recogn. 73 (2018), 47–64.10.1016/j.patcog.2017.07.011Search in Google Scholar
[22] W. Pantke, M. Dennhardt, D. Fecker, V. Margner and T. Fingscheidt, An historical handwritten Arabic dataset for segmentation-free word spotting – HADARA80P, in: Proceedings of International Conference on Frontiers in Handwriting Recognition, pp. 15–20, IEEE, 2014.10.1109/ICFHR.2014.11Search in Google Scholar
[23] R. Pintus, Y. Yang, E. Gobbetti and H. Rushmeier, An automatic word-spotting framework for medieval manuscripts, in: Digital Heritage, vol. 2, pp. 5–12, 2015.10.1109/DigitalHeritage.2015.7419446Search in Google Scholar
[24] P. Riba, J. Lladãs and A. Fornés, Handwritten word spotting by inexact matching of grapheme graphs, in: Proceedings of International Conference on Document Analysis and Recognition, pp. 781–785, IEEE, 2015.10.1109/ICDAR.2015.7333868Search in Google Scholar
[25] J. A. Rodriguez-Serrano and F. Perronnin, Handwritten word-spotting using hidden Markov models and universal vocabularies, Pattern Recogn. 42 (2009), 2106–2116.10.1016/j.patcog.2009.02.005Search in Google Scholar
[26] L. Rothacker and G. A. Fink, Segmentation-free query-by-string word spotting with bag-of-features HMMs, in: Proceedings of the 13th International Conference on Document Analysis and Recognition, pp. 661–665, 2015.10.1109/ICDAR.2015.7333844Search in Google Scholar
[27] L. Rothacker, M. Rusiñol and G. A. Fink, Bag-of-features HMMs for segmentation-free word spotting in handwritten documents, in: Proceedings of the International Conference on Document Analysis and Recognition, pp. 1305–1309, 2013.10.1109/ICDAR.2013.264Search in Google Scholar
[28] P. Roy, J. Ramel and N. Ragot, Word retrieval in historical document using character-primitives, in: Proceedings of International Conference on Document Analysis and Recognition, pp. 678–682, 2011.10.1109/ICDAR.2011.142Search in Google Scholar
[29] M. Rusinol, D. Aldavert, R. Toledo and J. Llados, Efficient segmentation-free keyword spotting in historical document collections, Pattern Recogn. 48 (2015), 545–555.10.1016/j.patcog.2014.08.021Search in Google Scholar
[30] R. Saabni and A. Bronstein, Fast keyword searching using boostmap based embedding, in: International Conference on Frontiers in Handwriting Recognition, pp. 734–739, IEEE, 2012.10.1109/ICFHR.2012.204Search in Google Scholar
[31] S. Sahoo, S. Nandi, S. Barua, P. Priyam, S. Bhowmik, S. Malakar and R. Sarkar, Handwritten Bangla word recognition using negative refraction based shape transformation, J. Intell. Fuzzy Syst. Appl. Eng. Technol. (2018) (in press).10.3233/JIFS-169712Search in Google Scholar
[32] R. Sarkar, S. Malakar, N. Das, S. Basu, M. Kundu and M. Nasipuri, Word extraction and character segmentation from text lines of unconstrained handwritten Bangla document images, J. Intell. Syst. 20 (2011), 227–260.10.1515/jisys.2011.013Search in Google Scholar
[33] R. Sarkar, N. Das, S. Basu, M. Kundu, M. Nasipuri and D. K. Basu, CMATERdb1: a database of unconstrained handwritten Bangla and Bangla-English mixed script document image, Int. J. Doc. Anal. Recogn. 15 (2012), 71–83.10.1007/s10032-011-0148-6Search in Google Scholar
[34] R. Sarkar, S. Halder, S. Malakar, N. Das, S. Basu and M. Nasipuri, Text line extraction from handwritten document pages based on line contour estimation, in: Proceedings of International Conference on Computing Communication & Networking Technologies, pp. 1–8, IEEE, 2012.10.1109/ICCCNT.2012.6395873Search in Google Scholar
[35] S. N. Srihari, C. Huang and H. Srinivasan, Search engine for handwritten documents, in: Proceedings of International Society for Optics and Photonics Electronic Imaging, pp. 66–75, 2005.10.1117/12.585883Search in Google Scholar
[36] S. Sudholt and G. A. Fink, PHOCNet: a deep convolutional neural network for word spotting in handwritten documents, in: Proceedings International Conference on Frontiers in Handwriting Recognition, pp. 277–282, 2016.10.1109/ICFHR.2016.0060Search in Google Scholar
[37] P. Wang, V. Eglin, C. Garcia, C. Largeron, J. Llados and A. Fornes, A novel learning-free word spotting approach based on graph representation, in: Proceedings of International Workshop on Document Analysis Systems, pp. 207–211, IEEE, 2014.10.1109/DAS.2014.46Search in Google Scholar
[38] P. Wang, V. Eglin, C. Garcia, C. Largeron, J. Llados and A. Fornes. A coarse-to-fine word spotting approach for historical handwritten documents based on graph embedding and graph edit distance, in: Proceedings of International Conference on Pattern Recognition, pp. 3074–3079, 2014.10.1109/ICPR.2014.530Search in Google Scholar
©2020 Walter de Gruyter GmbH, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 Public License.
Articles in the same Issue
- An Optimized K-Harmonic Means Algorithm Combined with Modified Particle Swarm Optimization and Cuckoo Search Algorithm
- Texture Feature Extraction Using Intuitionistic Fuzzy Local Binary Pattern
- Leaf Disease Segmentation From Agricultural Images via Hybridization of Active Contour Model and OFA
- Deadline Constrained Task Scheduling Method Using a Combination of Center-Based Genetic Algorithm and Group Search Optimization
- Efficient Classification of DDoS Attacks Using an Ensemble Feature Selection Algorithm
- Distributed Multi-agent Bidding-Based Approach for the Collaborative Mapping of Unknown Indoor Environments by a Homogeneous Mobile Robot Team
- An Efficient Technique for Three-Dimensional Image Visualization Through Two-Dimensional Images for Medical Data
- Combined Multi-Agent Method to Control Inter-Department Common Events Collision for University Courses Timetabling
- An Improved Particle Swarm Optimization Algorithm for Global Multidimensional Optimization
- A Kernel Probabilistic Model for Semi-supervised Co-clustering Ensemble
- Pythagorean Hesitant Fuzzy Information Aggregation and Their Application to Multi-Attribute Group Decision-Making Problems
- Using an Efficient Optimal Classifier for Soil Classification in Spatial Data Mining Over Big Data
- A Bayesian Multiresolution Approach for Noise Removal in Medical Magnetic Resonance Images
- Gbest-Guided Artificial Bee Colony Optimization Algorithm-Based Optimal Incorporation of Shunt Capacitors in Distribution Networks under Load Growth
- Graded Soft Expert Set as a Generalization of Hesitant Fuzzy Set
- Universal Liver Extraction Algorithm: An Improved Chan–Vese Model
- Software Effort Estimation Using Modified Fuzzy C Means Clustering and Hybrid ABC-MCS Optimization in Neural Network
- Handwritten Indic Script Recognition Based on the Dempster–Shafer Theory of Evidence
- An Integrated Intuitionistic Fuzzy AHP and TOPSIS Approach to Evaluation of Outsource Manufacturers
- Automatically Assess Day Similarity Using Visual Lifelogs
- A Novel Bio-Inspired Algorithm Based on Social Spiders for Improving Performance and Efficiency of Data Clustering
- Discriminative Training Using Noise Robust Integrated Features and Refined HMM Modeling
- Self-Adaptive Mussels Wandering Optimization Algorithm with Application for Artificial Neural Network Training
- A Framework for Image Alignment of TerraSAR-X Images Using Fractional Derivatives and View Synthesis Approach
- Intelligent Systems for Structural Damage Assessment
- Some Interval-Valued Pythagorean Fuzzy Einstein Weighted Averaging Aggregation Operators and Their Application to Group Decision Making
- Fuzzy Adaptive Genetic Algorithm for Improving the Solution of Industrial Optimization Problems
- Approach to Multiple Attribute Group Decision Making Based on Hesitant Fuzzy Linguistic Aggregation Operators
- Cubic Ordered Weighted Distance Operator and Application in Group Decision-Making
- Fault Signal Recognition in Power Distribution System using Deep Belief Network
- Selector: PSO as Model Selector for Dual-Stage Diabetes Network
- Oppositional Gravitational Search Algorithm and Artificial Neural Network-based Classification of Kidney Images
- Improving Image Search through MKFCM Clustering Strategy-Based Re-ranking Measure
- Sparse Decomposition Technique for Segmentation and Compression of Compound Images
- Automatic Genetic Fuzzy c-Means
- Harmony Search Algorithm for Patient Admission Scheduling Problem
- Speech Signal Compression Algorithm Based on the JPEG Technique
- i-Vector-Based Speaker Verification on Limited Data Using Fusion Techniques
- Prediction of User Future Request Utilizing the Combination of Both ANN and FCM in Web Page Recommendation
- Presentation of ACT/R-RBF Hybrid Architecture to Develop Decision Making in Continuous and Non-continuous Data
- An Overview of Segmentation Algorithms for the Analysis of Anomalies on Medical Images
- Blind Restoration Algorithm Using Residual Measures for Motion-Blurred Noisy Images
- Extreme Learning Machine for Credit Risk Analysis
- A Genetic Algorithm Approach for Group Recommender System Based on Partial Rankings
- Improvements in Spoken Query System to Access the Agricultural Commodity Prices and Weather Information in Kannada Language/Dialects
- A One-Pass Approach for Slope and Slant Estimation of Tri-Script Handwritten Words
- Secure Communication through MultiAgent System-Based Diabetes Diagnosing and Classification
- Development of a Two-Stage Segmentation-Based Word Searching Method for Handwritten Document Images
- Pythagorean Fuzzy Einstein Hybrid Averaging Aggregation Operator and its Application to Multiple-Attribute Group Decision Making
- Ensembles of Text and Time-Series Models for Automatic Generation of Financial Trading Signals from Social Media Content
- A Flame Detection Method Based on Novel Gradient Features
- Modeling and Optimization of a Liquid Flow Process using an Artificial Neural Network-Based Flower Pollination Algorithm
- Spectral Graph-based Features for Recognition of Handwritten Characters: A Case Study on Handwritten Devanagari Numerals
- A Grey Wolf Optimizer for Text Document Clustering
- Classification of Masses in Digital Mammograms Using the Genetic Ensemble Method
- A Hybrid Grey Wolf Optimiser Algorithm for Solving Time Series Classification Problems
- Gray Method for Multiple Attribute Decision Making with Incomplete Weight Information under the Pythagorean Fuzzy Setting
- Multi-Agent System Based on the Extreme Learning Machine and Fuzzy Control for Intelligent Energy Management in Microgrid
- Deep CNN Combined With Relevance Feedback for Trademark Image Retrieval
- Cognitively Motivated Query Abstraction Model Based on Associative Root-Pattern Networks
- Improved Adaptive Neuro-Fuzzy Inference System Using Gray Wolf Optimization: A Case Study in Predicting Biochar Yield
- Predict Forex Trend via Convolutional Neural Networks
- Optimizing Integrated Features for Hindi Automatic Speech Recognition System
- A Novel Weakest t-norm based Fuzzy Fault Tree Analysis Through Qualitative Data Processing and Its Application in System Reliability Evaluation
- FCNB: Fuzzy Correlative Naive Bayes Classifier with MapReduce Framework for Big Data Classification
- A Modified Jaya Algorithm for Mixed-Variable Optimization Problems
- An Improved Robust Fuzzy Algorithm for Unsupervised Learning
- Hybridizing the Cuckoo Search Algorithm with Different Mutation Operators for Numerical Optimization Problems
- An Efficient Lossless ROI Image Compression Using Wavelet-Based Modified Region Growing Algorithm
- Predicting Automatic Trigger Speed for Vehicle-Activated Signs
- Group Recommender Systems – An Evolutionary Approach Based on Multi-expert System for Consensus
- Enriching Documents by Linking Salient Entities and Lexical-Semantic Expansion
- A New Feature Selection Method for Sentiment Analysis in Short Text
- Optimizing Software Modularity with Minimum Possible Variations
- Optimizing the Self-Organizing Team Size Using a Genetic Algorithm in Agile Practices
- Aspect-Oriented Sentiment Analysis: A Topic Modeling-Powered Approach
- Feature Pair Index Graph for Clustering
- Tangramob: An Agent-Based Simulation Framework for Validating Urban Smart Mobility Solutions
- A New Algorithm Based on Magic Square and a Novel Chaotic System for Image Encryption
- Video Steganography Using Knight Tour Algorithm and LSB Method for Encrypted Data
- Clay-Based Brick Porosity Estimation Using Image Processing Techniques
- AGCS Technique to Improve the Performance of Neural Networks
- A Color Image Encryption Technique Based on Bit-Level Permutation and Alternate Logistic Maps
- A Hybrid of Deep CNN and Bidirectional LSTM for Automatic Speech Recognition
- Database Creation and Dialect-Wise Comparative Analysis of Prosodic Features for Punjabi Language
- Trapezoidal Linguistic Cubic Fuzzy TOPSIS Method and Application in a Group Decision Making Program
- Histopathological Image Segmentation Using Modified Kernel-Based Fuzzy C-Means and Edge Bridge and Fill Technique
- Proximal Support Vector Machine-Based Hybrid Approach for Edge Detection in Noisy Images
- Early Detection of Parkinson’s Disease by Using SPECT Imaging and Biomarkers
- Image Compression Based on Block SVD Power Method
- Noise Reduction Using Modified Wiener Filter in Digital Hearing Aid for Speech Signal Enhancement
- Secure Fingerprint Authentication Using Deep Learning and Minutiae Verification
- The Use of Natural Language Processing Approach for Converting Pseudo Code to C# Code
- Non-word Attributes’ Efficiency in Text Mining Authorship Prediction
- Design and Evaluation of Outlier Detection Based on Semantic Condensed Nearest Neighbor
- An Efficient Quality Inspection of Food Products Using Neural Network Classification
- Opposition Intensity-Based Cuckoo Search Algorithm for Data Privacy Preservation
- M-HMOGA: A New Multi-Objective Feature Selection Algorithm for Handwritten Numeral Classification
- Analogy-Based Approaches to Improve Software Project Effort Estimation Accuracy
- Linear Regression Supporting Vector Machine and Hybrid LOG Filter-Based Image Restoration
- Fractional Fuzzy Clustering and Particle Whale Optimization-Based MapReduce Framework for Big Data Clustering
- Implementation of Improved Ship-Iceberg Classifier Using Deep Learning
- Hybrid Approach for Face Recognition from a Single Sample per Person by Combining VLC and GOM
- Polarity Analysis of Customer Reviews Based on Part-of-Speech Subcategory
- A 4D Trajectory Prediction Model Based on the BP Neural Network
- A Blind Medical Image Watermarking for Secure E-Healthcare Application Using Crypto-Watermarking System
- Discriminating Healthy Wheat Grains from Grains Infected with Fusarium graminearum Using Texture Characteristics of Image-Processing Technique, Discriminant Analysis, and Support Vector Machine Methods
- License Plate Recognition in Urban Road Based on Vehicle Tracking and Result Integration
- Binary Genetic Swarm Optimization: A Combination of GA and PSO for Feature Selection
- Enhanced Twitter Sentiment Analysis Using Hybrid Approach and by Accounting Local Contextual Semantic
- Cloud Security: LKM and Optimal Fuzzy System for Intrusion Detection in Cloud Environment
- Power Average Operators of Trapezoidal Cubic Fuzzy Numbers and Application to Multi-attribute Group Decision Making
Articles in the same Issue
- An Optimized K-Harmonic Means Algorithm Combined with Modified Particle Swarm Optimization and Cuckoo Search Algorithm
- Texture Feature Extraction Using Intuitionistic Fuzzy Local Binary Pattern
- Leaf Disease Segmentation From Agricultural Images via Hybridization of Active Contour Model and OFA
- Deadline Constrained Task Scheduling Method Using a Combination of Center-Based Genetic Algorithm and Group Search Optimization
- Efficient Classification of DDoS Attacks Using an Ensemble Feature Selection Algorithm
- Distributed Multi-agent Bidding-Based Approach for the Collaborative Mapping of Unknown Indoor Environments by a Homogeneous Mobile Robot Team
- An Efficient Technique for Three-Dimensional Image Visualization Through Two-Dimensional Images for Medical Data
- Combined Multi-Agent Method to Control Inter-Department Common Events Collision for University Courses Timetabling
- An Improved Particle Swarm Optimization Algorithm for Global Multidimensional Optimization
- A Kernel Probabilistic Model for Semi-supervised Co-clustering Ensemble
- Pythagorean Hesitant Fuzzy Information Aggregation and Their Application to Multi-Attribute Group Decision-Making Problems
- Using an Efficient Optimal Classifier for Soil Classification in Spatial Data Mining Over Big Data
- A Bayesian Multiresolution Approach for Noise Removal in Medical Magnetic Resonance Images
- Gbest-Guided Artificial Bee Colony Optimization Algorithm-Based Optimal Incorporation of Shunt Capacitors in Distribution Networks under Load Growth
- Graded Soft Expert Set as a Generalization of Hesitant Fuzzy Set
- Universal Liver Extraction Algorithm: An Improved Chan–Vese Model
- Software Effort Estimation Using Modified Fuzzy C Means Clustering and Hybrid ABC-MCS Optimization in Neural Network
- Handwritten Indic Script Recognition Based on the Dempster–Shafer Theory of Evidence
- An Integrated Intuitionistic Fuzzy AHP and TOPSIS Approach to Evaluation of Outsource Manufacturers
- Automatically Assess Day Similarity Using Visual Lifelogs
- A Novel Bio-Inspired Algorithm Based on Social Spiders for Improving Performance and Efficiency of Data Clustering
- Discriminative Training Using Noise Robust Integrated Features and Refined HMM Modeling
- Self-Adaptive Mussels Wandering Optimization Algorithm with Application for Artificial Neural Network Training
- A Framework for Image Alignment of TerraSAR-X Images Using Fractional Derivatives and View Synthesis Approach
- Intelligent Systems for Structural Damage Assessment
- Some Interval-Valued Pythagorean Fuzzy Einstein Weighted Averaging Aggregation Operators and Their Application to Group Decision Making
- Fuzzy Adaptive Genetic Algorithm for Improving the Solution of Industrial Optimization Problems
- Approach to Multiple Attribute Group Decision Making Based on Hesitant Fuzzy Linguistic Aggregation Operators
- Cubic Ordered Weighted Distance Operator and Application in Group Decision-Making
- Fault Signal Recognition in Power Distribution System using Deep Belief Network
- Selector: PSO as Model Selector for Dual-Stage Diabetes Network
- Oppositional Gravitational Search Algorithm and Artificial Neural Network-based Classification of Kidney Images
- Improving Image Search through MKFCM Clustering Strategy-Based Re-ranking Measure
- Sparse Decomposition Technique for Segmentation and Compression of Compound Images
- Automatic Genetic Fuzzy c-Means
- Harmony Search Algorithm for Patient Admission Scheduling Problem
- Speech Signal Compression Algorithm Based on the JPEG Technique
- i-Vector-Based Speaker Verification on Limited Data Using Fusion Techniques
- Prediction of User Future Request Utilizing the Combination of Both ANN and FCM in Web Page Recommendation
- Presentation of ACT/R-RBF Hybrid Architecture to Develop Decision Making in Continuous and Non-continuous Data
- An Overview of Segmentation Algorithms for the Analysis of Anomalies on Medical Images
- Blind Restoration Algorithm Using Residual Measures for Motion-Blurred Noisy Images
- Extreme Learning Machine for Credit Risk Analysis
- A Genetic Algorithm Approach for Group Recommender System Based on Partial Rankings
- Improvements in Spoken Query System to Access the Agricultural Commodity Prices and Weather Information in Kannada Language/Dialects
- A One-Pass Approach for Slope and Slant Estimation of Tri-Script Handwritten Words
- Secure Communication through MultiAgent System-Based Diabetes Diagnosing and Classification
- Development of a Two-Stage Segmentation-Based Word Searching Method for Handwritten Document Images
- Pythagorean Fuzzy Einstein Hybrid Averaging Aggregation Operator and its Application to Multiple-Attribute Group Decision Making
- Ensembles of Text and Time-Series Models for Automatic Generation of Financial Trading Signals from Social Media Content
- A Flame Detection Method Based on Novel Gradient Features
- Modeling and Optimization of a Liquid Flow Process using an Artificial Neural Network-Based Flower Pollination Algorithm
- Spectral Graph-based Features for Recognition of Handwritten Characters: A Case Study on Handwritten Devanagari Numerals
- A Grey Wolf Optimizer for Text Document Clustering
- Classification of Masses in Digital Mammograms Using the Genetic Ensemble Method
- A Hybrid Grey Wolf Optimiser Algorithm for Solving Time Series Classification Problems
- Gray Method for Multiple Attribute Decision Making with Incomplete Weight Information under the Pythagorean Fuzzy Setting
- Multi-Agent System Based on the Extreme Learning Machine and Fuzzy Control for Intelligent Energy Management in Microgrid
- Deep CNN Combined With Relevance Feedback for Trademark Image Retrieval
- Cognitively Motivated Query Abstraction Model Based on Associative Root-Pattern Networks
- Improved Adaptive Neuro-Fuzzy Inference System Using Gray Wolf Optimization: A Case Study in Predicting Biochar Yield
- Predict Forex Trend via Convolutional Neural Networks
- Optimizing Integrated Features for Hindi Automatic Speech Recognition System
- A Novel Weakest t-norm based Fuzzy Fault Tree Analysis Through Qualitative Data Processing and Its Application in System Reliability Evaluation
- FCNB: Fuzzy Correlative Naive Bayes Classifier with MapReduce Framework for Big Data Classification
- A Modified Jaya Algorithm for Mixed-Variable Optimization Problems
- An Improved Robust Fuzzy Algorithm for Unsupervised Learning
- Hybridizing the Cuckoo Search Algorithm with Different Mutation Operators for Numerical Optimization Problems
- An Efficient Lossless ROI Image Compression Using Wavelet-Based Modified Region Growing Algorithm
- Predicting Automatic Trigger Speed for Vehicle-Activated Signs
- Group Recommender Systems – An Evolutionary Approach Based on Multi-expert System for Consensus
- Enriching Documents by Linking Salient Entities and Lexical-Semantic Expansion
- A New Feature Selection Method for Sentiment Analysis in Short Text
- Optimizing Software Modularity with Minimum Possible Variations
- Optimizing the Self-Organizing Team Size Using a Genetic Algorithm in Agile Practices
- Aspect-Oriented Sentiment Analysis: A Topic Modeling-Powered Approach
- Feature Pair Index Graph for Clustering
- Tangramob: An Agent-Based Simulation Framework for Validating Urban Smart Mobility Solutions
- A New Algorithm Based on Magic Square and a Novel Chaotic System for Image Encryption
- Video Steganography Using Knight Tour Algorithm and LSB Method for Encrypted Data
- Clay-Based Brick Porosity Estimation Using Image Processing Techniques
- AGCS Technique to Improve the Performance of Neural Networks
- A Color Image Encryption Technique Based on Bit-Level Permutation and Alternate Logistic Maps
- A Hybrid of Deep CNN and Bidirectional LSTM for Automatic Speech Recognition
- Database Creation and Dialect-Wise Comparative Analysis of Prosodic Features for Punjabi Language
- Trapezoidal Linguistic Cubic Fuzzy TOPSIS Method and Application in a Group Decision Making Program
- Histopathological Image Segmentation Using Modified Kernel-Based Fuzzy C-Means and Edge Bridge and Fill Technique
- Proximal Support Vector Machine-Based Hybrid Approach for Edge Detection in Noisy Images
- Early Detection of Parkinson’s Disease by Using SPECT Imaging and Biomarkers
- Image Compression Based on Block SVD Power Method
- Noise Reduction Using Modified Wiener Filter in Digital Hearing Aid for Speech Signal Enhancement
- Secure Fingerprint Authentication Using Deep Learning and Minutiae Verification
- The Use of Natural Language Processing Approach for Converting Pseudo Code to C# Code
- Non-word Attributes’ Efficiency in Text Mining Authorship Prediction
- Design and Evaluation of Outlier Detection Based on Semantic Condensed Nearest Neighbor
- An Efficient Quality Inspection of Food Products Using Neural Network Classification
- Opposition Intensity-Based Cuckoo Search Algorithm for Data Privacy Preservation
- M-HMOGA: A New Multi-Objective Feature Selection Algorithm for Handwritten Numeral Classification
- Analogy-Based Approaches to Improve Software Project Effort Estimation Accuracy
- Linear Regression Supporting Vector Machine and Hybrid LOG Filter-Based Image Restoration
- Fractional Fuzzy Clustering and Particle Whale Optimization-Based MapReduce Framework for Big Data Clustering
- Implementation of Improved Ship-Iceberg Classifier Using Deep Learning
- Hybrid Approach for Face Recognition from a Single Sample per Person by Combining VLC and GOM
- Polarity Analysis of Customer Reviews Based on Part-of-Speech Subcategory
- A 4D Trajectory Prediction Model Based on the BP Neural Network
- A Blind Medical Image Watermarking for Secure E-Healthcare Application Using Crypto-Watermarking System
- Discriminating Healthy Wheat Grains from Grains Infected with Fusarium graminearum Using Texture Characteristics of Image-Processing Technique, Discriminant Analysis, and Support Vector Machine Methods
- License Plate Recognition in Urban Road Based on Vehicle Tracking and Result Integration
- Binary Genetic Swarm Optimization: A Combination of GA and PSO for Feature Selection
- Enhanced Twitter Sentiment Analysis Using Hybrid Approach and by Accounting Local Contextual Semantic
- Cloud Security: LKM and Optimal Fuzzy System for Intrusion Detection in Cloud Environment
- Power Average Operators of Trapezoidal Cubic Fuzzy Numbers and Application to Multi-attribute Group Decision Making














