Skip to content
BY-NC-ND 3.0 license Open Access Published by De Gruyter March 16, 2017

Feature Selection Using Harmony Search for Script Identification from Handwritten Document Images

  • Pawan Kumar Singh EMAIL logo , Supratim Das , Ram Sarkar and Mita Nasipuri

Abstract

The feature selection process can be considered a problem of global combinatorial optimization in machine learning, which reduces the irrelevant, noisy, and non-contributing features, resulting in acceptable classification accuracy. Harmony search algorithm (HSA) is an evolutionary algorithm that is applied to various optimization problems such as scheduling, text summarization, water distribution networks, vehicle routing, etc. This paper presents a hybrid approach based on support vector machine and HSA for wrapper feature subset selection. This approach is used to select an optimized set of features from an initial set of features obtained by applying Modified log-Gabor filters on prepartitioned rectangular blocks of handwritten document images written in either of 12 official Indic scripts. The assessment justifies the need of feature selection for handwritten script identification where local and global features are computed without knowing the exact importance of features. The proposed approach is also compared with four well-known evolutionary algorithms, namely genetic algorithm, particle swarm optimization, tabu search, ant colony optimization, and two statistical feature dimensionality reduction techniques, namely greedy attribute search and principal component analysis. The acquired results show that the optimal set of features selected using HSA gives better accuracy in handwritten script recognition.

1 Introduction

Many strategies have been exploited for the task of feature selection (FS), in an effort to identify more compact and better-quality feature subsets. The development of nature-inspired stochastic search techniques allows multiple good-quality feature subsets to be discovered without resorting to exhaustive search. Selecting a subset of features from the entire feature set is often beneficial because it lessens the computation time along with helping achieve better classification accuracy. At a higher dimension, it becomes very difficult to visualize the data and there may be some non-contributing features that may even mislead the classifier’s decision [31]. Identification of those features at higher dimension, though a challenging task, motivates the researchers to work in this domain. For a large feature dimensionality, it is almost impossible to check the accuracies considering all possible feature subsets. Thus, one may apply the hill climbing (HC) method by successively adding/removing feature(s) into/from the original feature set. However, the HC method gets stuck at local minima. Thus, the need to explore the search space using some heuristic search techniques will be of utmost importance. Practical problems that arise during analyzing the real-world data are often related to the number of features (so called curse of dimensionality). The inability of the classification tools to identify and extract patterns or rules easily may also be attributed to high interdependency among individual features, or the behavior of the combined features. Techniques such as text processing and classification can benefit greatly from FS, once the noisy, irrelevant, redundant, or misleading features are removed. Given a feature set with size n, there will be 2n possible feature subsets. The FS problem is to find a minimal feature subset while retaining a suitably high accuracy in classifying the data [25].

The FS approaches can generally be divided into two groups: filter and wrapper approaches. The filter approach operates independently of any learning algorithm. These methods rank the features by some criteria and neglect all features that do not achieve a sufficient score. Due to their computational efficiency, the filter methods are very popular for high-dimension data. Some popular filter methods are F-score criterion [7], mutual information [40], information gain [1], correlation [17], etc. On the other hand, the wrapper approaches involve a predetermined learning model while selecting features on measuring the parameters of the particular machine learning model [17]. Filter approaches are usually less computationally intensive than wrappers, but they produce a feature set that is biased to a specific type of predictive model. This unbiased tuning means a feature set from a filter is more general than the set from a wrapper, usually giving lower prediction performance than a wrapper. However, the feature set selected using the former method does not contain the assumptions of a prediction model, and so is more useful for exposing the relationships among the original features.

The rest of the paper is organized as follows: Section 2 presents a brief review of some of the previous approaches to FS, and Section 3 illustrates the motivation behind the proposed work. Section 4 describes the Modified log-Gabor filter-based feature extraction procedure for handwritten Indic script identification, whereas in Section 5, the basics of the harmony search (HS)-based feature subset selection process are discussed. Section 6 describes the performance of the proposed script identification system, and finally, Section 7 concludes the work.

2 Related Works

There has been an immense amount of work in the area of FS, which credits its importance. Sivagaminathan and Ramakrishnan [35] presented a hybrid method based on ant colony optimization (ACO) and artificial neural networks to address FS. The proposed hybrid model was demonstrated using datasets from the domain of medical diagnosis, yielding promising results. El Alami proposed a novel text categorization algorithm [9] called chaos genetic FS optimization in order to choose a subset of available features by eliminating unnecessary features to the classification task. The proposed algorithm selected the optimal subsets in both empirical and theoretical works in machine learning, and presented a general framework for text categorization. Experimental results showed that the proposed algorithm simplified the FS process effectively and obtained higher classification accuracy with only a subset of original features. Bermejo et al. [3] proposed a stochastic algorithm based on the greedy randomized adaptive search procedure (GRASP) meta-heuristic, with the main goal of speeding up the feature subset selection process, basically by reducing the number of wrapper evaluations to be carried out. GRASP is a multi-start constructive method that formulates a solution in its first stage, and then applies a modifier stage on that solution for an improved result. Several instances of the proposed GRASP method were experimentally tested and compared with state-of-the-art algorithms over 12 high-dimensional datasets. The statistical analysis of the results illustrated that the proposal was comparable in accuracy and cardinality of the selected subset to previous algorithms, but required significantly fewer evaluations. Forsati et al. [11] formulated the problem of FS as an optimization problem and introduced a novel FS procedure in order to achieve better classification results. Experiments over a standard benchmark demonstrated that applying bee colony optimization in the context of FS was a feasible approach. Das et al. [4] proposed a methodology where local regions of varying heights and widths were created dynamically on the pattern images. Genetic algorithm (GA) was then applied on these local regions to sample the optimal set of local regions containing best distinguishing information among the pattern classes. An optimal feature set was then extracted from those selected regions. Other popular optimization techniques like simulated annealing and HC had also been evaluated and, finally, a comparison with the proposed technique was done on a dataset of handwritten Bangla digits. Vieira et al. [39] developed modified binary particle swarm optimization (MBPSO) for feature subset selection. It optimized the kernel parameters of support vector machine (SVM) as well as controlled the premature convergence of the algorithm. Zeng et al. [41] introduced a novel hybrid FS method based on rough conditional mutual information and naive Bayesian classifier, which could also be used to filter the irrelevant and redundant features. Subsequently, to reduce the feature and improve classification accuracy, a wrapper approach based on naive Bayesian classifier was used to search the optimal feature subset in the space of a candidate feature subset that was selected by filter model. The results showed that the approach obtained not only high classification accuracy, but also selected the least number of features. Nazir et al. [27] proposed an efficient gender classification technique for real-world face images (labeled faces in the wild). This work extracted facial local features using local binary pattern (LBP), and then these features were fused with clothing features to enhance the classification accuracy. In the following step, PSO and GA were combined to select the most important features set that more clearly represented the gender, and thus the data dimensionality was reduced. Roy et al. [29] implemented artificial bee colony optimization to pin point the set of local regions offering the optimal discriminating feature set for handwritten numeral and character recognition. Initially, eight-directional gradient features were extracted from every region of different levels of partitions created using a center of gravity-based quad-tree partitioning approach. Then, using this approach, at each level, the sampling process was done based on SVM in every single region. De Stefano et al. [5] presented a novel GA-based FS algorithm in which feature subsets were evaluated by means of a specifically devised separability index. This index measured the statistical properties of the feature subset and did not depend on any specific classification scheme. The proposed index represented an extension of the Fisher linear discriminant method and used covariance matrices for estimating how class probability distributions were spread out in the considered N-dimensional feature space. A key property of the approach was that it does not require any a priori knowledge about the number of features to be used in the feature subset. Roy et al. [30] proposed a new FS methodology on the basis of features combined class separability power, using the framework of axiomatic fuzzy set theory, which provided the rules for logic operations needed to interpret the combinations of features from the fuzzy feature set. Based on these combinational rules, the class separability power of the combined features was determined, and subsequently the most powerful subset of the feature set was selected. Mohammadi and Abadeh [26] introduced a new feature-based blind steganalysis method for detecting stego images from the cover images in JPEG images using an FS technique based on the artificial bee colony optimization technique. Kashef and Nezamabadi-pour [18] presented a novel feature selection algorithm based on ACO, called advanced binary ACO. Features were treated as graph nodes to construct a graph model and were fully connected to each other. The performance of the proposed algorithm was compared with the performance of binary GA, binary PSO (BPSO), catfish BPSO, improved binary gravitational search algorithm, and some prominent ACO-based algorithms on the task of FS on 12 well-known UCI datasets. Simulation results verified that the algorithm provided a suitable feature subset with good classification accuracy using a smaller feature set than competing FS methods. Ahmad [2] used principal component analysis (PCA) for feature reduction and subset selection in which features were primarily projected into a principal space and then elected based on their eigenvalues, but the feature with the highest eigenvalue may not had the guarantee to provide optimal sensitivity for the classifier. To avoid this problem, two evolutionary optimization approaches like GA and PSO had been used to search the most discriminative subset of transformed features. Finally, it was observed that a feature subset selection based on PSO provided better performance as compared to GA. Singh et al. [32] proposed a PSO- and F-score-based FS algorithm for selecting the significant features that contributed to improve the classification accuracy. The performance of their proposed method was evaluated with various classifiers such as SVM, naive Bayes, k-NN, and decision tree. The experimental results illustrated that the proposed method outperformed some existing methods.

In particular, HS is a recently developed technique mimicking musicians’ experiences, which has been effectively utilized to cope with FS problems. Kumar et al. [22] presented a novel variance-based HS (VHS) algorithm for solving optimization problems. The problem of setting a constant parameter in the algorithm was removed, and the technique was named as explorative HS. The proposed algorithm was then employed for a data clustering problem on four real-life datasets chosen from the UCI machine learning repository. The results indicated that the VHS-based clustering outperformed the existing well-known clustering algorithms. Kumar et al. [21] introduced a parameter adaptive HS (PAHS) algorithm for solving optimization problems. The two important parameters of HS algorithm (HSA), namely harmony memory consideration rate (HMCR) and pitch adjusting rate (PAR), were allowed to dynamically change in their proposed algorithm in order to get the global optimal solution. Evaluation of the proposed algorithm for data clustering on five real-life UCI datasets showed better results than other algorithms. Kumar et al. [23] assumed the problem of automatic data clustering as the searching of optimal number of clusters in order to optimize the attained partitions. The PAHS algorithm (as described in Ref. [21]) was employed as an underlying optimization strategy. The data points of the different cluster centroids were assigned on the basis of newly developed weighted Euclidean distance. The algorithm was compared with four existing clustering techniques, and the same was further applied for automatic segmentation of grayscale and color images. Kumar et al. [20] proposed a novel automatic unsupervised FS method based on gravitational search algorithm, called AFSGSA (automatic FS using the gravitational search algorithm). The algorithm requires no prior knowledge of the data to be classified and the number of features to be selected. A novel fitness function was also included in order to make the search more efficient. Comparison of the performance of AFSGSA concluded the efficiency and efficacy of the proposed FS technique over other existing techniques for the same. In this paper, we present a wrapper FS method based on HS to discover a minimal feature subset from a problem domain while retaining a suitably high accuracy in representing the original data in the feature space. Applying this FS method, we have achieved some significant improvement in the recognition accuracy of the Modified log-Gabor filter-based handwritten Indic script identification technique.

3 Motivation

India is a multilingual country with 23 constitutionally recognized languages [38] written in 12 major scripts. Besides these, hundreds of other languages are used in India, each one with a number of dialects. The officially recognized languages are Hindi, Bengali, Punjabi, Marathi, Gujarati, Oriya, Sindhi, Assamese, Nepali, Urdu, Sanskrit, Tamil, Telugu, Kannada, Malayalam, Kashmiri, Manipuri, Konkani, Maithali, Santhali, Bodo, Dogari, and English. The 12 major scripts used to write these languages are Devanagari, Bangla, Oriya, Gujarati, Gurumukhi, Tamil, Telugu, Kannada, Malayalam, Manipuri, Roman, and Urdu. Of these, Urdu is derived from the Persian script and is written from right to left. The first 10 scripts are originated from the early Brahmi script (300 BC) and are also referred to as Indic scripts. Due to the fact that any optical character recognition (OCR) system can recognize only a script of particular type, it is impossible to design a single recognizer that can identify a variety of scripts/languages. An alternate solution to handle this problem is to use a pool of OCRs corresponding to different scripts. This, in turn, requires the identification of the document’s script before feeding it to the corresponding OCR system. Here lies the problem of script identification. In general, script identification can be achieved at any of the three levels: (a) page level, (b) text-line level, and (c) word level. Identifying scripts at page level can sometimes be too convoluted and protracted due to large computational complexity. Again, the identification of scripts at text-line or word level requires the accurate segmentation of the document pages into their corresponding text-lines and words. That is, the accuracy of the script identification, in turn, depends on the accuracies of its text-line and word segmentation methods, respectively. In addition, identifying text words of different scripts with only a few numbers of characters may not always be feasible because at word level, the number of characters present in a single word may not be always informative. Thus, in the present work, we have performed script identification at block level, where entire document pages are decomposed gradually into their corresponding smaller subblocks using a quad-tree-based page subdivision approach.

A comprehensive literature survey on Indic script identification (described in Ref. [34]) indicates that a lot of researchers have contributed to the handwritten script identification problem by inventing/deriving a bundle of features; however, the exploration of the optimum feature set for this problem has not received any attention from the research community till date in spite of its importance. In general, it is noticed that heuristic search techniques, namely GA, PSO, ACO, and TS, have mostly been applied to reduce the complexity of the FS process. However, in this paper, an HS-based FS procedure is described instead, owing to its simplicity and stochastic behavior (which helps in escaping from trapping at local optima). In comparison with GA-, ACO-, TS-, and PSO-based searches, HS is less explored by the research community, although it has gained huge popularity in recent times. The present HS-based search is also compared with some preceding entrenched optimization methods, and the experimental results also confirm the effectiveness of HS-based search.

4 Proposed Work

The proposed work employs a two-pass approach. In the first pass, the texture-based features are extracted by applying Modified log-Gabor filters [33] designed at multiple scales and multiple orientations. These features are estimated at different sub-image levels of the document images decomposed by the quad-tree-based approach. This approach is recursively applied to subdivided image segments until the image segment does not provide any significant information. Thus, at the first level, the whole document image is divided into four equal blocks (L1, L2, L3, and L4), whereas at the second level, L1 block is again subdivided into another four subblocks, namelyL11, L12, L13, and L14. Similarly, each of the other blocks, i.e. L2, L3, and L4, are also subdivided into four subblocks. Thus, at second-level decomposition, a total of 16 blocks is realized. Therefore, the total number of subblocks at l-th level can be written as 4l, where l denotes the level of decomposition. For an image of size M×N, the size of each subblock at l-th level is (M/2l)×(N/2l). For the present work, the maximal value of l is chosen to be 4. Figure 1 shows the quad-tree decomposition for a handwritten Tamil document image. It is evident from Figure 1E that the text block at fourth level has just half the word images of two consecutive text-lines. If the page decomposition is performed beyond fourth level, we would be left with hardly one or two characters of the word image, which may not be sufficient for the script identification purpose. Thus, we limit our quad-tree-based approach for page decomposition till l=4.

Figure 1: Illustration of (A) Handwritten Tamil Document Page after (B) First-Level (L1), (C) Second-Level (L11), (D) Third-Level (L111), and (E) Fourth-Level (L1111) Decomposition.
Figure 1:

Illustration of (A) Handwritten Tamil Document Page after (B) First-Level (L1), (C) Second-Level (L11), (D) Third-Level (L111), and (E) Fourth-Level (L1111) Decomposition.

In the second pass, the HSA is applied to select the feature subset where the objective function is the performance of the SVM classifier. The block diagram of the proposed two-pass approach is shown in Figure 2.

Figure 2: Schematic Diagram of the Proposed Two-Pass Approach for Handwritten Script Identification.
Figure 2:

Schematic Diagram of the Proposed Two-Pass Approach for Handwritten Script Identification.

4.1 Modified Log-Gabor Filters

The log-Gabor function was proposed by Field [10] as a means to overcome the limitations of Gabor function. Field suggested that when viewed on a logarithmic frequency scale, coding of natural images by filters having Gaussian transfer functions is superior. The log-Gabor function has a transfer function of the following form:

(1)G(w)=e(log(w/w0)2)(2(log(k/w0)2)

where wo is the filter’s center frequency. To obtain constant shape ratio filters, the term k/wo must also be held constant for all chosen values of wo.

The log-Gabor functions do not have any DC component, and their transfer function has an extended tail at the high frequency end. The studies of Field on the statistics of ordinary images claim that these images have amplitude spectra that gradually fall off at approximately 1/w. In order to encode those images that possess such spectral characteristic filters, similar spectra must be used. Further, log-Gabor functions having extended tails are able to encode images more efficiently than ordinary Gabor functions, as these log-Gabor filters neither overrepresent the lower-frequency components nor underrepresent the high-frequency components in any encoding.

For proper texture analysis of an image using the features extracted using log-Gabor filter, the frequency components are necessary. In order to retain the spatial information eliminated by a simple Fourier transform, a windowed Fourier transform (WFT) is used. WFT involves multiplication of the image X(n0, n1) by the window function w(n0, n1) followed by application of the Fourier transform on the result to obtain

(2)τ(n0(0),n1(0),u0,u1)=X(n0,n1)w(n0n0(0),n1n1(0))e2πi(u0n0+u1n1)dn0dn1=e2πi(u0n0(0)+u1n1(0))[X(n0(0),n1(0))(w(n0(0),n1(0))e2πi(u0n0(0)+u1n1(0)))]=e2πi(u0n0(0)+u1n1(0))[Xmu0,u1](n0(0),n1(0)).

Here, (n0(0), n1(0)) is the position in the image where the horizontal and vertical frequencies u0 and u1 are computed. The WFT is basically a convolution of the image with the filter mu0,u1. For texture analysis, both spatial and frequency information are desired. Thus, we try to achieve a good trade-off between these two with respect to the uncertainty principle of Fourier transform. Gabor transforms use a Gaussian function as the optimally concentrated function in the spatial as well as in the frequency domain [10]. Here, non-isotropic Gaussian is of the form

(3)mf,(n0,n1)=12πσ2λe12σ2(n02λ2+n12)e2πifn0
(4)Mf,φ(u0,u1)=e2π2σ2[(u0f)2λ2+u12]

with the center frequency f=u02+u12 and the rotated coordinates (ń0, ń1)=(n0 cos ∅+n1sin ∅−n0 sin ∅+n1 cos ∅)).

Here, 1λ is the aspect ratio.

Due to the convolution theorem, the filter interpretation of the Gabor transform allows the efficient computation of the Gabor coefficients Gf,∅ (n0, n1) by multiplication of the Fourier transformed image τ (u0, u1) with the Fourier transform of the Gabor filter Mf,∅(u0, u1). The inverse Fourier transform is then applied on the resultant vector as defined below:

(5)Gf,φ(n0,n1)=FFT1{τ(u0,u1).Mf,(u0,u1)}

The images, after low pass filtering, are passed as input to a function that computes the Gabor energy feature from them. This is done by multiplying in frequency domain [which is equivalent to two-dimensional (2D) convolution in the time-space domain]. The input image is then passed to a function to yield a Gabor array, which is the array equivalent of the image after Gabor filtering. The function displays the image equivalent of the magnitude and the real part of the Gabor array pixels.

Consider that there are ns scales and nonumber of orientations, resulting in ns×nodifferent filters. Let J denote the Fourier transform of the input image, Gs,othe Gabor filter at scale s and orientation o, and Vs,othe output of the convolution of Gs,oand J.

(6)Vs,o=JGs,o.

The local responses of each of the Gabor filters can also be represented in terms of amplitude As,o(x, y) and energy Es,o(x, y), as defined below:

(7)As,o=|Vs,o(x,y)|

and

(8)Es,o(x,y)=|Real{Vs,o(x,y)}||Img{Vs,o(x,y)}|,

where (x, y) denotes the 2D coordinates of a pixel, and Real and Img denote the real and imaginary parts of the filter responses, respectively. Next, the median overall orientations for a fixed scale s for As,oand Es,o are calculated as follows [14]:

(9)As(x,y)=median{o=1,2,no}As,o(x,y)

and

(10)Es(x,y)=median{o=1,2,no}Es,o(x,y).

For the present work, features based on Modified log-Gabor filters have been extracted for five scales (ns=1, 2, 3, 4, and 5) and 12 orientations (no=0°, 15°, 30°, … …., 165°, 180°), where each filter has to be convolved with the input image to obtain 60 (i.e. 12*5) different representations (response matrices) for the same image. These response matrices are then converted to feature vectors. Each input image provides us with one feature vector consisting each of 60 elements calculated using amplitude As(x, y) and energy Es(x, y) values. Finally, a set of 120 multiscale and multidirectional oriented features has been used for the block-level recognition of 12 official scripts.

The algorithm of the feature extraction methodology can be described as follows:

Step 1: Input an RGB document image and convert it into a grayscale image I (x, y), where x and y are the spatial coordinates of the image.

Step 2: Select an integer value l, where l indicates the level of decomposition. For the present work, the optimal value of l is varied from 2 to 4.

Step 3: Perform l-level quad-tree decomposition to generate 4l blocks. Feature values are generated from each of these segmented blocks.

Step 4: Convolve the block image of step 3 with modified log-Gabor filter with five different scales (ns=1, 2, 3, 4, and 5) and 12 different orientations (no=0°, 15°, 30°, … …., 165°, 180°). Thus, for each block, we get a total of 60 convolved filter outputs.

Step 5: Acquire the absolute value of these filtered output block images and calculate the amplitude and energy values.

Step 6: Repeat step 3 to step 5 to extract a total of 120 features from each of the script block images.

5 HS Optimization

HS is a music-based meta-heuristic optimization algorithm, proposed by Geem [12] in 2000. It was inspired by the observation that the aspiration of a musician is to search for a perfect state of harmony. The search process in optimization can be compared to a jazz musician’s improvisation process. Likewise, an optimal solution to an optimization problem should be the desirable solution available to the problem under the given objectives and limited by constraints. Both processes intend to produce the best or optimum output. In order to explain the HS in more detail, let us first understand the improvisation process by a skilled musician. When a musician is improvising, he or she has three possible choices: (a) play any famous piece of music (a series of pitches in harmony) exactly from his or her memory, (b) play something similar to a known piece (thus adjusting the pitch slightly), or (c) compose new or random notes. Geem et al. [13] formalized these three options into a quantitative optimization process in 2001, and the three corresponding components became usage of harmony memory (HM), pitch adjusting, and randomization.

The usage of HM ensures that the best harmonies pass over to the new HM. This usage is controlled by assigning a parameter δ, known as HM accepting or considering rate. The value of δ normally lies between 0 and 1, but typically, it is set to 0.7. The second component called pitch adjustment, ρ∈[0.1, 0.5], is used to control the degree of the adjustment, and for most applications, this is set to 0.3. The third component called randomization is used to increase the diversity of the solutions. Although adjusting pitch has a similar role with randomization, it is limited to certain local pitch adjustment and thus corresponds to a local search. The use of randomization can drive the system further to explore diverse solutions so as to find the global optimality. It is given by

(11)Prand=1δ

Pitch adjustment is determined by a pitch bandwidth Pband and a pitch adjusting rate ρ. The linear adjustment of pitch is done in the following way:

(12)xnew=xold+Pbandε

where xold is the existing pitch or solution from the HM and xnew is the new pitch after the pitch adjusting action. This essentially produces a new solution around the existing quality solution by varying the pitch slightly by a small random amount. Here, ε is a random number generator in the range of [−1, 1].

5.1 Characteristics of HSA

While reviewing other meta-heuristic algorithms, it has been found that the success of any optimization process depends on two major contradicting components: diversification and intensification. Proper diversification or exploration makes sure the search in the parameter space can explore as many locations and regions as possible in an efficient and effective manner. It also ensures that the evolving system will not be trapped in biased local optima. On the other hand, the appropriate intensification or exploitation intends to exploit the history and experience of the search process. Its aim is to ensure to speed up the convergence by reducing the randomness and limiting diversification.

The two subcomponents, namely pitch adjustment and randomization, control the diversification in the HSA. These two parameters may contribute toward the high efficiency of the HS method. The level of efficiency of the first subcomponent in composing “new music”, or generating new solutions, via randomization would be at least similar in comparison to other algorithms attained by randomization. On the other hand, an additional subcomponent, pitch adjustment denoted by ρ, is calculated by regulating the pitch in the given bandwidth by a small random amount relative to the existing pitch or solution from the HM. Local solutions are refined by using this parameter. The preservation of good local solutions are ensured by both δ and ρ, while the global search space will be efficiently explored by randomization and δ. The vulnerability of this is that it is a controlled diversification around the good solutions (good harmonics and pitches), as well as it operates almost like an escalation factor. The intensification is mainly represented in the HSA by δ. A high δ implies a high chance of the selection or inheritance of good solutions from the memory, whereas a low δ means that the solutions will converge more slowly.

5.2 HSA-Based Feature Subset Selection

The HSA consists of the following five steps [24]:

Step 1: Initialization of the Optimization Problem and Algorithm Parameters

The optimization problem can be defined as follows:

Maximize f(x) subject to xiX, X=x1, x2, … … ….. xn, where f(x) is the objective function, Xi is the set of possible range of values for each decision variable, that is Lxi<Xi<Uxi, where Uxi and Lxi are the upper and lower bounds of each decision variable. The control parameters of HS are the HM size (HMS), i.e. the number of solution vectors in the HM (in each iteration), and the maximum number of iterations (Maxltr) or stopping criterion.

Step 2: HM Initialization

HM is a memory location where all the solution vectors (sets of decision variables) are stored. The HM matrix is initially filled with as many randomly generated solution vectors as the HMS.

(13)HM=[x11x21xN1x12x22xN2x1HMSx2HMSxNHMS]

Step 3: New Harmony Improvisation

In this step, a new harmony vector x=(x1,x2,x3,x3,,xN) is generated based on three rules: (a) memory consideration, (b) pitch adjustment, and (c) random selection. Generating a new harmony is called “improvisation”. In the memory consideration, the value of the first decision variable x1 for the new vector is chosen from the values already existing in the current HM, i.e. from the set {x11,,x1HMS}, with a probability δ. Values of the other decision variables x2,x3,x3,,xN are also chosen in the same manner. The value of δ, varying in the range 0 and 1, is the rate of choosing one value from the previous values stored in the HM, while (1−δ) is the rate of randomly selecting a new value from the possible range of values.

(14)xi={xiϵ{xi1,xi2,xi3,.,xiHMS}with probability δxiϵXiwith probability(1δ)

Every component obtained by the memory consideration is further examined to determine whether it should be pitch adjusted or not. This operation uses the parameter ρ as follows:

(15)xi={(xi±Pbandε)with probability ρxiwith probability(1ρ)

where Pband is the pitch bandwidth and ε is a random number generator in the range of [−1, 1].

Step 4: Update HM

If the new harmony vector, x=(x1,x2,x3,x3,,xN), is better than the worst harmony in the HM, from the point of view of objective function value, the new harmony is included in the HM and the existing worst harmony is excluded from the same.

Step 5: Repetition of Steps 3 and 4 Until the Stopping Criterion Is Satisfied

If the stopping criterion (i.e. maxItr) is satisfied, the computation is terminated. Otherwise, Steps 3 and 4 are repeated.

This is summarized in the flowchart shown in Figure 3.

Figure 3: Flowchart of the HSA Procedure.
Figure 3:

Flowchart of the HSA Procedure.

5.3 Example of FS Using HSA

All the data used in this example are hypothetical. Let us assume we have a total of 20 features, i.e. f1, f2, f3, … …. f20, and we want to select the best five features among them. Thus, the total number of variables taken under consideration is five. Here, our objective function is the classification accuracy evaluated by a given machine learning algorithm. The goal is to maximize this classification accuracy using HSA. Let us consider the HMS to be 6.

HMSClassification Score
Set #1<f1, f5, f3, f2, f9>40%
Set #2<f2, f7, f4, f1, f6>60%
Set #3<f10, f1, f2, f3, f7>70%
Set #4<f12, f17, f3, f19, f1>50%
Set #5<f1, f10, f20, f13, f5>30%
Set #6<f2, f9, f5, f14, f18>63%

Here, each solution vector is generated randomly. Each musician must select a value from his or her note domain at the time of improvisation to generate a new solution vector. This selection of note, from the existing domain of notes of that particular musician, is controlled by two parameters (i.e. δ and ρ). If we set δ=0.9 and ρ=0.1, then the first musician chooses a new note from the domain {f1, f2, f10, f12, f1, f2} with 0.9 probability; however, there still exists a probability of 0.1, i.e. (1−0.9), that the first musician may choose from the range of all possible attribute values [i.e. (f1, f2, f3, … … .. f20)]. Let us assume that the first musician has chosen f17 value from all possible existing note values; then, as per our HSA, the rest of the musicians cannot choose the f17 value. Thus, f17 needs to be excluded from the domain of the second musician; thus, the second musician can choose a new note from the domain {f5, f7, f1, f10, f9} with 0.9 probability as δ=0.9. Similarly, there exists a probability of 0.1 that the second musician may choose from the range of all possible attributes values [i.e. (f1, f2, f3, …, f16, f18, f19, f20)]. This FS based on δ value may not be the final selection in the new solution vector, as ρ can shift the previously selected feature. For example, here, the first musician has already chosen f17; however, as ρ=0.1, after making these choices, the first musician can choose left or right neighbors of f17 with 0.1 probability. If Pband=2, thus the first musician has total choices of five values { f15, f16, f17, f18, f19}. Here, Pband indicates that the first musician can reach up to the position along the left or right neighboring direction. Let us assume, at the end, that the first musician has chosen f18. In this way, all the musicians need to select note values. At the final stage of the improvisation process, let the new solution vector is (f18, f1, f20, f14, f4). Now, with this feature set, testing is done using any supervised classifier. Let us assume that we obtained 80% classification accuracy using the newly generated solution vector. This is better than the worst scored solution vector [i.e. <f1, f10, f20, f13, f5> with 30% classification accuracy] in the HM. Thus, we update the memory with the newly generated solution vector by replacing the worst solution vector. This type of improvisation process continues until MaxItr is met.

5.4 Application of HSA to Modified Log-Gabor Filter Technique

This subsection describes the selection of optimal features extracted from the modified log-Gabor filter procedure. Here, our main objective is to reduce the number of features using HSA so that the redundant or misleading features get eliminated, which, in turn, increases the classification accuracy. Initially, let us assume 20% of the total number of features (i.e. 120) would provide a good classification score. This statement is also validated by increasing the total number of features by 20% in the next experiment. In this way, the experimentation is done by taking 20%, 40%, 60%, and 80% of the original feature set. That is, the increment factor (IF) is set to 20%. Here, we have a total of 120 features, i.e. f1, f2, f3, … … .. f120, and suppose we want to select 24 optimal features among them (as our initial assumption is that 20% of total feature can give us a good result). Thus, the total number of variables taken under consideration is 24. Let us consider the initial value of HMS to be 10. At a later stage, this value is made to vary, and it is found that the initial value is the sufficient value as an increase of HMS value does not affect the overall result in our experiment.

The pseudo code for selection of optimal feature subset for the proposed work is as follows.

Here, we have a total of 120 features, i.e. f1, f2, f3, … … .. f120, and initially we have tried with 24 features in order to find the optimal set of feature of length 24.

Here, 24 features are used to define the objective function f(x). Let us consider we have features fi, fj, fk, … … .. fm, where the subscript of each feature denotes the corresponding feature number and 1<=i, j, k, … . m<=120 and ijk≠ … … .≠m. Here, our objective function is the maximum classification accuracy evaluated by a given machine learning algorithm. The goal is to maximize this classification accuracy by applying HSA.

Begin

Initialize PAR and bandwidth

Initialize HMCR, HMS, IF, and MaxItr

Initialize the number of decision variable v (for this experiment, v is initialized to 20% of the original feature set)

Fill the HMS with some probable solution vectors by choosing random harmonies between f1 and f120

Forc=1 to ((100 / IF)−1)

// loop will execute four times, i.e. 20%, 40%, 60%, and 80%, as IF=20%

{

while(t<MaxItr)

//t indicating current iteration number must be initialized with 1

{

For each musician (xi )

//here total number of musician depends on v

{

Generate a random number rn1, where 0≤rn1≤1

if (rn1<HMCR)

{

Randomly choose a note from the existing domain of xi

Generate a random number rn2, where 0≤rn2≤1

if (rn2<PAR)

{

Select the neighboring value of already selected note, this range of choosing the neighboring value depends on b [here, b needs to be chosen randomly].

}

}

else

{

Randomly choose a value for xi between (f1, f120)

}

// If the current value of xi is found to be same with the previous value, then we need to continue the same procedure until distinct feature values appear. This is due to the fact that the same feature cannot be present twice in the probable solution.

} // end For

Evaluate the classification score for newly generated solution vector.

If the newly generated solution vector is better than the worst solution vector in HMS, then replace the worst solution vector with newly generated solution vector.

t=t+1

} // end while

Increment v by IF

} // end For

6 Results and Analysis

The experiments are carried out on 12 official scripts of India, namelyBangla, Devanagari, Gujarati, Gurumukhi, Kannada, Malayalam, Manipuri, Oriya, Tamil, Telugu, Urdu, and Roman. All the experiments are implemented in MATLAB 2010 (The Mathworks, Natick, MA, USA) under a Windows XP (Microsoft, Redmond, WA, USA) environment on an Intel Core2 Duo 2.4 GHz processor with 2 GB of RAM. A total dataset of 120 handwritten document pages (10 pages from each script) is collected from different people of varying age, sex, educational qualification, etc. These documents are digitized using an HP flatbed scanner and stored as 24-bit bitmap images. The gray-tone images may contain noisy pixels that have been removed by using Gaussian filter [14]. A set of 160, 640, and 2560 text blocks for each document of different scripts are prepared by applying second-level, third-level, and fourth-level quad-tree-based page segmentation approaches. Sample handwritten text block images of second-level decomposition for all the above-mentioned scripts are shown in Figure 4. A 3-fold cross-validation scheme is used for testing the present approach. That is, for the second-level decomposition approach, a total of 1280 text blocks are used for the training purpose and the remaining 640 text blocks are used for the testing purpose. Similarly, for third-level decomposition, 5120 text blocks are taken for the training and the remaining 2560 text blocks are used for testing the system. Again, for fourth-level decomposition, a set of 20,480 and 10,240 text blocks are utilized for training and testing the present system, respectively. The proposed approach is then evaluated using six different conventional classifiers, namely Naïve Bayes, Multi Layer Perceptron (MLP), SVM, Random Forest, Bagging, MultiClass classifier. The graphical comparison of the identification accuracies of the said classifiers achieved with Modified log-Gabor filter-based features for second-level, third-level, and fourth-level page-segmentation approaches are shown with the help of a bar chart in Figure 5. It can be seen from the figure that the best identification accuracy is achieved by the SVM classifier, which is found to be 92.19%, 90.08%, and 89.26% for second-level, third-level, and fourth-level decomposition, respectively.

Figure 4: Samples of Text Block Images, Obtained at Second-Level Quad-Tree-Based Decomposition Scheme, Written in (A) Bangla, (B) Devanagari, (C) Gujarati, (D) Gurumukhi, (E) Kannada, (F) Malayalam, (G) Manipuri, (H) Oriya, (I) Tamil, (J) Telugu, (K) Urdu, and (L) Roman scripts.
Figure 4:

Samples of Text Block Images, Obtained at Second-Level Quad-Tree-Based Decomposition Scheme, Written in (A) Bangla, (B) Devanagari, (C) Gujarati, (D) Gurumukhi, (E) Kannada, (F) Malayalam, (G) Manipuri, (H) Oriya, (I) Tamil, (J) Telugu, (K) Urdu, and (L) Roman scripts.

Figure 5: Graph Showing the Recognition Accuracies of Modified Log-Gabor Filter-Based Feature Extraction Method for Second-Level, Third-Level, and Fourth-Level Decomposition Approaches Using Six Different Classifiers.
Figure 5:

Graph Showing the Recognition Accuracies of Modified Log-Gabor Filter-Based Feature Extraction Method for Second-Level, Third-Level, and Fourth-Level Decomposition Approaches Using Six Different Classifiers.

However, it is difficult to decide the optimum set of features that contributed positively in the above script identification experiments. Thus, the HS-based FS method is tested by selecting 20%, 40%, 60%, and 80% of the original feature vector (rounding off to the lowest nearest integer). Here, the objective function is the obviously the classification accuracy, which is evaluated as follows:

(16)Classification Accuracy=Number of correctly classified text blocksTotal number of text blocks×100(%).

As the best classification accuracy has been achieved using the SVM classifier, we have chosen the same for further experimentations. The SVM classifier effectively maps pattern vectors to a high-dimensional feature space where a “best” separating hyperplane (the “maximal margin” hyperplane) is constructed. To construct an optimal hyperplane, SVM employs an iterative training algorithm to minimize an error function, which can be defined as follows:

Find W∈ℝm and b∈ℝ to

Minimize  12WTW+Ci=1Nξi

subject to constraints yi(WT ϕ(xi)+b)≥1−ξi, i=1, 2, …….,N

ξi0,i=1,2.,N,

where C is the penalty parameter of the error term, W is the vector of coefficients, b is a constant, and ξi represents parameters for handling non-separable data (inputs). The index i labels the N training cases. Note thaty∈±1 represents the class labels and xi represents the independent variables. The kernel ϕ is used to transform data from the input (independent) to the feature space. It should be noted that the larger the C, the more the error is penalized. Thus, C should be chosen with care to avoid overfitting. There are number of kernels that can be used in SVM models. These include linear, polynomial, radial basis function, and sigmoid. For the present work, we have applied polynomial kernel, which can be written as

(17)K(Xi,Xj)=(γXi.Xj+C)d

where K(Xi, Xj)=ϕ(Xi).ϕ(Xj) is the kernel function that represents a dot product of input data points mapped into the higher-dimensional feature space by transformation ϕ.

For the present work, the optimal values of δ and ρ have been experimentally set to 0.7 and 0.3, respectively. Keeping these values fixed, a comparison between the different values of HMS (HMS=5, 10, 15, and 20) varying the number of iterations (MaxItr=30, 50, and 70) and features (24, 48, 72, and 96) has been performed, which is detailed in Table 1. It can be observed from the table that for HMS=10 and MaxItr=50, the algorithm attains the highest recognition accuracies of 98.18%, 96.56%, and 94.38% for second-level, third-level, and fourth-level decomposition, respectively, which is significantly higher than the preceding outcomes. Apart from this, the optimal number of features required by the HSA is reduced to 72, which comprise almost 60% of the original feature set described in Section 3. In the present work, statistical performance measures with respect to different parameters, namely kappa statistics, mean absolute error, root mean square error, true positive rate, false positive rate, precision, recall, F-measure, and area under ROC are computed for 12 different scripts using the SVM classifier. The said measurements at the second, third, and fourth levels of page decomposition are detailed in Table 2A–C, respectively.

Table 1:

Recognition Accuracies of the Present HSA for Different Parameter Values of HMS at (A) Second Level, (B) Third Level, and (C) Fourth Level (Best Case Is Shaded in Gray and Styled in Bold).

Number of

Optimal

features

selected
MaxItrClassification accuracy (%)
HMS
5101520
(A)
Second level
243091.0692.2493.6992.5
5091.5993.1893.3592.33
7093.2993.3592.7392.39
483097.0196.9597.1797.17
5097.1696.8797.2796.99
7096.9997.3997.2797.21
723097.6597.9597.9598.01
5097.9598.1897.9598.01
7097.9597.8498.0197.95
963098.0198.0798.1298.12
5098.1898.0798.1298.18
7098.1298.1898.1298.18
(B)
tdird level243091.1892.3592.3591.62
5091.1892.1292.0291.62
7091.8692.1992.5392.15
483094.6694.5295.1094.98
5094.8994.9395.0694.98
7095.0795.1795.1095.04
723096.2196.2495.595.68
5096.2196.5695.3995.37
7095.2495.5795.2195.68
963095.6896.2495.6895.68
5095.6896.2495.6395.68
7095.6895.7595.6895.68
(C)
Fourth level243091.5592.8692.392.39
5091.5591.0692.391.18
7091.7892.4292.8692.39
483092.5593.0893.0893.25
5092.3593.0893.0593.17
7093.1994.2193.6893.25
723093.8994.2193.293.2
5093.5694.3893.293.15
7093.8993.7593.5793.09
963093.5993.6893.9893.25
5094.0993.2893.9893.39
7093.5993.6793.2593.25
Table 2:

Statistical Performance Measures along with Their Respective Means (Styled in Bold and Shaded in Gray) Achieved for Best Case of the SVM Classifier on 12 Different Scripts at (A) Second-Level, (B) Third-Level, and (C) Four-Level Page Decomposition.

ScriptStatistical performance measures
Kappa statisticsMAERMSETPRFPRPrecisionRecallF-measureAUC
(A)
Bangla0.980.16210.28190.9630.0030.9690.9630.9660.999
Devanagari0.9500.0010.9870.9500.9680.993
Gujarati0.9940.0010.9940.9940.9941.000
Gurumukhi0.9880.0030.950.9880.9810.994
Kannada1.0000.0030.9761.0000.9881.000
Malayalam1.0000.0010.9881.0000.9941.000
Manipuri0.9820.0020.9820.9820.9820.980
Oriya0.9810.0020.9690.9810.9751.000
Tamil0.9810.0020.9810.9810.9810.999
Telugu0.9810.0030.9810.9810.9810.999
Urdu0.9810.0010.980.9810.9840.999
Roman0.9810.0010.9940.9810.9871.000
Average0.980.16210.28190.9820.0020.9820.9820.9820.998
(B)
Bangla0.96220.02270.08050.9780.0030.9680.9780.9730.998
Devanagari0.9360.0020.9790.9360.9570.993
Gujarati0.9800.0050.9490.9800.9640.995
Gurumukhi0.9470.0040.9630.9470.9550.994
Kannada0.9810.0030.9750.9810.9780.997
Malayalam0.9840.0020.9810.9840.9830.999
Manipuri0.9660.0030.9660.9660.9660.995
Oriya0.9550.0030.9740.9550.9640.994
Tamil0.9890.0070.9380.9890.9630.997
Telugu0.9220.0070.9320.9220.9270.982
Urdu0.9860.0010.9870.9860.9870.999
Roman0.9640.0020.9780.9640.9710.997
Average0.96220.02270.08050.9660.0030.9660.9660.9660.995
(C)
Bangla0.93820.03270.10180.9860.0070.9370.9860.9610.994
Devanagari0.8610.0060.9380.8610.8980.968
Gujarati0.9650.0070.9330.9650.9490.990
Gurumukhi0.9180.0050.9500.9180.9340.983
Kannada0.9700.0020.9780.9700.9740.997
Malayalam0.9820.0030.9670.9820.9750.998
Manipuri0.9440.0060.9440.9440.9440.986
Oriya0.9480.0040.9590.9480.9530.990
Tamil0.9630.0120.8930.9630.9270.980
Telugu0.8820.0100.9020.8820.8920.961
Urdu0.9450.0050.9540.9450.9500.990
Roman0.9630.0020.9750.9630.9690.996
Average0.93820.03270.10180.9440.0060.9440.9440.9440.986
  1. MAE, mean absolute error; RMSE, root mean square error; TPR, true positive rate; FPR, false positive rate; AUC, area under ROC.

6.1 Sensitivity to Noise

In this subsection, the accuracy of feature subset selection using HSA has been measured by studying the impact of Gaussian noise to the handwritten script document images at all the three levels of decomposition. This is done to validate the robustness of the present methodology in the presence of noise. Presently, we have added 10%, 20%, and 30% Gaussian noises into the pages, and experimented at all the three levels of page decomposition. An example showing the addition of Gaussian noise on a sample handwritten Kannada image at second-level decomposition is illustrated in Figure 6. The Modified log-Gabor filter (as described in Section 3) is again applied to extract the original features, and the number of features is the optimized HSA. The values of parameters of HSA for which the highest identification accuracies have been obtained (in the previous subsection) are taken into consideration for the current experiment. The classification is done using the SVM classifier. A comparative study of the script identification accuracies achieved in the presence of noise at all the three levels of page decomposition are shown with the help of a bar chart in Figure 7. It is to be noted from the figure that the present methodology shows a considerable amount of resistivity to noise. The script recognition accuracies have been found to decrease only about 1%, 2%, and 3% than the present recognition accuracies for 10%, 20%, and 30% addition of Gaussian noise at second-, third-, and fourth-level decompositions, respectively. Thus, it can be concluded from this experiment that the present methodology is less sensitive to noise to a large extent.

Figure 6: Illustration of (A) Original Handwritten Kannada Image at Second-Level Decomposition, and Addition of Gaussian Noise at (B) 10%, (C) 20%, and (D) 30%.
Figure 6:

Illustration of (A) Original Handwritten Kannada Image at Second-Level Decomposition, and Addition of Gaussian Noise at (B) 10%, (C) 20%, and (D) 30%.

Figure 7: Comparison of the Impact of Added Gaussian Noise to Handwritten Script Identification Accuracies (Evaluated Using the SVM Classifier) at All Three Levels of Page Decomposition.
Figure 7:

Comparison of the Impact of Added Gaussian Noise to Handwritten Script Identification Accuracies (Evaluated Using the SVM Classifier) at All Three Levels of Page Decomposition.

6.2 Statistical Significance Tests

The statistical significance of the experimental setup has also been measured as an essential part for validating the performance of the multiple classifiers using multiple datasets. For statistical comparison of multiple classifiers, two or more learning algorithms are first trained and tested on suitable datasets and then evaluated using the classification accuracies. Datasets are randomly divided into small datasets with different sample sizes. Performances of different classifiers are then carried out for each randomly created dataset. The only requirement for performing non-parametric tests is that the compiled results provide reliable estimates of the performances of the classification algorithms on each dataset. In the usual experimental setups, these numbers come from cross-validation or from repeated stratified random splits onto training and testing datasets. The term “sample size” refers to the number of datasets used, not the number of training/testing samples drawn from each individual set. The sample size can therefore be as small as five and is usually well below 30. In the present work, we have performed some statistical significance tests for validating the results of the best classifier among six different classifiers at second-level decomposition. These tests have been performed at two stages: (a) before applying HSA and (b) after applying HSA. The theoretical discussions related to these tests are already described in one of our previous works [16].

Case I: Application of Statistical Significance Tests prior to HSA

A safe and robust non-parametric Friedman test [14] has been conducted along with the corresponding post-hoc tests. For the experimentation, the number of datasets (N) and the number of classifiers (k) are set as 12 and 6, respectively. These datasets are chosen randomly from the test set. The performances of the classifiers on different datasets are shown in Table 3. On the basis of these performances, the classifiers are then ranked for each dataset separately, the best-performing algorithm gets the rank 1, the second best gets the rank 2, and so on (see Table 3). Average ranks are assigned in case of ties.

Table 3:

Recognition Accuracies of Six Classifiers and Their Corresponding Ranks Using 12 Different Datasets Before Performing HSA (Ranks in Parentheses Are Used for Performing Friedman Test).

DatasetClassifier
Naïve bayes (%)MLP (%)SVM (%)Random forest (%)Bagging (%)Multiclass classifier (%)
#190 (5)93 (3)98 (1)88 (6)91 (4)95 (2)
#288 (6)93 (2)96 (1)91 (3)90 (4)89 (5)
#393 (3)91 (4)95 (2)90 (5.5)90 (5.5)97 (1)
#491 (4.5)95 (2)97 (1)93 (3)90 (6)91 (4.5)
#590 (6)95 (2)96 (1)93 (4)91 (5)95 (3)
#696 (2)90 (5.5)98 (1)90 (5.5)91 (4)93 (3)
#793 (2)91 (5)95 (1)91 (5)92 (3)91 (5)
#888 (6)95 (3)96 (1.5)92 (4)90 (5)96 (1.5)
#995 (2)94 (3)97 (1)90 (6)91 (4.5)91 (4.5)
#1092 (2)86 (5.5)94 (1)90 (3)86 (5.5)88 (4)
#1190 (4.5)90 (4.5)95 (1)93 (2)91 (3)88 (6)
#1289 (6)94 (2.5)96 (1)94 (2.5)90 (5)92 (4)
Mean rankR1=4.08R2=3.5R3=1.125R4=4.125R5=4.54R6=3.625
  1. The mean ranks scored by individual classifiers are shaded in gray and styled in bold.

Let rji be the rank of the j-th classifier on i-th dataset. Then, the mean of the ranks of the j-th classifier over all the N datasets will be computed as

(18)Rj=1Ni=1Nrji

The null hypothesis states that all the classifiers are equivalent and so their ranks Rj should be equal. To justify it, the Friedman statistic [16] is computed as follows:

(19)χF2=12Nk(k+1)[jRj2k(k+1)24]

Under the current experimentation, this statistic is distributed according to χF2 with k−1 (=5) degrees of freedom. Using Eq. (19), the value of χF2 is calculated as 25.47. From the table of critical values (see any standard statistical book), the value of χF2 with 5 degrees of freedom is 11.07 for α=0.05 (where α is known as the level of significance). It can be seen that the computed χF2 differs significantly from the standard χF2. Thus, the null hypothesis is rejected.

Iman et al. [16] have derived a superior statistic using the following formula:

(20)FF=(N1)χF2N(k1)χF2

FF is distributed according to the F-distribution with k−1 (=5) and (k−1)(N−1) (=55) degrees of freedom. Using Eq. (20), the value of FF is calculated as 8.115. The critical value of F (5, 55) for α=0.05 is 4.43 (see any standard statistical book), which shows a significant difference between the standard and calculated values of FF. Thus, both the Friedman and Iman et al. statistics reject the null hypothesis.

As the null hypothesis is rejected, a post-hoc test known as the Nemenyi test [16] is carried out for pair-wise comparisons of the best- and worst-performing classifiers. The performances of two classifiers are significantly different if the corresponding average ranks differ by at least the critical difference (CD), which is expressed as

(21)CD=qαk(k+1)6N

For Nemenyi’s test, the value of q0.05 for six classifiers is 2.850 (see Table 5a of Ref. [6]). Thus, the CD is calculated as 2.85067612, i.e. 2.17 using Eq. (21). As the difference between the mean ranks of the best and worst classifier is much greater than the CD, we can conclude that there is a significant difference between the performing ability of the classifiers. For comparing all the classifiers with a control classifier (here SVM), we have applied the Bonferroni-Dunn test [16]. For this test, CD is calculated using the same Eq. (21). However, here, the value of q0.05 for six classifiers is 2.576 (see Table 5b of Ref. [6]). Thus, the CD for the Bonferroni-Dunn test is calculated as 2.57667612, i.e. 1.96. As the difference between the mean ranks of any classifier and SVM is always greater than CD, the chosen control classifier performs significantly better than other classifiers. A graphical representation of the said post-hoc tests for comparison of six different classifiers is shown in Figure 8.

Figure 8: Graphical Representation of Comparison of Multiple Classifiers for (A) Nemenyi’s Test and (B) Bonferroni-Dunn’s Test Before Applying HSA.
Figure 8:

Graphical Representation of Comparison of Multiple Classifiers for (A) Nemenyi’s Test and (B) Bonferroni-Dunn’s Test Before Applying HSA.

After performing the above-mentioned statistical significance tests over the 12 datasets and six classifiers, it can be concluded that the SVM classifier outperforms all other classifiers at second-level decomposition. Similarly, the performances of these tests at the remaining two levels also confirm that the SVM classifier is found to be statistically significant than the other five classifiers.

Case II: Application of Statistical Significance Tests subsequent to HSA

Similar to the preceding case, the number of datasets (N) and the number of classifiers (k) are chosen as 12 and 6, respectively. The performances of the classifiers on different datasets are shown in Table 4.

Table 4:

Recognition Accuracies of Six Classifiers and Their Corresponding Ranks Using 12 Different Datasets after Applying HSA (Ranks in Parentheses Are Used for Performing Friedman Test).

DatasetClassifier
Naïve bayes (%)MLP (%)SVM (%)Random forest (%)Bagging (%)Multiclass classifier (%)
#193 (6)99 (2.5)100 (1)97 (5)99 (2.5)98 (4)
#299 (2.5)99 (2.5)100 (1)96 (5.5)96 (5.5)97 (4)
#398 (4.5)93 (6)100 (1)99 (2.5)98 (4.5)99 (2.5)
#493 (6)98 (3)99 (1.5)99 (1.5)97 (4)96 (5)
#595 (6)99 (2.5)100 (1)99 (2.5)97 (5)98 (4)
#699 (2.5)96 (6)100 (1)97 (5)98 (4)99 (2.5)
#791 (6)99 (3)100 (1)99 (3)99 (3)98 (5)
#893 (6)98 (3.5)100 (1)98 (3.5)96 (5)99 (2)
#995 (6)100 (1.5)100 (1.5)97 (5)99 (3.5)99 (3.5)
#1099 (2.5)99 (2.5)100 (1)97 (5.5)97 (5.5)98 (4)
#1194 (6)98 (4.5)100 (1)99 (2.5)98 (4.5)99 (2.5)
#1298 (3.5)97 (5.5)100 (1)98 (3.5)99 (2)97 (5.5)
Mean rankR1=4.79R2=3.58R3=1.083R4=3.75R5=4.08R6=3.708
  1. The mean ranks scored by individual classifiers are shaded in gray and styled in bold.

Using the Friedman statistic given in Eq. (19), the value of χF2 is calculated as 27.057. It can be seen that the computed χF2 differs significantly from the standard χF2. Using the Iman et al. statistic written in Eq. (20), the tabulated value of FF is calculated as 9.035, which is much greater than the observed value of FF for (5,55) degrees of freedom. Thus, for both tests, the null hypothesis is rejected.

For Nemenyi’s test, the value of CD is calculated as 2.85067612, i.e. 2.17, using the same Eq. (21). As the difference between the mean ranks of the best and worst classifier is much greater than the CD, we can conclude that there is a significant difference between the performing ability of the classifiers. For comparing all the classifiers with a control classifier using the Bonferroni-Dunn test, the value of CD is calculated as 2.57667612, i.e. 1.96. As the difference between the mean ranks of any classifier and SVM is always greater than CD, the chosen control classifier performs significantly better than other classifiers. A graphical representation of the above-mentioned post-hoc tests for comparison of six different classifiers is shown in Figure 9.

Figure 9: Graphical Representation of Comparison of Multiple Classifiers for (A) Nemenyi’s Test and (B) Bonferroni-Dunn’s Test After Applying HSA.
Figure 9:

Graphical Representation of Comparison of Multiple Classifiers for (A) Nemenyi’s Test and (B) Bonferroni-Dunn’s Test After Applying HSA.

After performing the above-mentioned statistical significance tests over the 12 datasets and six classifiers, it can be concluded that the SVM classifier outperforms all other classifiers at second-level decomposition. Similarly, the performance of these tests at the remaining two levels also confirms that the SVM classifier is found to be statistically significant than the other five classifiers.

6.3 Comparison with Popular Meta-heuristic Algorithms

For the present work, the optimization performance of HSA is also compared with four popular meta-heuristic algorithms, namely GA [15], TS [36], PSO [19], and ACO [8]. The GA is implemented to contain 20 chromosomes and made to run for 100 generations in each trial. The crossover rate and mutation rate are set to 1.0 and 0.1, respectively. The TS algorithm contains a list size of 10 tabus with a neighborhood size of 20, and this algorithm is made to run for 100 iterations. For the PSO algorithm, the number of particles used is 20. The two factors rand1 and rand2 are random numbers between (0, 1), whereas c1 and c2 are acceleration (learning) factors, with c1=c2=2. The inertia weight w is taken as 0.8, and the maximum number of iterations used in our PSO is 100. Lastly, the ACO algorithm uses the following parameter values: (a) maximum number of iterations=100, (b) evaporation rate=0.3, and (c) maximum local search=10. The comparison is measured with respect to three parameters: (a) optimal feature subset, (b) classification accuracy, and (c) execution time. The optimal feature subset denotes the optimal number of features required to achieve the maximum recognition accuracy. The execution time is measured as the time required to run the given optimization algorithm until the corresponding classification accuracy is realized using the SVM classifier with 3-fold cross-validation scheme. The comparison enlisted in Table 5 confirms that the present HSA not only achieves the highest classification accuracy but also involves the minimum computational time. Again, two well-known statistical feature dimensionality reduction techniques, named PCA [28] and greedy attribute search (GAS) [37], are also implemented and evaluated on the same dataset. From the experiments (shown in Table 6), it is found that both PCA and GAS utilize a feature dimensionality of 70 yielding recognition accuracies of 94.58%, 93.97%, 92.34% and 93.7%, 94.01%, 90.25% at second, third, and fourth levels of page segmentations, respectively. Although PCA selects fewer number of features as well as requires less computational time than HSA, HSA achieves almost about 4%, 3%, and 2% more recognition accuracies than the PCA at second-, third-, and fourth-level decomposition, respectively. Therefore, it is rather clear from the preceding observations that the present HSA outperforms some previously pioneered optimization algorithms used mainly for FS.

Table 5:

Comparison of HSA with the Best Case of the Optimization Performances of GA, TS, PSO, and ACO Algorithms.

Optimization algorithmsPage segmentation levelOptimum feature subsetClassification accuracy (%)Execution time (s)
GA [15]2nd8092.321506.85
3rd93.542059.68
4th91.022945.36
TS [36]2nd8092.651521.35
3rd93.762038.96
4th92.452852.28
PSO [19]2nd7593.551355.29
3rd93.981968.03
4th90.462571.54
ACO [8]2nd7595.261425.36
3rd92.852098.22
4th91.642464.55
HS2nd7298.181203.49
3rd96.561755.75
4th94.382236.20
  1. Highest recognition accuracy is shaded in gray and styled in bold.

Table 6:

Comparison of HSA with PCA and GAS Techniques.

Statistical feature dimensionality reduction techniquesPage segmentation levelOptimum feature subsetClassification accuracy (%)Execution time (s)
PCA [27]2nd7094.581115.95
3rd93.971568.34
4th92.342055.6
GAS [37]2nd7093.71328.55
3rd94.011945.27
4th90.252385.09
HS2nd7298.181203.49
3rd96.561755.75
4th94.382236.20
  1. Highest recognition accuracy is shaded in gray and styled in bold.

7 Conclusion

This paper has investigated the HSA for the selection of the optimal feature subset in handwritten script identification problem. The proposed HSA evolved upon the HM, which adaptively stores updated solutions during the evolution. This algorithm has been evaluated on a limited dataset of script blocks written in 12 official Indian scripts. For obtaining the optimal classification accuracy, we have conducted a series of experiments to test the efficiency of HSA with different parameter settings. The proposed HSA achieves the highest classification accuracy of 98.18% at the second-level block decomposition approach, which is almost about 6% more than the earlier attained result while only considering almost 60% of the original feature set. The improved result justifies the need for selecting the subset of designed features for this type of pattern classification problem, where both local and global features are computed without knowing the exact importance of the features from all regions of a particular script block image. Although the results show a rather encouraging trend, much work can still be done to further increase the potential of the present work.

In particular, investigations regarding how the algorithm parameters may be better tuned are necessary. It would be useful to develop a better stopping criterion, which will, in turn, lessen the computation time. In addition, a method may be devised to dynamically specify the size of the HM, which affects the convergence rate and search parameters.

Acknowledgments

The authors are thankful to the Center for Microprocessor Application for Training Education and Research and Project on Storage Retrieval and Understanding of Video for Multimedia of Computer Science and Engineering Department, Jadavpur University, for providing infrastructure facilities during the progress of the work. The current work reported here was partially funded by University with Potential for Excellence, Phase-II, UGC, Government of India.

Bibliography

[1] M. H. Aghdam, N. Ghasem-Aghaee and M. E. Basiri, Text feature selection using ant colony optimization, Expert Syst. Appl.36 (2009), 6843–6853.10.1016/j.eswa.2008.08.022Search in Google Scholar

[2] I. Ahmad, Feature selection using particle swarm optimization in intrusion detection, Int. J. Distrib. Sensor Netw.9 (2015), 1–8.10.1155/2015/806954Search in Google Scholar

[3] P. Bermejo, J. Gámez and J. Puerta, A GRASP algorithm for fast hybrid (filter-wrapper) feature subset selection in high dimensional datasets, Pattern Recogn. Lett.32 (2011), 701–711.10.1016/j.patrec.2010.12.016Search in Google Scholar

[4] N. Das, R. Sarkar, S. Basu, M. Kundu and M. Nasipuri, A genetic algorithm based region sampling for selection of local features in handwritten digit recognition application, Appl. Soft Comput.12 (2012), 1592–1606.10.1016/j.asoc.2011.11.030Search in Google Scholar

[5] C. De Stefano, F. Fontanella, C. Marrocco and A. Scotto di Freca, AGA-based feature selection approach with an application to handwritten character recognition, Pattern Recogn. Lett.35 (2014), 130–141.10.1016/j.patrec.2013.01.026Search in Google Scholar

[6] J. Demsar, Statistical comparisons of classifiers over multiple data sets, J. Mach. Learn. Res.7 (2006), 1–30.Search in Google Scholar

[7] S. Ding, Feature selection based F-score and ACO algorithm in support vector machine, in: Proc. of 2nd International Symposium on Knowledge Acquisition and Modeling, 2009.10.1109/KAM.2009.137Search in Google Scholar

[8] M. Dorigo, The ant system: optimization by a colony of cooperating agents, IEEE Trans. Syst. Man Cybern. B26 (1996), 1–13.10.1109/3477.484436Search in Google Scholar PubMed

[9] M. El Alami, A filter model for feature subset selection based on genetic algorithm, Knowl. Based Syst.22 (2009), 356–362.10.1016/j.knosys.2009.02.006Search in Google Scholar

[10] D. J. Field, Relations between the statistics of natural images and the response properties of cortical cells, J. Opt. Soc. Am. A4 (1987), 2379–2394.10.1364/JOSAA.4.002379Search in Google Scholar PubMed

[11] R. Forsati, A. Moayedikia and A. Keikha, A novel approach for feature election based on the bee colony optimization, Int. J. Comput. Appl.43 (2012), 13–16.10.5120/6122-8329Search in Google Scholar

[12] Z. W. Geem, Optimal cost design of water distribution networks using harmony search, Dissertation, Korea University, 2000.Search in Google Scholar

[13] Z. W. Geem, J. H. Kim and G. V. Loganathan, A new heuristic optimization algorithm: harmony search, Simulation76 (2001), 60–68.10.1177/003754970107600201Search in Google Scholar

[14] R. C. Gonzalez and R. E. Woods, Digital Image Processing, vol. I. Prentice-Hall, India, 1992.Search in Google Scholar

[15] J. H. Holland, Adaptation in Natural and Artificial Systems, University of Michigan Press, New York, 1975.Search in Google Scholar

[16] R. L. Iman and J. M. Davenport, Approximations of the critical region of the Friedman statistic, in: Communications in Statistics, pp. 571–595, 1980.10.1080/03610928008827904Search in Google Scholar

[17] M. Kabir, M. Islam and K. Murase, A new local search based hybrid genetic algorithm for feature selection, Neurocomputing74 (2011), 2914–2928.10.1016/j.neucom.2011.03.034Search in Google Scholar

[18] S. Kashef and H. Nezamabadi-pour, An advanced ACO algorithm for feature subset selection, Neurocomputing147 (2015), 271–279.10.1016/j.neucom.2014.06.067Search in Google Scholar

[19] J. Kennedy and R. C. Eberhart, Particle swarm optimization, in: Proc. of IEEE International Conference on Neural Networks, pp. 1942–1948, 1995.Search in Google Scholar

[20] V. Kumar, J. K. Chhabra and D. Kumar, Automatic unsupervised feature selection using gravitational search algorithm, IETE J. Res.61 (2014), 22–31.10.1080/03772063.2014.987702Search in Google Scholar

[21] V. Kumar, J. K. Chhabra and D. Kumar, Parameter adaptive harmony search algorithm for unimodal and multimodal optimization problems, J. Comput. Sci.5 (2014), 144–155.10.1016/j.jocs.2013.12.001Search in Google Scholar

[22] V. Kumar, J. K. Chhabra and D. Kumar, Variance based harmony search algorithm for unimodal and multimodal optimization problems with application to clustering, Cybernet. Syst.45 (2014), 486–511.10.1080/01969722.2014.929349Search in Google Scholar

[23] V. Kumar, J. K. Chhabra and D. Kumar, Automatic data clustering using parameter adaptive harmony search algorithm and its application to image segmentation, J. Intell. Syst.25 (2015), 595–610.10.1515/jisys-2015-0004Search in Google Scholar

[24] K. S. Lee and Z. W. Geem, A new meta-heuristic algorithm for continuous engineering optimization: harmony search theory and practice, Comput. Methods Appl. Mech. Eng.194 (2004), 3902–3933.10.1016/j.cma.2004.09.007Search in Google Scholar

[25] S. W. Lin, K. C. Ying, S. C. Chen and Z. J. Lee, Particle swarm optimization for parameter determination and feature selection of support vector machines, Expert Syst. Appl.35 (2008), 1817–1824.10.1016/j.eswa.2007.08.088Search in Google Scholar

[26] F. G. Mohammadi and M. S. Abadeh, Imagesteg analysis using a bee colony based feature selection algorithm, Eng. Appl. Artif. Intell.31 (2014), 35–43.10.1016/j.engappai.2013.09.016Search in Google Scholar

[27] M. Nazir, A. Majid-Mirza and S. Ali-Khan, PSO-GA based optimized feature selection using facial and clothing information for gender classification, J. Appl. Res. Technol.12 (2014), 145–152.10.1016/S1665-6423(14)71614-1Search in Google Scholar

[28] K. Pearson, On lines and planes of closest fit to systems of points in space, Philos. Mag.2 (1901), 559–572.10.1080/14786440109462720Search in Google Scholar

[29] A. Roy, N. Das, R. Sarkar, S. Basu, M. Kundu and M. Nasipuri, Region selection in handwritten character recognition using artificial bee colony optimization, in: Proc. of 3rdInternational Conference on Emerging Applications of Information Technology (EAIT), pp. 183–186, 2012.10.1109/EAIT.2012.6407891Search in Google Scholar

[30] A. Roy, N. Das, R. Sarkar, S. Basu, M. Kundu and M. Nasipuri, An axiomatic fuzzy set theory based feature selection methodology for handwritten numeral recognition, in: Proc. of 48th Annual Convention of Computer Society of India, vol. I, AISC248, pp. 133–140, 2014.10.1007/978-3-319-03107-1_16Search in Google Scholar

[31] A. K. Saxena and V. K. Dubey, A survey on feature selection algorithms, Int. J. Recent Innov. Trends Comput. Commun.3 (2015), 1895–1899.10.17762/ijritcc2321-8169.150431Search in Google Scholar

[32] D. A. A. G. Singh, E. Jebamalar Leavline, K. Valliyappan and M. Srinivasan, Enhancing the performance of classifier using particle swarm optimization (PSO)-based dimensionality reduction, Int. J. Energ. Inform. Commun.6 (2015), 19–26.10.14257/ijeic.2015.6.5.03Search in Google Scholar

[33] P. K. Singh, I. Chatterjee and R. Sarkar, Page-level handwritten script identification using modified log-Gabor filter based features, in: Proc. of 2nd IEEE International Conference on Recent Trends in Information Systems (ReTIS), pp. 225–230, 2015.10.1109/ReTIS.2015.7232882Search in Google Scholar

[34] P. K. Singh, R. Sarkar and M. Nasipuri, Offline script identification from multilingual Indic-script documents: a state-of-the-art, Comput. Sci. Rev.15–16 (2015), 1–28.10.1016/j.cosrev.2014.12.001Search in Google Scholar

[35] R. Sivagaminathan and S. Ramakrishnan, A hybrid approach for feature subset selection using neural networks and ant colony optimization, Expert Syst. Appl.33 (2007), 49–60.10.1016/j.eswa.2006.04.010Search in Google Scholar

[36] C. S. Sung and H. W. Jin, A tabu-search-based heuristic for clustering, Pattern Recogn.33 (2000), 849–858.10.1016/S0031-3203(99)00090-4Search in Google Scholar

[37] H. Vafaie and I. F. Imam, Feature selection methods: genetic algorithms vs. greedy-like search, in: Proc. of 3rd International Fuzzy Systems and Intelligent Control Conference, Louisville, KY, March 1994.Search in Google Scholar

[38] V. M. Vanishree, Provision for linguistic diversity and linguistic minorities in India, Available at: http://www.languageinindia.com/feb2011/vanishreemastersfinal.pdf. Retrieved 5 February, 2016.Search in Google Scholar

[39] S. Vieira, L. Mendonc, G. Farinha and J. Sousa, Modified binary PSO for feature selection using SVM applied to mortality prediction of septic patients, Appl. Soft Comput.13 (2013), 3494–3504.10.1016/j.asoc.2013.03.021Search in Google Scholar

[40] L. T. Vinh, S. Lee, Y. T. Park and B. J. d’Auriol, A novel feature selection method based on normalized mutual information, Appl. Intell.37 (2010), 100–120.10.1007/s10489-011-0315-ySearch in Google Scholar

[41] Z. Zeng, H. Zhang, R. Zhang and Y. Zhang, A hybrid feature selection method based on rough conditional mutual information and naive Bayesian classifier, ISRN Appl. Math. (2014), 1–11. Article ID: 382738.10.1155/2014/382738Search in Google Scholar

Received: 2016-6-1
Published Online: 2017-3-16
Published in Print: 2018-7-26

©2018 Walter de Gruyter GmbH, Berlin/Boston

This article is distributed under the terms of the Creative Commons Attribution Non-Commercial License, which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Downloaded on 10.6.2024 from https://www.degruyter.com/document/doi/10.1515/jisys-2016-0070/html
Scroll to top button