Keywords

1 Introduction

Biometric authentication is a valuable technique for the automated recognition of individuals. The palmprint recognition (PPR) biometric technique is regarded as reliable and efficient as it has a high recognition rate but is relatively low cost and user friendly. The methods used for PPR can roughly be categorised according to palm features on which the scheme is based. (1) Lines and minutiae. The success of this technique is dependent upon very high resolution of at least 500 dpi [2] to capture the main lines and minutiae of the palm [1]. Edge detectors, such as the Sobel filter [3] or templates [4], can be used to extract the principal lines within the print. Detection of minutiae points is achieved by locating the ridge bifurcation points and endings [5]. (2) Texture. Local binary patterns (LBP) schemes use texture to collect the global pattern of lines [6], ridges and wrinkles. Examples of LBP that accurately recognise, even very low-resolution palm-print images (PPIs) include Gabor transform [7], Palmcode [8], Fusion code [9], Competitive code [10] and Contour Code [11]. (3) Appearance. There are a number of techniques that are appearance based, including linear discriminant analysis (LDA) [12], principal component analysis (PCA) [1], kernel discriminant analysis (KDA) [13] and kernel PCA (KPCA) [13]. The PCA mixture model (PCAMM) is a generative-model-based approach that can also be used to discern palm appearance [14]. Schemes may also combine a mixture of line and minutiae, edge detection or appearance techniques. These hybrid schemes are considered superior as another can compensate limitations in one scheme.

In [15], Fei et al. introduced a method that used a double half-orientation bank of half-Gabor filters, designed for half-orientation extraction. A double-orientation code based on Gabor filters and nonlinear matching scheme was described in [15]. The methods presented in [15, 16] were evaluated using multispectral palmprints from the MS-PolyU database, a proposed method for palmprint recognition (PPR) based on BSIF and Histogram of oriented gradients (HOG) [17]. The palmprint recognition scheme presented in this study combines a novel spiral feature extraction with a sparse representation of binarized statistical image features (BSIF). BSIF are comparable to LBP, with exception to the method by which filters are learned [18]. In contrast to LBP, where filters are manually predefined, the texture descriptors of BSIF are learned from natural images; a binary string is created from a PPI by convolving with filters. In this study, a K-nearest-neighbour (KNN) classifier [19] was used to conduct subject verification on the fusion features acquired.

The main contributions of this paper are summarized as follows:

  • A new PPR method is proposed based on feature selection from a fusion of BSIF and spiral features.

  • Extensive experiments are carried out on palm-print databases (PPDBs): the PolyU contact PPDB with 356 subjects [20].

  • Comprehensive analysis is performed by comparing the proposed scheme with nine different state-of-the-art contemporary schemes based on BSIF-SRC [23], LBP [14], palmcode [8], fusion code [9], the Gabor transform with KDA [7], A double-orientation code [15, 16], BSIF and HOG [17] and the Gabor transform with sparse representation [23].

The rest of this paper is structured as follows. Section 2 discusses the proposed scheme for robust PPR. Section 3 discusses the experimental setup, protocols, and results. Section 4 draws conclusions.

2 Spiral Local Image Features

The proposed method pipeline for PPR based on BSIF and spiral features is depicted in Fig. 1.

Fig. 1.
figure 1

Proposed method pipeline

Effective PPR demands accurate ROI extraction, the identification of a region of the palm print that is rich in features, such as principal lines, ridges and wrinkle, which is then extracted. Using the algorithm devised by Zhang et al. [22], which computes the centre of mass and valley regions to align the palm print.

2.1 Moment Statistics

Statistical moments were used to extract the spiral image feature. Starting at the central block, the image was divided into N blocks (Each block has same size such as BSIF Filter) and based on the algorithm illustrated in Fig. 3 the route was followed as depicted in Fig. 2. The spiral feature vector of a block \( k \) is: \( s_{k} = \left[ {w_{k} \times \left( {\text{var} (x_{k} ,y_{k} ),Moy(x_{k} ,y_{k} ),skew(x_{k} ,y_{k} ),kurt(x_{k} ,y_{k} )} \right)} \right] \)

Fig. 2.
figure 2

Spiral feature extraction scheme

where \( w_{k} \) represents the weight of the block \( k \) in the image according to spiral ordering. The final spiral feature vector is:

$$ S = \left[ {\begin{array}{*{20}c} {w_{1} \times \left( {\text{var} (x_{1} ,y_{1} ),Moy(x_{1} ,y_{1} ),skew(x_{1} ,y_{1} ),kurt(x_{1} ,y_{1} )} \right), \ldots ,} \\ {w_{N} \times \left( {\text{var} (x_{N} ,y_{N} ),Moy(x_{N} ,y_{N} ),skew(x_{N} ,y_{N} ),kurt(x_{N} ,y_{N} )} \right)} \\ \end{array} } \right] $$

where N is the number of blocks in the image

2.2 Binarized Statistical Image Features

BSIF was first put forward by Kannala and Rahtu [18]. The method represents a binary code string for the pixels of a given image; the code value of a pixel is considered as a local descriptor of the image surrounding the pixel. Given an image \( {\text{I}}_{\text{p}} \) and a linear filter \( {\text{W}}_{\text{i}} \) of the same size, the filter response \( {\text{R}}_{\text{i}} \) is found by

$$ R_{i} = \sum\nolimits_{m,n} {I_{p} (m,n)W_{i} (m,n)} , $$
(1)

where m and n denote the size of the PPI patch, \( W_{i} \) denotes the number of linear filters \( \forall i = \left\{ {1,2, \ldots \ldots .,n} \right\} \) whose response can be calculated and binarized to obtain the binary string as follows [24]:

$$ b_{i} = \left\{ {\begin{array}{*{20}c} 1 & {if\text{ }R_{i} \succ 0} \\ 0 & {otherwise} \\ \end{array} } \right.. $$
(2)

The BSIF codes are presented as a histogram of pixel binary codes, which can effectively distinguish the texture features of the PPI. The filter size and the length of bit strings are important to evaluate effectively the BSIF descriptor for palm-print verification.

In this study, eight different filter sizes (3 × 3, 5 × 5, 7 × 7, 9 × 9, 11 × 11, 13 × 13, 15 × 15 and 17 × 17) with four different bit lengths (6, 7, 9 and 11) were assessed (see Fig. 3). The 17 × 17 filter with an 11-bit length was selected based upon the superior experimental accuracy achieved with this combination.

Fig. 3.
figure 3

Samples from PolyU database and corresponding BSIF codes

2.3 Global Features

Compared to score-level fusion, feature-level fusion has a faster response time. The drawback of the system that has limited its widespread uptake is that it struggles to fuse incompatible feature vectors derived from multiple modalities. Creating a linked series of extracted features is the simplest type of feature-level fusion, but concatenating two feature vectors may result in a large feature vector, leading to the ‘curse of dimensionality’ problem. To offset the poor learning performance of the high-dimensional fused feature vector data, PCA is used.

Therefore, the range of all features should be normalized so that each feature contributes approximately proportionately to the final distance. In this study, the Min–Max normalisation scheme was used to normalise the feature vectors in the range [0.1]. By concatenating the normalised feature vectors of BSIF and spiral features into a single fused vector, the definitive fused vector is achieved. Let the normalized feature vectors of an image I be \( E = \left[ {e_{1} \ldots e_{k} \ldots e_{N} } \right] \) and \( S = \left[ {s_{1} \ldots s_{k} \ldots s_{N} } \right] \) respectively for BSIF and spiral extraction. Then, the fused vector is represented as:

$$ Fused_{vector} = \left[ {e_{1} \ldots e_{N} \, s_{1} \ldots s_{N} } \right] $$
(3)

To find the optimal projection bases, PCA draws upon statistical distribution of the set of the given features to generate new features [23]. This method aims to locate the projection of the feature vector on a set of basis vectors, in order to create the new feature vector.

2.4 Extreme Learning Machine

Single-hidden layer feedforward neural networks have often been learnt by utilizing ELM [17]. The iterative tuning of obscured nodes is not necessary after random initialization using ELM has been performed. Thus, learning must occur only for the input weight parameters. If \( (x_{j} ,y_{j} ),j = \left[ {1, \ldots ,q} \right] \) indicates A as the training sample with \( x_{j} \in R^{M} \) and \( y_{j} \in R^{M} \), then the following equation can be used for the ELM’s output function with L obscured neurons:

$$ f_{l} \left( x \right) = \sum\limits_{i = 1}^{L} {g_{i} } w_{i} (x) =\Omega (x)G. $$
(4)

The \( m\text{ > }1 \) output nodes associated with the \( L \) hidden nodes by the output weight vector \( G = \left[ {g_{1} , \ldots ,g_{L} } \right] \). With the nonlinear activation function represented by \( \Omega (x) = \left[ {w_{1} (x), \ldots ,w_{L} (x)} \right] \).

3 Experimental Results

3.1 Palmprint Databases

To test the validity of the scheme, exhaustive experiments were conducted using data available famous PPDBs: The PolyU database contains 7,752 palmprint images collected from 386 palms of 193 individuals. Of them, 131 individuals are males and the rest are females. The palmprint images were collected in two sessions with the average interval over two months. In each session, about 10 palmprint images were captured from each palm. So there are 386 classes of palm in the PolyU database, each of which contains about 20 palmprint images. The images in the PolyU database were cropped to a size of 128 × 128 pixels [20, 21].

3.2 Palmprint Identification

For this study, the BSIF histogram and spiral feature vectors were combined to form the last feature vector. To detect the best results, BSIF filters with different sizes and bits were chosen. The accuracy is shown in Fig. 4. The probe samples are sorted so that they are associated with the class to which they are most similar. For this study, based on our previous work [17] and to validate our identification algorithms we have calculated the Recognition Rate. For PolyU (352 * 10 * 2 images) we have taken 10 images of one user for training and remaining 10 for testing.

Fig. 4.
figure 4

Average accuracy for PolyU palm-print databases

The 17 × 17 filter with an 11-bit length was selected based on the superior experimental accuracy achieved with this combination. The rate of errors in the identification process is used to estimate the performance of palmprint identification. The performance of the fused SPIRAL and BSIF features are applied to different databases is presented in Table 1. It shows that, compared to other feature extraction methods [7,8,9, 14,15,16,17, 23, 25], the experimental method provides a superior performance, with an EER from 0 to 2.06. The performance of the method applied on the PolyU PPDB is presented in Table 1. It shows that compared to other feature extraction methods the experimental method has a superior performance, with an EER of 2.06%. This is almost half the error rate of scheme [23], which of all the single BSIF features schemes included, has the next lowest ERR. This reinforces the applicability of the proposed scheme for PPR.

Table 1. Performance of the proposed method on the PolyU palm-print database

4 Conclusion

The accuracy and reliability of PPR is dependent upon the precision of the features represented. In this paper, a novel PPR approach is presented that uses extracted spiral features that are fused with BSIF. The experimental method was validated by conducting extensive tests on PolyU database. The performance results were compared to the performance results of six well-established state-of-the-art schemes. The results demonstrate that the experimental scheme was comparable or superior to the other methods, justifying it as being an efficient and robust tool for accurate PPR. In the future, we need to do more research on fusing two biometric modalities, which are the palmprint and iris to extract more discriminative information for bimodal identification system.