Please use this identifier to cite or link to this item:
http://theses.ncl.ac.uk/jspui/handle/10443/2128
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ahmad, Muhammad Imran. | - |
dc.date.accessioned | 2014-03-03T16:36:20Z | - |
dc.date.available | 2014-03-03T16:36:20Z | - |
dc.date.issued | 2013 | - |
dc.identifier.uri | http://hdl.handle.net/10443/2128 | - |
dc.description | Thesis | en_US |
dc.description.abstract | Multimodal biometric systems that integrate the biometric traits from several modalities are able to overcome the limitations of single modal biometrics. Fusing the information at an earlier level by consolidating the features given by different traits can give a better result due to the richness of information at this stage. In this thesis, three novel methods are derived and implemented on face and palmprint modalities, taking advantage of the multimodal biometric fusion at feature level. The benefits of the proposed method are the enhanced capabilities in discriminating information in the fused features and capturing all of the information required to improve the classification performance. Multimodal biometric proposed here consists of several stages such as feature extraction, fusion, recognition and classification. Feature extraction gathers all important information from the raw images. A new local feature extraction method has been designed to extract information from the face and palmprint images in the form of sub block windows. Multiresolution analysis using Gabor transform and DCT is computed for each sub block window to produce compact local features for the face and palmprint images. Multiresolution Gabor analysis captures important information in the texture of the images while DCT represents the information in different frequency components. Important features with high discrimination power are then preserved by selecting several low frequency coefficients in order to estimate the model parameters. The local features extracted are fused in a new matrix interleaved method. The new fused feature vector is higher in dimensionality compared to the original feature vectors from both modalities, thus it carries high discriminating power and contains rich statistical information. The fused feature vector also has larger data points in the feature space which is advantageous for the training process using statistical methods. The underlying statistical information in the fused feature vectors is captured using GMM where several numbers of modal parameters are estimated from the distribution of fused feature vector. Maximum likelihood score is used to measure a degree of certainty to perform recognition while maximum likelihood score normalization is used for classification process. The use of likelihood score normalization is found to be able to suppress an imposter likelihood score when the background model parameters are estimated from a pool of users which include statistical information of an imposter. The present method achieved the highest recognition accuracy 97% and 99.7% when tested using FERET-PolyU dataset and ORL-PolyU dataset respectively. | en_US |
dc.description.sponsorship | Universiti Malaysia Perlis and Ministry of Higher Education Malaysia | en_US |
dc.language.iso | en | en_US |
dc.publisher | Newcastle University | en_US |
dc.title | Feature extraction and information fusion in face and palmprint multimodal biometrics | en_US |
dc.type | Thesis | en_US |
Appears in Collections: | School of Electrical and Electronic Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Ahmad, M.I. 13.pdf | Thesis | 1.72 MB | Adobe PDF | View/Open |
dspacelicence.pdf | License | 43.82 kB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.