首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
In this paper, we demonstrate that use of a recently proposed feature set, termed Maximum Auto-Correlation Values, which utilizes information from the source part of the speech signal, significantly improves the robustness of a text independent identity verification system. We also propose an adaptive fusion technique for integration of audio and visual information in a multi-modal verification system. The proposed technique explicitly measures the quality of the speech signal, adjusting the amount of contribution of the speech modality to the final verification decision. Results on the VidTIMIT database indicate that the proposed approach outperforms existing adaptive and non-adaptive fusion techniques. For a wide range of audio SNRs, the performance of the multi-modal system utilizing the proposed technique is always found to be better than the performance of the face modality.  相似文献   

2.
With an aim of extracting robust facial features under pose variations, this paper presents two directional projections corresponding to extraction of vertical and horizontal local face image features. The matching scores computed from both horizontal and vertical features are subsequently fused at score level via an extreme learning machine that optimizes the total error rate for performance enhancement. In order to benchmark the performance, both the feature extraction and fusion results are compared with that of popular face recognition methods such as principal components analysis and linear discriminant analysis in terms of equal error rate and CPU time. Our empirical experiments using four data sets show encouraging results under considerable horizontal pose variations.  相似文献   

3.
In this paper, we propose to extract localized random features directly from partial face image matrix for cancelable identity verification. Essentially, the extracted random features consist of compressed horizontal and vertical facial information obtained from a structured projection of the raw face images. For template security reason, the face appearance information is concealed via averaging several templates over different transformations. The match score outputs of these cancelable templates are then fused through a total error rate minimization. Extensive experiments were carried out to evaluate and benchmark the performance of the proposed method based on the AR, FERET, ORL, Sheffield and BERC databases. Our empirical results show encouraging performances in terms of verification accuracy as well as satisfying four cancelable biometric properties.  相似文献   

4.
5.
The use of personal identity verification systems with multi-modal biometrics has been proposed in order to increase the performance and robustness against environmental variations and fraudulent attacks. Usually multi-modal fusion of biometrics is performed in parallel at the score-level by combining the individual matching scores. This parallel strategy exhibits some drawbacks: (i) all available biometrics are necessary to perform fusion, thus the verification time depends on the slowest system; (ii) some users could be easily recognizable using a certain biometric instead of another one and (iii) the system invasiveness increases. A system characterized by the serial combination of multiple biometrics can be a good trade-off between verification time, performance and acceptability. However, these systems have been poorly investigated, and no support for designing the processing chain has been given so far. In this paper, we propose a novel serial scheme and a simple mathematical model able to predict the performance of two serially combined matchers as function of the selected processing chain. Our model helps the designer in finding the processing chain allowing a trade-off, in particular, between performance and matching time. Experiments carried out on well-known benchmark data sets made up of face and fingerprint images support the usefulness of the proposed methodology and compare it with standard parallel fusion.  相似文献   

6.
介绍了一种二代身份证识别验证系统,该系统针对身份证照片样本单一的问题,提出一种将二代身份证照片从单一样本虚拟为多样本的方法。该系统在在一定程度上减弱了人脸姿态的变化对识别率的影响,并在实际采集的数据库中验证了该方法的有效性。  相似文献   

7.
随着RFID识别技术的广泛应用以及人脸识别技术的发展,结合两种技术将会提高身份识别的安全性和有效性。基于此,整合了RFID射频卡信息处理、Adaboost人脸检测、LDA人脸识别等模式识别处理技术,实现了一个身份智能验证系统,可应用于大型活动、会议和体育赛事的安检以及公司的考勤中,具有重要的实际应用价值。  相似文献   

8.
针对人脸认证中的小样本问题和Gabor小波特征提取的不足,提出一种有效的人脸认证算法。对预处理后的图像进行2D双树复小波变换,将每幅图像不同尺度下多个方向的小波系数幅值作为特征矢量,表征重要的局部信息;将提取的特征矢量向判别共同矢量空间投影,进一步提取具有判别能力的特征,同时进行降维;根据用户特定阈值进行认证。ORL人脸库和FERET子库上的实验结果验证了算法的有效性。  相似文献   

9.
《Information Fusion》2002,3(4):267-276
Multimodal fusion for identity verification has already shown great improvement compared to unimodal algorithms. In this paper, we propose to integrate confidence measures during the fusion process. We present a comparison of three different methods to generate such confidence information from unimodal identity verification systems. These methods can be used either to enhance the performance of a multimodal fusion algorithm or to obtain a confidence level on the decisions taken by the system. All the algorithms are compared on the same benchmark database, namely XM2VTS, containing both speech and face information. Results show that some confidence measures did improve statistically significantly the performance, while other measures produced reliable confidence levels over the fusion decisions.  相似文献   

10.
In this paper a method is proposed for person identification and verification. The method proposed in Viola and Jones (2001) is used to detect the face region in the image. The detected face region is processed to determine the locations of the eyes and mouth. The facial and mouth features are extracted relative to the locations of the eyes and mouth. A new feature called fovea intensity comparison code (FICC) is obtained from intensity values of the face/mouth region. The dimension of the FICC is reduced using principal component analysis (PCA). Euclidean distance matching is used for identification and verification. The performance of the system is evaluated in real time in the laboratory environment, and the system achieves a recognition rate (RR) of 99.0% and an equal error rate (EER) of about 0.84% for 50 subjects. The performance of the system is also evaluated for the eXtended Multi Modal Verification for Teleservices and Security (XM2VTS) database, and the system achieves a recognition rate of 100% an equal error rate (EER) of about 0.23%.  相似文献   

11.
In this letter we propose a piece-wise linear (PL) classifier for use as the decision stage in a two-modal verification system, comprised of a face and a speech expert. The classifier utilizes a fixed decision boundary that has been specifically designed to account for the effects of noisy audio conditions. Experimental results on the VidTIMIT database show that in clean conditions, the proposed classifier is outperformed by a traditional weighted summation decision stage (using both fixed and adaptive weights). Using white Gaussian noise to corrupt the audio data resulted in the PL classifier obtaining better performance than the fixed approach and similar performance to the adaptive approach. Using a more realistic noise type, namely “operations room” noise from the NOISEX-92 corpus, resulted in the PL classifier obtaining better performance than both the fixed and adaptive approaches. The better results in this case stem from the PL classifier not making a direct assumption about the type of noise that causes the mismatch between training and testing conditions (unlike the adaptive approach). Moreover, the PL classifier has the advantage of having a fixed (non-adaptive, thus simpler) structure.  相似文献   

12.
This paper describes a real time vision system that allows us to localize faces in video sequences and verify their identity. These processes are image processing techniques based on the radial basis function (RBF) neural network approach. The robustness of this system has been evaluated quantitatively on eight video sequences. We have adapted our model for an application of face recognition using the Olivetti Research Laboratory (ORL), Cambridge, UK, database so as to compare the performance against other systems. We also describe three hardware implementations of our model on embedded systems based on the field programmable gate array (FPGA), zero instruction set computer (ZISC) chips, and digital signal processor (DSP) TMS320C62, respectively. We analyze the algorithm complexity and present results of hardware implementations in terms of the resources used and processing speed. The success rates of face tracking and identity verification are 92% (FPGA), 85% (ZISC), and 98.2% (DSP), respectively. For the three embedded systems, the processing speeds for images size of 288 /spl times/ 352 are 14 images/s, 25 images/s, and 4.8 images/s, respectively.  相似文献   

13.
Graph embedding (GE) is a unified framework for dimensionality reduction techniques. GE attempts to maximally preserve data locality after embedding for face representation and classification. However, estimation of true data locality could be severely biased due to limited number of training samples, which trigger overfitting problem. In this paper, a graph embedding regularization technique is proposed to remedy this problem. The regularization model, dubbed as Locality Regularization Embedding (LRE), adopts local Laplacian matrix to restore true data locality. Based on LRE model, three dimensionality reduction techniques are proposed. Experimental results on five public benchmark face datasets such as CMU PIE, FERET, ORL, Yale and FRGC, along with Nemenyi Post-hoc statistical of significant test attest the promising performance of the proposed techniques.  相似文献   

14.
When combining outputs from multiple classifiers, many combination rules are available. Although easy to implement, fixed combination rules are optimal only in restrictive conditions. We discuss and evaluate their performance when the optimality conditions are not fulfilled. Fixed combination rules are then compared with trainable combination rules on real data in the context of face-based identity verification. The face images are classified by combining the outputs of five different face verification experts. It is demonstrated that a reduction in the error rates of up to 50% over the best single expert is achieved on the XM2VTS database, using either fixed or trainable combination rules.  相似文献   

15.
In this paper, we propose a novel image-based identity verification method. This method first uses the training images of the claimed identity to represent the testing sample and then exploits the representation result to determine the verification result, that is, accept or reject. The proposed method not only has sound theoretical foundation but also is simple and easy to implement. Moreover, our method greatly outperforms previously image-based identity verification methods.  相似文献   

16.
一种融合PCA 和KFDA 的人脸识别方法   总被引:2,自引:0,他引:2       下载免费PDF全文
陈才扣  杨静宇  杨健 《控制与决策》2004,19(10):1147-1150
提出一种融合PCA和KFDA的人脸识别方法,即在进行非线性映射之前,首先利用经典的主分量分析(C—PCA)进行降维,然后执行KFDA.为进一步降低整个算法的计算时问,又提出一种I—PCA KFDA方法,它直接基于图像矩阵的主分量分析(I—PCA).ORL标准人脸库的试验结果表明,与现有的核Fisher鉴别分析方法相比,两种方法可将特征抽取的速度分别提高3倍和7倍,其识别精度没有丝毫的降低.  相似文献   

17.
18.
This work explores the use of speech enhancement for enhancing degraded speech which may be useful for text dependent speaker verification system. The degradation may be due to noise or background speech. The text dependent speaker verification is based on the dynamic time warping (DTW) method. Hence there is a necessity of the end point detection. The end point detection can be performed easily if the speech is clean. However the presence of degradation tends to give errors in the estimation of the end points and this error propagates into the overall accuracy of the speaker verification system. Temporal and spectral enhancement is performed on the degraded speech so that ideally the nature of the enhanced speech will be similar to the clean speech. Results show that the temporal and spectral processing methods do contribute to the task by eliminating the degradation and improved accuracy is obtained for the text dependent speaker verification system using DTW.  相似文献   

19.
We address the pose mismatch problem which can occur in face verification systems that have only a single (frontal) face image available for training. In the framework of a Bayesian classifier based on mixtures of gaussians, the problem is tackled through extending each frontal face model with artificially synthesized models for non-frontal views. The synthesis methods are based on several implementations of maximum likelihood linear regression (MLLR), as well as standard multi-variate linear regression (LinReg). All synthesis techniques rely on prior information and learn how face models for the frontal view are related to face models for non-frontal views. The synthesis and extension approach is evaluated by applying it to two face verification systems: a holistic system (based on PCA-derived features) and a local feature system (based on DCT-derived features). Experiments on the FERET database suggest that for the holistic system, the LinReg-based technique is more suited than the MLLR-based techniques; for the local feature system, the results show that synthesis via a new MLLR implementation obtains better performance than synthesis based on traditional MLLR. The results further suggest that extending frontal models considerably reduces errors. It is also shown that the local feature system is less affected by view changes than the holistic system; this can be attributed to the parts based representation of the face, and, due to the classifier based on mixtures of gaussians, the lack of constraints on spatial relations between the face parts, allowing for deformations and movements of face areas.  相似文献   

20.
Unified model in identity subspace for face recognition   总被引:1,自引:0,他引:1       下载免费PDF全文
Human faces have two important characteristics: (1) They are similar objects and the specific variations of each face are similar to each other; (2) They are nearly bilateral symmetric. Exploiting the two important properties, we build a unified model in identity subspace (UMIS) as a novel technique for face recognition from only one example image per person. An identity subspace spanned by bilateral symmetric bases, which compactly encodes identity information, is presented. The unified model, trained on an obtained training set with multiple samples per class from a known people group A, can be generalized well to facial images of unknown individuals, and can be used to recognize facial images from an unknown people group B with only one sample per subject. Extensive experimental results on two public databases (the Yale database and the Bern database) and our own database (the ICT-JDL database) demonstrate that the UMIS approach is significantly effective and robust for face recognition.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号