2006 IEEE International Conference on Acoustics Speed and Signal Processing Proceedings
DOI: 10.1109/icassp.2006.1661465
|View full text |Cite
|
Sign up to set email alerts
|

Fusion of Talking Face Biometric Modalities for Personal Identity Verification

Abstract: We describe a personal identity verification system based on lip dynamics biometric. The lip shape is represented in terms of a B-spline model, tracked over time. The coordinates of the 11 control points of the B-spline model are used as features for each frame. An utterance consisting of N frames produces a sequence of 22 dimensional feature vectors that is matched to the template using dynamic time warping. The verification error rate achived by the systems on the XM2VTS database is about 14%. By fusing the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 12 publications
0
6
0
Order By: Relevance
“…As shown in Table I, the most commonly used database and protocol are XM2VTS [25] (used by 3 authors) and Lausanne Protocols [26] respectively. The best performance obtained using lip features only on this database is by [14](HTER of 13.35%). Multi-modal fusion with two face detectors and two audio systems [23] yields HTER of 0.15%.…”
Section: A Summary Of Relevant Workmentioning
confidence: 92%
See 2 more Smart Citations
“…As shown in Table I, the most commonly used database and protocol are XM2VTS [25] (used by 3 authors) and Lausanne Protocols [26] respectively. The best performance obtained using lip features only on this database is by [14](HTER of 13.35%). Multi-modal fusion with two face detectors and two audio systems [23] yields HTER of 0.15%.…”
Section: A Summary Of Relevant Workmentioning
confidence: 92%
“…A final point to note is that this experiment investigates how our system compares with the state-of-the-art benchmarks. The best baseline performance obtained using lip features only on this database was by [14](HTER of 13.35%) as shown in Table I. Multi-modal fusion with two face detectors and two audio systems [23] yielded HTER of 0.15% as shown in Table II.…”
Section: Evaluation Of the Locp-top Descriptor For Speaker Authentmentioning
confidence: 94%
See 1 more Smart Citation
“…The best performance obtained using lip features only on this database is by [14](HTER of 13.35%). Multi-modal fusion with two face detectors and two audio systems [23] yields HTER of 0.15%.…”
Section: A Summary Of Relevant Workmentioning
confidence: 92%
“…The first case study, which illustrates the merit of both multimodal and intramodal fusion, detailed in [43], involves the fusion of face, voice and lip dynamics biometric modalities. The system which used off-the-shelf conventional technologies was evaluated on the XM2VTS data base [44] producing the results obtained according to the Lausanne Experimental Protocol in Configuration I [44], as shown in Table I.…”
Section: Benefits Of Multiple Biometric Expert Fusionmentioning
confidence: 99%