Publications:Audio--visual person authentication using lip-motion from orientation maps

From ISLAB/CAISR

Do not edit this section

Keep all hand-made modifications below

Title Audio–visual person authentication using lip-motion from orientation maps
Author Maycel-Isaac Faraj and Josef Bigun
Year 2007
PublicationType Journal Paper
Journal Pattern Recognition Letters
HostPublication
Conference
DOI http://dx.doi.org/10.1016/j.patrec.2007.02.017
Diva url http://hh.diva-portal.org/smash/record.jsf?searchId=1&pid=diva2:238553
Abstract This paper describes a new identity authentication technique by a synergetic use of lip-motion and speech. The lip-motion is defined as the distribution of apparent velocities in the movement of brightness patterns in an image and is estimated by computing the velocity components of the structure tensor by 1D processing, in 2D manifolds. Since the velocities are computed without extracting the speaker’s lip-contours, more robust visual features can be obtained in comparison to motion features extracted from lip-contours. The motion estimations are performed in a rectangular lip-region, which affords increased computational efficiency. A person authentication implementation based on lip-movements and speech is presented along with experiments exhibiting a recognition rate of 98%. Besides its value in authentication, the technique can be used naturally to evaluate the “liveness” of someone speaking as it can be used in text-prompted dialogue. The XM2VTS database was used for performance quantification as it is currently the largest publicly available database (≈300 persons) containing both lip-motion and speech. Comparisons with other techniques are presented.