Paper
20 August 1993 Locating the mouth region in images of human faces
H. J. Grech-Cini, Gerard T. McKee
Author Affiliations +
Proceedings Volume 2059, Sensor Fusion VI; (1993) https://doi.org/10.1117/12.150249
Event: Optical Tools for Manufacturing and Advanced Automation, 1993, Boston, MA, United States
Abstract
Being able to see the face of a speaker can improve speech recognition performance by as much as a shift from 20% to 80% intelligibility under certain circumstances. Lip movements provide a major source of visual cues in speech recognition. In our research we are concerned with locating, tracking, characterizing, and exploiting the lip movements for this purpose. In this paper we focus on the first of these problems. Using a technique based on n-Tuples we locate the `eye-nose-region' (ENR) of the face in images and infer the location of the mouth via a `face model.' We describe this method in detail and present initial test results.
© (1993) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
H. J. Grech-Cini and Gerard T. McKee "Locating the mouth region in images of human faces", Proc. SPIE 2059, Sensor Fusion VI, (20 August 1993); https://doi.org/10.1117/12.150249
Lens.org Logo
CITATIONS
Cited by 10 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Mouth

Speech recognition

Stochastic processes

Image fusion

Visualization

Binary data

Sensor fusion

RELATED CONTENT

Visual words for lip-reading
Proceedings of SPIE (April 28 2010)
Unifying Voice And Hand Indication Of Spatial Layout
Proceedings of SPIE (March 01 1990)
Color-based lip localization method
Proceedings of SPIE (April 28 2010)

Back to Top