Gaze-contingent automatic speech recognition

Neil Cooke, Martin Russell

Research output: Contribution to journalArticle

7 Citations (Scopus)

Abstract

There has been progress in improving speech recognition using a tightly-coupled modality such as lip movement; and using additional input interfaces to improve recognition of commands in multimodal human computer interfaces such as speech and pen-based systems. However, there has been little work that attempts to improve the recognition of spontaneous, conversational speech by adding information from a loosely-coupled modality. The study investigated this idea by integrating information from gaze into an automatic speech recognition (ASR) system. A probabilistic framework for multimodal recognition was formalised and applied to the specific case of integrating gaze and speech. Gaze-contingent ASR systems were developed from a baseline ASR system by redistributing language model probability mass according to the visual attention. These systems were tested on a corpus of matched eye movement and related spontaneous conversational British English speech segments (n = 1355) for a visual-based, goal-driven task. The best performing systems had similar word error rates to the baseline ASR system and showed an increase in keyword spotting accuracy. The core values of this work may be useful for developing robust speech-centric multimodal decoding system functions.
Original languageEnglish
Pages (from-to)369-380
Number of pages12
JournalIET Signal Processing
Volume2
Issue number4
DOIs
Publication statusPublished - 1 Jan 2008

Fingerprint

Dive into the research topics of 'Gaze-contingent automatic speech recognition'. Together they form a unique fingerprint.

Cite this