Speaker Detection and Applications to Cross-Modal Analysis of Planning Meetings Conference Paper uri icon

abstract

  • Detection of meeting events is one of the most important tasks in multimodal analysis of planning meetings. Speaker detection is a key step for extraction of most meaningful meeting events. In this paper, we present an approach of speaker localization using combination of visual and audio information in multimodal meeting analysis. When talking, people make a speech accompanying mouth movements and hand gestures. By computing correlation of audio signals, mouth movements, and hand motion, we detect a talking person both spatially and temporally. Three kinds of features are extracted for speaker localization. Hand movements are expressed by hand motion efforts; audio features are expressed by computing 12 mel-frequency cepstral coefficients from audio signals, and mouth movements are expressed by normalized cross-correlation coefficients of mouth area between two successive frames. A time delay neural network is trained to learn the correlation relationships, which is then applied to perform speaker localization. Experiments and applications in planning meeting environments are provided. 2009 IEEE.

name of conference

  • 2009 11th IEEE International Symposium on Multimedia

published proceedings

  • 2009 11th IEEE International Symposium on Multimedia

author list (cited authors)

  • Fang, B., Xiong, Y., & Quek, F.

citation count

  • 0

complete list of authors

  • Fang, Bing||Xiong, Yingen||Quek, Francis

publication date

  • January 2009