It is of prime importance in everyday human life to cope with and respond appropriately to events that are not foreseen by prior experience. Machines to a large extent lack the ability to respond appropriately to such inputs. An important class of unexpected events is defined by incongruent combinations of inputs from different modalities and therefore multimodal information provides a crucial cue for the identification of such events, e.g., the sound of a voice is being heard while the person in the field-of-view does not move her lips. In the project DIRAC ("Detection and Identification of Rare Audio-visual Cues") we have been developing algorithmic approaches to the detection of such events, as well as an experimental hardware platform to test it. An audio-visual platform ("AWEAR" - audio-visual wearable device) has been constructed with the goal to help users with disabilities or a high cognitive load to deal with unexpected events. Key hardware components include stereo panoramic vision sensors and 6-channel worn-behind-the-ear (hearing aid) microphone arrays. Data have been recorded to study audio-visual tracking, a/v scene/object classification and a/v detection of incongruencies. © 2008 ACM.

The DIRAC AWEAR audio-visual platform for detection of unexpected and incongruent events / Anemüller, Jörn; Bach, Jörg Hendrik; Caputo, Barbara; Havlena, Michal; Jie, Luo; Kayser, Hendrik; Leibe, Bastian; Motlicek, Petr; Pajdla, Tomas; Pavel, Misha; Torii, Akihiko; Gool, Luc Van; Zweig, Alon; Hermansky, Hynek. - STAMPA. - (2008), pp. 289-292. (Intervento presentato al convegno 10th International Conference on Multimodal Interfaces, ICMI'08 tenutosi a Chania, Crete; Greece nel 20-22 October 2008) [10.1145/1452392.1452451].

The DIRAC AWEAR audio-visual platform for detection of unexpected and incongruent events

CAPUTO, BARBARA;
2008

Abstract

It is of prime importance in everyday human life to cope with and respond appropriately to events that are not foreseen by prior experience. Machines to a large extent lack the ability to respond appropriately to such inputs. An important class of unexpected events is defined by incongruent combinations of inputs from different modalities and therefore multimodal information provides a crucial cue for the identification of such events, e.g., the sound of a voice is being heard while the person in the field-of-view does not move her lips. In the project DIRAC ("Detection and Identification of Rare Audio-visual Cues") we have been developing algorithmic approaches to the detection of such events, as well as an experimental hardware platform to test it. An audio-visual platform ("AWEAR" - audio-visual wearable device) has been constructed with the goal to help users with disabilities or a high cognitive load to deal with unexpected events. Key hardware components include stereo panoramic vision sensors and 6-channel worn-behind-the-ear (hearing aid) microphone arrays. Data have been recorded to study audio-visual tracking, a/v scene/object classification and a/v detection of incongruencies. © 2008 ACM.
2008
10th International Conference on Multimodal Interfaces, ICMI'08
Audio-visual; Augmented cognition; Event detection; Multimodal interaction; Sensor platform; Computer Science Applications1707 Computer Vision and Pattern Recognition; Human-Computer Interaction
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
The DIRAC AWEAR audio-visual platform for detection of unexpected and incongruent events / Anemüller, Jörn; Bach, Jörg Hendrik; Caputo, Barbara; Havlena, Michal; Jie, Luo; Kayser, Hendrik; Leibe, Bastian; Motlicek, Petr; Pajdla, Tomas; Pavel, Misha; Torii, Akihiko; Gool, Luc Van; Zweig, Alon; Hermansky, Hynek. - STAMPA. - (2008), pp. 289-292. (Intervento presentato al convegno 10th International Conference on Multimodal Interfaces, ICMI'08 tenutosi a Chania, Crete; Greece nel 20-22 October 2008) [10.1145/1452392.1452451].
File allegati a questo prodotto
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/951736
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? ND
social impact