Human-computer interfaces and multimodal interaction are increasingly used in everyday life. Environments equipped with sensors are able to acquire and interpret a wide range of information, thus assisting humans in several application areas, such as behaviour understanding, event detection, action recognition, and many others. In these areas, the suitable processing of this information is a key factor to properly structure multimodal data. In particular, heterogeneous devices and different acquisition times can be exploited to improve recognition results. On the basis of these assumptions, in this paper, a multimodal system based on Allen’s temporal logic combined with a prevision method is proposed. The main target of the system is to correlate user’s events with system’s reactions. After the post-processing data coming from different acquisition devices (e.g., RGB images, depth maps, sounds, proximity sensors), the system manages the correlations between recognition/detection results and events, in real-time, thus creating an interactive environment for users. To increase the recognition reliability, a predictive model is also associated with the method. Modularity of the system grants a full dynamic development and upgrade with customized modules. Finally, comparisons with other similar systems are shown, thus underlining the high flexibility and robustness of the proposed event management method.

MIFTel: a multimodal interactive framework based on temporal logic rules / Avola, D.; Cinque, L.; Del Bimbo, A.; Marini, M. R.. - In: MULTIMEDIA TOOLS AND APPLICATIONS. - ISSN 1380-7501. - 79:19-20(2020), pp. 13533-13558. [10.1007/s11042-019-08590-1]

MIFTel: a multimodal interactive framework based on temporal logic rules

Avola D.
Primo
;
Cinque L.;Del Bimbo A.;Marini M. R.
2020

Abstract

Human-computer interfaces and multimodal interaction are increasingly used in everyday life. Environments equipped with sensors are able to acquire and interpret a wide range of information, thus assisting humans in several application areas, such as behaviour understanding, event detection, action recognition, and many others. In these areas, the suitable processing of this information is a key factor to properly structure multimodal data. In particular, heterogeneous devices and different acquisition times can be exploited to improve recognition results. On the basis of these assumptions, in this paper, a multimodal system based on Allen’s temporal logic combined with a prevision method is proposed. The main target of the system is to correlate user’s events with system’s reactions. After the post-processing data coming from different acquisition devices (e.g., RGB images, depth maps, sounds, proximity sensors), the system manages the correlations between recognition/detection results and events, in real-time, thus creating an interactive environment for users. To increase the recognition reliability, a predictive model is also associated with the method. Modularity of the system grants a full dynamic development and upgrade with customized modules. Finally, comparisons with other similar systems are shown, thus underlining the high flexibility and robustness of the proposed event management method.
2020
Event management; Human-computer interaction; Multimodal interaction; Natural interaction; Temporal logic
01 Pubblicazione su rivista::01a Articolo in rivista
MIFTel: a multimodal interactive framework based on temporal logic rules / Avola, D.; Cinque, L.; Del Bimbo, A.; Marini, M. R.. - In: MULTIMEDIA TOOLS AND APPLICATIONS. - ISSN 1380-7501. - 79:19-20(2020), pp. 13533-13558. [10.1007/s11042-019-08590-1]
File allegati a questo prodotto
File Dimensione Formato  
Avola_MIFTel_2020.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 3.11 MB
Formato Adobe PDF
3.11 MB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1411211
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 5
  • ???jsp.display-item.citation.isi??? 5
social impact