A Dynamic Bayesian Network Based Framework for Multimodal Context-Aware Interactions (IUI'25)
Автор: Interactive Structures Lab
Загружено: 2025-08-08
Просмотров: 84
Описание:
"A Dynamic Bayesian Network Based Framework for Multimodal Context-Aware Interactions" is a research project by Meta Reality Labs and the Human-Computer Interaction Institute at Carnegie Mellon University.
Authors: Violet Yinuo Han, Tianyi Wang, Hyunsung Cho, Kashyap Todi, Ajoy Savio Fernandes, Andre Levi, Zheng Zhang, Tovi Grossman, Alexandra Ion, Tanya R Jonker
Published in IUI'25
DOI: https://interactive-structures.org/pu...
//Abstract
Multimodal context-aware interactions integrate multiple sensory inputs, such as gaze, gestures, speech, and environmental signals, to provide adaptive support across diverse user contexts. Building such systems is challenging due to the complexity of sensor fusion, real-time decision-making, and managing uncertainties from noisy inputs. To address these challenges, we propose a hybrid approach combining a dynamic Bayesian network (DBN) with a large language model (LLM). The DBN offers a probabilistic framework for modeling variables, relationships, and temporal dependencies, enabling robust, real-time inference of user intent, while the LLM incorporates world knowledge for contextual reasoning. We demonstrate our approach with a tri-level DBN implementation for tangible interactions, integrating gaze and hand actions to infer user intent in real time. A user evaluation with 10 participants in an everyday office scenario showed that our system can accurately and efficiently infer user intentions, achieving 0.83 per frame accuracy, even in complex environments. These results validate the effectiveness of the DBN+LLM framework for multimodal context-aware interactions.
Повторяем попытку...
Доступные форматы для скачивания:
Скачать видео
-
Информация по загрузке: