Logo Logo
Hilfe
Hilfe
Switch Language to English

Ramirez-Amaro, Karinne; Minhas, Humera Noor; Zehetleitner, Michael; Beetz, Michael und Cheng, Gordon (2017): Added Value of Gaze-Exploiting Semantic Representation to Allow Robots Inferring Human Behaviors. In: ACM Transactions on interactive intelligent systems, Bd. 7, Nr. 1, 5

Volltext auf 'Open Access LMU' nicht verfügbar.

Abstract

Neuroscience studies have shown that incorporating gaze view with third view perspective has a great influence to correctly infer human behaviors. Given the importance of both first and third person observations for the recognition of human behaviors, we propose a method that incorporates these observations in a technical system to enhance the recognition of human behaviors, thus improving beyond third person observations in a more robust human activity recognition system. First, we present the extension of our proposed semantic reasoning method by including gaze data and external observations as inputs to segment and infer human behaviors in complex real-world scenarios. Then, from the obtained results we demonstrate that the combination of gaze and external input sources greatly enhance the recognition of human behaviors. Our findings have been applied to a humanoid robot to online segment and recognize the observed human activities with better accuracy when using both input sources;for example, the activity recognition increases from 77% to 82% in our proposed pancake-making dataset. To provide completeness of our system, we have evaluated our approach with another dataset with a similar setup as the one proposed in this work, that is, the CMU-MMAC dataset. In this case, we improved the recognition of the activities for the egg scrambling scenario from 54% to 86% by combining the external views with the gaze information, thus showing the benefit of incorporating gaze information to infer human behaviors across different datasets.

Dokument bearbeiten Dokument bearbeiten