z-logo
open-access-imgOpen Access
Egocentric Human Activities Recognition With Multimodal Interaction Sensing
Author(s) -
Yuzhe Hao,
Asako Kanezaki,
Ikuro Sato,
Rei Kawakami,
Koichi Shinoda
Publication year - 2024
Publication title -
ieee sensors journal
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.681
H-Index - 121
eISSN - 1558-1748
pISSN - 1530-437X
DOI - 10.1109/jsen.2023.3349191
Subject(s) - signal processing and analysis , communication, networking and broadcast technologies , components, circuits, devices and systems , robotics and control systems
Egocentric human activity recognition (ego-HAR) has received attention in fields where human intentions in a video must be estimated. However, the performance of existing methods is limited due to insufficient information about the subject’s motion in egocentric videos. To overcome the problem, we proposed to use two hands’ inertial sensor data as supplements for egocentric videos to do the ego-HAR task. For this purpose, we construct a publicly available dataset, egocentric video, and inertial sensor data kitchen (EvIs-Kitchen), which contains well-synchronized egocentric videos and two-hand inertial sensor data and includes interaction-focus actions as recognition targets. We also designed the optimal choices of input combination and component variants through experiments under two-branch late-fusion architecture. The results show our multimodal setup outperforms any other single-modal methods on EvIs-Kitchen.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here