z-logo
open-access-imgOpen Access
Cost-Efficient Fall Risk Assessment with Attention Augmented Vision Machine Learning on Sit-To-Stand Test Videos
Author(s) -
Chunhua Pan,
Boting Qu,
Rui Miao,
Xin Wang
Publication year - 2025
Publication title -
ieee access
Language(s) - English
Resource type - Magazines
SCImago Journal Rank - 0.587
H-Index - 127
eISSN - 2169-3536
DOI - 10.1109/access.2025.3598002
Subject(s) - aerospace , bioengineering , communication, networking and broadcast technologies , components, circuits, devices and systems , computing and processing , engineered materials, dielectrics and plasmas , engineering profession , fields, waves and electromagnetics , general topics for engineers , geoscience , nuclear engineering , photonics and electrooptics , power, energy and industry applications , robotics and control systems , signal processing and analysis , transportation
Falls among the elderly and other at-risk populations pose a significant public health challenge, necessitating innovative methods to detect and intervene early. While automated fall risk assessment methods using wearable sensors or depth cameras have been proposed, the physical and psychological burden of wearing extra sensors, and the high cost of specialized equipment like Kinect, limits their practical adoption. To tackle this challenge, this paper presents a novel machine learning-based fall risk assessment approach called FRAVM, which operates on Five times Sit-To-Stand (FSTS) test videos captured with standard, widely available cameras to identify individuals requiring fall prevention interventions. To enhance the practicality of FRAVM, 3D pose estimation is applied to generate vision-rich 3D body keypoints, mitigating the challenges posed by restricted camera angles. Median-average filtering is used to reduce noise caused by video shaking and pose estimation inaccuracies, while a new Dynamic Time Warping (DTW)-based matching algorithm is designed to handle interference from irrelevant individuals appearing in the video. Furthermore, a novel Attention-augmented Spatial-Temporal Graph Convolutional Network (AST-GCN) is developed for reliably identifying the action in each frame, enabling accurate computation of key kinematic features for fall risk prediction. Experimental evaluations on a dataset of 450 FSTS test videos demonstrate the high performance of FRAVM, with detection accuracy, F1 score, and ROC-AUC achieving 88.64%, 88.58%, and 98.86%, respectively. The ablation analysis confirms that the components in FRAVM are essential for achieving optimal results.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom