
RGB‐D static gesture recognition based on convolutional neural network
Author(s) -
Xie Bin,
He Xiaoyu,
Li Yi
Publication year - 2018
Publication title -
the journal of engineering
Language(s) - English
Resource type - Journals
ISSN - 2051-3305
DOI - 10.1049/joe.2018.8327
Subject(s) - computer science , gesture recognition , rgb color model , artificial intelligence , gesture , convolutional neural network , computer vision , sign language , feature extraction , segmentation , feature (linguistics) , sketch recognition , pattern recognition (psychology) , philosophy , linguistics
In the area of human–computer interaction (HCI) and computer vision, gesture recognition has always been a research hotspot. With the appearance of depth camera, gesture recognition using RGB‐D camera has gradually become mainstream in this field. However, how to effectively use depth information to construct a robust gesture recognition system is still a problem. In this paper, an RGB‐D static gesture recognition method based on fine‐tuning Inception V3 is proposed, which can eliminate the steps of gesture segmentation and feature extraction in traditional algorithms. Compared with general CNN algorithms, the authors adopt a two‐stage training strategy to fine‐tune the model. This method sets a feature concatenate layer of RGB and depth images in the CNN structure, using depth information to promote the performance of gesture recognition. Finally, on the American Sign Language (ASL) Recognition dataset, the authors compared their method with other traditional machine learning methods, CNN algorithms, and the RGB input only method. Among three groups of comparative experiments, the authors’ method reached the highest accuracy of 91.35%, reaching the state‐of‐the‐art currently on ASL dataset.