z-logo
open-access-imgOpen Access
Temporal-Sound based User Interface for Smart Home
Author(s) -
Kido Tani,
Nobuyuki Umezu
Publication year - 2021
Language(s) - English
Resource type - Conference proceedings
DOI - 10.5121/csit.2021.112107
Subject(s) - spectrogram , gesture , computer science , desk , accelerometer , interface (matter) , acceleration , speech recognition , artificial intelligence , computer vision , physics , bubble , classical mechanics , maximum bubble pressure method , parallel computing , operating system
We propose a gesture-based interface to control a smart home. Our system replaces existing physical controls with our temporal sound commands using accelerometer. In our preliminary experiments, we recorded the sounds generated by six different gestures (knocking the desk, mouse clicking, and clapping) and converted them into spectrogram images. Classification learning was performed on these images using a CNN. Due to the difference between the microphones used, the classification results are not successful for most of the data. We then recorded acceleration values, instead of sounds, using a smart watch. 5 types of motions were performed in our experiments to execute activity classification on these acceleration data using a machine learning library named Core ML provided by Apple Inc.. These results still have much room to be improved.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here