z-logo
open-access-imgOpen Access
Controlling External Devices by Using Speech and Simple Eye Movements
Author(s) -
Duy Long Dao,
Thanh An Le,
Anh Tai Tran,
Minh Huong Nguyen Thi
Publication year - 2020
Publication title -
kalpa publications in engineering
Language(s) - English
Resource type - Conference proceedings
ISSN - 2515-1770
DOI - 10.29007/jk28
Subject(s) - bluetooth , computer science , wheelchair , eye movement , movement (music) , process (computing) , arduino , computer vision , artificial intelligence , simple (philosophy) , control (management) , control system , image processing , human–computer interaction , simulation , image (mathematics) , embedded system , engineering , telecommunications , wireless , operating system , philosophy , epistemology , world wide web , electrical engineering , aesthetics
In Vietnam, the number of people with limb defects or diseases that affect the ability of limbs is increasing [1]. They were reduced or incapable of living, unable to walk, to run, and utilities around them became useless. Therefore, it is imperative to build a patient support system in controlling external devices. Recognizing the problem, the author has studied, researched, and built a wheelchair control system based on speech recognition and eye movement. In this study, the eye movement system is at the simple level: glance left and right. The system uses an image processing method to locate the pupil. The external device for testing is a car model that control by Arduino connected via Bluetooth. Experimental evaluation results show that the eye movement system achieves high accuracy of over 70%. However, the eye movement system by image processing takes a lot of time to process for a command. All of the above shows the feasibility of developing an external control device for people with disabilities.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here