
Controlling External Devices by Using Speech and Simple Eye Movements
Author(s) -
Duy Long Dao,
Thanh An Le,
Anh Tai Tran,
Minh Huong Nguyen Thi
Publication year - 2020
Publication title -
kalpa publications in engineering
Language(s) - English
Resource type - Conference proceedings
ISSN - 2515-1770
DOI - 10.29007/jk28
Subject(s) - bluetooth , computer science , wheelchair , eye movement , movement (music) , process (computing) , arduino , computer vision , artificial intelligence , simple (philosophy) , control (management) , control system , image processing , human–computer interaction , simulation , image (mathematics) , embedded system , engineering , telecommunications , wireless , operating system , philosophy , epistemology , world wide web , electrical engineering , aesthetics
In Vietnam, the number of people with limb defects or diseases that affect the ability of limbs is increasing [1]. They were reduced or incapable of living, unable to walk, to run, and utilities around them became useless. Therefore, it is imperative to build a patient support system in controlling external devices. Recognizing the problem, the author has studied, researched, and built a wheelchair control system based on speech recognition and eye movement. In this study, the eye movement system is at the simple level: glance left and right. The system uses an image processing method to locate the pupil. The external device for testing is a car model that control by Arduino connected via Bluetooth. Experimental evaluation results show that the eye movement system achieves high accuracy of over 70%. However, the eye movement system by image processing takes a lot of time to process for a command. All of the above shows the feasibility of developing an external control device for people with disabilities.