Abstract
A three-electrode human-computer interaction system, based on digital processing of the Electromyogram (EMG) signal, is presented. This system can effectively help disabled individuals paralyzed from the neck down to interact with computers or communicate with people through computers using point-and-click graphic interfaces. The three electrodes are placed on the right frontalis, the left temporalis and the right temporalis muscles in the head, respectively. The signal processing algorithm used translates the EMG signals during five kinds of facial movements (left jaw clenching, right jaw clenching, eyebrows up, eyebrows down, simultaneous left & right jaw clenching) into five corresponding types of cursor movements (left, right, up, down and left-click), to provide basic mouse control. The classification strategy is based on three principles: the EMG energy of one channel is typically larger than the others during one specific muscle contraction; the spectral characteristics of the EMG signals produced by the frontalis and temporalis muscles during different movements are different; the EMG signals from adjacent channels typically have correlated energy profiles. The algorithm is evaluated on 20 pre-recorded EMG signal sets, using Matlab simulations. The results show that this method provides improvements and is more robust than other previous approaches.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Jonathan R. Wolpaw, Niels Birbaumer, Dennis J. McFarland, Gert Pfurtscheller and Theresa M. Vaughan, “Brain-computer interfaces for communication and control,”Clinical Neurophysiology, vol. 113, Issue 6, pp. 767-791, June 2002.
Eric C. Leuthardt, Gerwin Schalk, Jonathan R. Wolpaw, Jeffrey G. Ojemann and Daniel W. Moran, “A brain–computer interface using electrocorticographic signals in humans,” J. Neural Eng., vol.1, issue 2,June 2004,pp. 63-71 .
A. B Barreto, S. D Scargle, and M. Adjouadi, “A Real-Time Assistive Computer Interface for Users with Motor Disabilities,” SIGCAPH Newsletter, Issue 64, 1999, pp. 6-16.
A. B Barreto, S. D Scargle, and M. Adjouadi, “ A Practical EMG-based Human-Computer Interface for Users with Motor Disabilities,” Journal Of Rehabilitation Research And Development, vol. 37, Issue 1, January - February 2000, pp.53-63.
T. Itou, M. Terao, J. Nagata and M. Yoshida, “Mouse cursor control system using EMG,” 2001 Conference Proceedings of the 23rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, vol. 2, 2001, pp. 1368-1369.
M. Yoshida, T. Itou and J. Nagata, “Development of EMG controlled mouse cursor,” Conference Proceedings, Second Joint EMBS-BMES Conference 2002.24th Annual International Conference of the Engineering in Medicine and Biology Society, Annual Fall Meeting of the Biomedical Engineering Society, vol. 3, 2002, pp. 2436.
Craig A. Chin, Armando Barreto, Gualberto Cremades and Malek Adjouadi, “Integrated electromyogram and eye-gaze tracking cursor control system for computer users with motor disabilities,” Journal of Rehabilitation Research & Development, vol. 45, Number 1, 2008, pp.161-174.
G.C. Chang, W.J. Kang, J.J. Luh, C.K. Cheng, J.S. Lai, J.J. Chen and T.S. Kuo, “ Real-time implementation of electromyogram pattern recognition as a control command of man-machine interface,” Med Eng Phys., vol. 45,1996,18(7),PP.529–37.
B. LeVeau and G. Andersson, “Output forms: data analysis and applications,” Interpretation of the electromyographic signal, selected topics in surface electromyography for use in the occupational setting: expert perspective, U .S . Department of Health and Human Services, NIOSH Pub. No. 91-100, March 1992.
A. Sasaki, H. Hashimoto and C. Ishii, “Driving Electric Car by Using EMG Interface,” Cybernetics and Intelligent Systems, 2006 IEEE Conference, June 2006, PP.1-5.
Henry Gray, Anatomy of the Human Body. 20th edition, 1918.
S. Shahid,J. Walker, G.M Lyons, C.A. Byrne, A.V. Nene , “Application of higher order statistics techniques to EMG signals to characterize the motor unit action potential,” IEEE Transactions on Biomedical Engineering,2005Vol.52(No.7).
C. Chin,A. Barreto,Jing Zhaiand Chao Li, “ New classification algorithm for electromyography-based computer cursor control system,” SoutheastCon, 2005. Proceedings. IEEE, 8-10 April 2005, PP.428- 432.
T.E Hutchinson, K.P. White, W.N Martin, K.C. Reichert, L.A. Frey “Human-computer interaction using eye-gaze input,”IEEE Transactions on Systems, Man and Cybernetics, Volume 19, Issue 6, 1989, pp. 1527-1534.
R.J.K. Jacob, “The use of eye movements in human-computer interaction techniques: what you look at is what you get,” ACM Transactions on Information Systems, Volume 9, Issue 2, 1991, pp. 152-169.
Lankford Chris; “Effective eye-gaze input into Windows,” Proceedings of the symposium on Eye tracking research & applications, 2000, pp. 23-27.
Acknowledgements
This work was sponsored by NSF grants CNS-0520811, CNS-0426125, and HRD-0833093.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer Science+Business Media B.V.
About this paper
Cite this paper
Ren, P., Barreto, A., Adjouadi, M. (2010). Multi-step EMG Classification Algorithm for Human-Computer Interaction. In: Sobh, T., Elleithy, K. (eds) Innovations in Computing Sciences and Software Engineering. Springer, Dordrecht. https://doi.org/10.1007/978-90-481-9112-3_31
Download citation
DOI: https://doi.org/10.1007/978-90-481-9112-3_31
Published:
Publisher Name: Springer, Dordrecht
Print ISBN: 978-90-481-9111-6
Online ISBN: 978-90-481-9112-3
eBook Packages: Computer ScienceComputer Science (R0)