1
|
Hu F, Qian M, He K, Zhang WA, Yang X. A Novel Multi-Feature Fusion Network With Spatial Partitioning Strategy and Cross-Attention for Armband-Based Gesture Recognition. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3878-3890. [PMID: 39466868 DOI: 10.1109/tnsre.2024.3487216] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/30/2024]
Abstract
Effectively integrating the time-space-frequency information of multi-modal signals from armband sensor, including surface electromyogram (sEMG) and accelerometer data, is critical for accurate gesture recognition. Existing approaches often neglect the abundant spatial relationships inherent in multi-channel sEMG signals obtained via armband sensors and face challenges in harnessing the correlations across multiple feature domains. To address this issue, we propose a novel multi-feature fusion network with spatial partitioning strategy and cross-attention (MFN-SPSCA) to improve the accuracy and robustness of gesture recognition. Specifically, a spatiotemporal graph convolution module with a spatial partitioning strategy is designed to capture potential spatial feature of multi-channel sEMG signals. Additionally, we design a cross-attention fusion module to learn and prioritize the importance and correlation of multi-feature domain. Extensive experiment demonstrate that the MFN-SPSCA method outperforms other state-of-the-art methods on self-collected dataset and the Ninapro DB5 dataset. Our work addresses the challenge of recognizing gestures from the multi-modal data collected by armband sensor, emphasizing the importance of integrating time-space-frequency information. Codes are available at https://github.com/ZJUTofBrainIntelligence/MFN-SPSCA.
Collapse
|
2
|
Ben Haj Amor A, El Ghoul O, Jemni M. Sign Language Recognition Using the Electromyographic Signal: A Systematic Literature Review. SENSORS (BASEL, SWITZERLAND) 2023; 23:8343. [PMID: 37837173 PMCID: PMC10574929 DOI: 10.3390/s23198343] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/30/2023] [Revised: 09/23/2023] [Accepted: 10/03/2023] [Indexed: 10/15/2023]
Abstract
The analysis and recognition of sign languages are currently active fields of research focused on sign recognition. Various approaches differ in terms of analysis methods and the devices used for sign acquisition. Traditional methods rely on video analysis or spatial positioning data calculated using motion capture tools. In contrast to these conventional recognition and classification approaches, electromyogram (EMG) signals, which measure muscle electrical activity, offer potential technology for detecting gestures. These EMG-based approaches have recently gained attention due to their advantages. This prompted us to conduct a comprehensive study on the methods, approaches, and projects utilizing EMG sensors for sign language handshape recognition. In this paper, we provided an overview of the sign language recognition field through a literature review, with the objective of offering an in-depth review of the most significant techniques. These techniques were categorized in this article based on their respective methodologies. The survey discussed the progress and challenges in sign language recognition systems based on surface electromyography (sEMG) signals. These systems have shown promise but face issues like sEMG data variability and sensor placement. Multiple sensors enhance reliability and accuracy. Machine learning, including deep learning, is used to address these challenges. Common classifiers in sEMG-based sign language recognition include SVM, ANN, CNN, KNN, HMM, and LSTM. While SVM and ANN are widely used, random forest and KNN have shown better performance in some cases. A multilayer perceptron neural network achieved perfect accuracy in one study. CNN, often paired with LSTM, ranks as the third most popular classifier and can achieve exceptional accuracy, reaching up to 99.6% when utilizing both EMG and IMU data. LSTM is highly regarded for handling sequential dependencies in EMG signals, making it a critical component of sign language recognition systems. In summary, the survey highlights the prevalence of SVM and ANN classifiers but also suggests the effectiveness of alternative classifiers like random forests and KNNs. LSTM emerges as the most suitable algorithm for capturing sequential dependencies and improving gesture recognition in EMG-based sign language recognition systems.
Collapse
Affiliation(s)
| | - Oussama El Ghoul
- Mada—Assistive Technology Center Qatar, Doha P.O. Box 24230, Qatar;
| | - Mohamed Jemni
- Arab League Educational, Cultural, and Scientific Organization, Tunis 1003, Tunisia
| |
Collapse
|
3
|
Wang X, Tang L, Zheng Q, Yang X, Lu Z. IRDC-Net: An Inception Network with a Residual Module and Dilated Convolution for Sign Language Recognition Based on Surface Electromyography. SENSORS (BASEL, SWITZERLAND) 2023; 23:5775. [PMID: 37447625 DOI: 10.3390/s23135775] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Revised: 06/06/2023] [Accepted: 06/15/2023] [Indexed: 07/15/2023]
Abstract
Deaf and hearing-impaired people always face communication barriers. Non-invasive surface electromyography (sEMG) sensor-based sign language recognition (SLR) technology can help them to better integrate into social life. Since the traditional tandem convolutional neural network (CNN) structure used in most CNN-based studies inadequately captures the features of the input data, we propose a novel inception architecture with a residual module and dilated convolution (IRDC-net) to enlarge the receptive fields and enrich the feature maps, applying it to SLR tasks for the first time. This work first transformed the time domain signal into a time-frequency domain using discrete Fourier transformation. Second, an IRDC-net was constructed to recognize ten Chinese sign language signs. Third, the tandem CNN networks VGG-net and ResNet-18 were compared with our proposed parallel structure network, IRDC-net. Finally, the public dataset Ninapro DB1 was utilized to verify the generalization performance of the IRDC-net. The results showed that after transforming the time domain sEMG signal into the time-frequency domain, the classification accuracy (acc) increased from 84.29% to 91.70% when using the IRDC-net on our sign language dataset. Furthermore, for the time-frequency information of the public dataset Ninapro DB1, the classification accuracy reached 89.82%; this value is higher than that achieved in other recent studies. As such, our findings contribute to research into SLR tasks and to improving deaf and hearing-impaired people's daily lives.
Collapse
Affiliation(s)
- Xiangrui Wang
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| | - Lu Tang
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| | - Qibin Zheng
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| | - Xilin Yang
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| | - Zhiyuan Lu
- School of Rehabilitation Science and Engineering, University of Health and Rehabilitation Sciences, Qingdao 266072, China
| |
Collapse
|
4
|
David JP, Helbig T, Witte H. SenGlove—A Modular Wearable Device to Measure Kinematic Parameters of The Human Hand. Bioengineering (Basel) 2023; 10:bioengineering10030324. [PMID: 36978716 PMCID: PMC10045424 DOI: 10.3390/bioengineering10030324] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2023] [Revised: 02/24/2023] [Accepted: 02/27/2023] [Indexed: 03/08/2023] Open
Abstract
For technical or medical applications, the knowledge of the exact kinematics of the human hand is key to utilizing its capability of handling and manipulating objects and communicating with other humans or machines. The optimal relationship between the number of measurement parameters, measurement accuracy, as well as complexity, usability and cost of the measuring systems is hard to find. Biomechanic assumptions, the concepts of a biomechatronic system and the mechatronic design process, as well as commercially available components, are used to develop a sensorized glove. The proposed wearable introduced in this paper can measure 14 of 15 angular values of a simplified hand model. Additionally, five contact pressure values at the fingertips and inertial data of the whole hand with six degrees of freedom are gathered. Due to the modular design and a hand size examination based on anthropometric parameters, the concept of the wearable is applicable to a large variety of hand sizes and adaptable to different use cases. Validations show a combined root-mean-square error of 0.99° to 2.38° for the measurement of all joint angles on one finger, surpassing the human perception threshold and the current state-of-the-art in science and technology for comparable systems.
Collapse
Affiliation(s)
- Jonas Paul David
- Fachgebiet Biomechatronik, Institut für Mechatronische Systemintegration, Fakultät für Maschinenbau, Technische Universität Ilmenau, 98693 Ilmenau, Germany (T.H.)
- neuroConn GmbH, Albert-Einstein-Straße 3, 98693 Ilmenau, Germany
| | - Thomas Helbig
- Fachgebiet Biomechatronik, Institut für Mechatronische Systemintegration, Fakultät für Maschinenbau, Technische Universität Ilmenau, 98693 Ilmenau, Germany (T.H.)
| | - Hartmut Witte
- Fachgebiet Biomechatronik, Institut für Mechatronische Systemintegration, Fakultät für Maschinenbau, Technische Universität Ilmenau, 98693 Ilmenau, Germany (T.H.)
- Correspondence: ; Tel.: +49-(0)-3677-69-2456
| |
Collapse
|
5
|
Pan TY, Tsai WL, Chang CY, Yeh CW, Hu MC. A Hierarchical Hand Gesture Recognition Framework for Sports Referee Training-Based EMG and Accelerometer Sensors. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:3172-3183. [PMID: 32776885 DOI: 10.1109/tcyb.2020.3007173] [Citation(s) in RCA: 18] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/27/2023]
Abstract
To cultivate professional sports referees, we develop a sports referee training system, which can recognize whether a trainee wearing the Myo armband makes correct judging signals while watching a prerecorded professional game. The system has to correctly recognize a set of gestures related to official referee's signals (ORSs) and another set of gestures used to intuitively interact with the system. These two gesture sets involve both large motion and subtle motion gestures, and the existing sensor-based methods using handcrafted features do not work well on recognizing all kinds of these gestures. In this work, deep belief networks (DBNs) are utilized to learn more representative features for hand gesture recognition, and selective handcrafted features are combined with the DBN features to achieve more robust recognition results. Moreover, a hierarchical recognition scheme is designed to first recognize the input gesture as a large or subtle motion gesture, and the corresponding classifiers for large motion gestures and subtle motion gestures are further used to obtain the final recognition result. Moreover, the Myo armband consists of eight-channel surface electromyography (sEMG) sensors and an inertial measurement unit (IMU), and these heterogeneous signals can be fused to achieve better recognition accuracy. We take basketball as an example to validate the proposed training system, and the experimental results show that the proposed hierarchical scheme considering DBN features of multimodality data outperforms other methods.
Collapse
|
6
|
Karnam NK, Dubey SR, Turlapaty AC, Gokaraju B. EMGHandNet: A hybrid CNN and Bi-LSTM architecture for hand activity classification using surface EMG signals. Biocybern Biomed Eng 2022. [DOI: 10.1016/j.bbe.2022.02.005] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
7
|
Farooq U, Rahim MSM, Sabir N, Hussain A, Abid A. Advances in machine translation for sign language: approaches, limitations, and challenges. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-06079-3] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
8
|
Lv Z, Xiao F, Wu Z, Liu Z, Wang Y. Hand gestures recognition from surface electromyogram signal based on self-organizing mapping and radial basis function network. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2021.102629] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
9
|
Jiang S, Kang P, Song X, Lo B, Shull P. Emerging Wearable Interfaces and Algorithms for Hand Gesture Recognition: A Survey. IEEE Rev Biomed Eng 2021; 15:85-102. [PMID: 33961564 DOI: 10.1109/rbme.2021.3078190] [Citation(s) in RCA: 26] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Hands are vital in a wide range of fundamental daily activities, and neurological diseases that impede hand function can significantly affect quality of life. Wearable hand gesture interfaces hold promise to restore and assist hand function and to enhance human-human and human-computer communication. The purpose of this review is to synthesize current novel sensing interfaces and algorithms for hand gesture recognition, and the scope of applications covers rehabilitation, prosthesis control, sign language recognition, and human-computer interaction. Results showed that electrical, dynamic, acoustical/vibratory, and optical sensing were the primary input modalities in gesture recognition interfaces. Two categories of algorithms were identified: 1) classification algorithms for predefined, fixed hand poses and 2) regression algorithms for continuous finger and wrist joint angles. Conventional machine learning algorithms, including linear discriminant analysis, support vector machines, random forests, and non-negative matrix factorization, have been widely used for a variety of gesture recognition applications, and deep learning algorithms have more recently been applied to further facilitate the complex relationship between sensor signals and multi-articulated hand postures. Future research should focus on increasing recognition accuracy with larger hand gesture datasets, improving reliability and robustness for daily use outside of the laboratory, and developing softer, less obtrusive interfaces.
Collapse
|
10
|
Kudrinko K, Flavin E, Zhu X, Li Q. Wearable Sensor-Based Sign Language Recognition: A Comprehensive Review. IEEE Rev Biomed Eng 2020; 14:82-97. [PMID: 32845843 DOI: 10.1109/rbme.2020.3019769] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Sign language is used as a primary form of communication by many people who are Deaf, deafened, hard of hearing, and non-verbal. Communication barriers exist for members of these populations during daily interactions with those who are unable to understand or use sign language. Advancements in technology and machine learning techniques have led to the development of innovative approaches for gesture recognition. This literature review focuses on analyzing studies that use wearable sensor-based systems to classify sign language gestures. A review of 72 studies from 1991 to 2019 was performed to identify trends, best practices, and common challenges. Attributes including sign language variation, sensor configuration, classification method, study design, and performance metrics were analyzed and compared. Results from this literature review could aid in the development of user-centred and robust wearable sensor-based systems for sign language recognition.
Collapse
|
11
|
Mendes Junior JJA, Freitas MLB, Campos DP, Farinelli FA, Stevan SL, Pichorim SF. Analysis of Influence of Segmentation, Features, and Classification in sEMG Processing: A Case Study of Recognition of Brazilian Sign Language Alphabet. SENSORS (BASEL, SWITZERLAND) 2020; 20:E4359. [PMID: 32764286 PMCID: PMC7471999 DOI: 10.3390/s20164359] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/09/2020] [Revised: 07/26/2020] [Accepted: 08/01/2020] [Indexed: 11/17/2022]
Abstract
Sign Language recognition systems aid communication among deaf people, hearing impaired people, and speakers. One of the types of signals that has seen increased studies and that can be used as input for these systems is surface electromyography (sEMG). This work presents the recognition of a set of alphabet gestures from Brazilian Sign Language (Libras) using sEMG acquired from an armband. Only sEMG signals were used as input. Signals from 12 subjects were acquired using a MyoTM armband for the 26 signs of the Libras alphabet. Additionally, as the sEMG has several signal processing parameters, the influence of segmentation, feature extraction, and classification was considered at each step of the pattern recognition. In segmentation, window length and the presence of four levels of overlap rates were analyzed, as well as the contribution of each feature, the literature feature sets, and new feature sets proposed for different classifiers. We found that the overlap rate had a high influence on this task. Accuracies in the order of 99% were achieved for the following factors: segments of 1.75 s with a 12.5% overlap rate; the proposed set of four features; and random forest (RF) classifiers.
Collapse
Affiliation(s)
- José Jair Alves Mendes Junior
- Graduate Program in Electrical Engineering and Industrial Informatics (CPGEI), Federal University of Technology–Paraná (UTFPR), Curitiba (PR) 80230-901, Brazil; (J.J.A.M.J.); (F.A.F.); (S.F.P.)
| | - Melissa La Banca Freitas
- Graduate Program in Electrical Engineering (PPGEE), Federal University of Technology–Paraná (UTFPR), Ponta Grossa (PR) 84017-220, Brazil;
| | - Daniel Prado Campos
- Graduate Program in Biomedical Engineering (PPGEB), Federal University of Technology–Paraná (UTFPR), Curitiba (PR) 80230-901, Brazil;
| | - Felipe Adalberto Farinelli
- Graduate Program in Electrical Engineering and Industrial Informatics (CPGEI), Federal University of Technology–Paraná (UTFPR), Curitiba (PR) 80230-901, Brazil; (J.J.A.M.J.); (F.A.F.); (S.F.P.)
| | - Sergio Luiz Stevan
- Graduate Program in Electrical Engineering (PPGEE), Federal University of Technology–Paraná (UTFPR), Ponta Grossa (PR) 84017-220, Brazil;
| | - Sérgio Francisco Pichorim
- Graduate Program in Electrical Engineering and Industrial Informatics (CPGEI), Federal University of Technology–Paraná (UTFPR), Curitiba (PR) 80230-901, Brazil; (J.J.A.M.J.); (F.A.F.); (S.F.P.)
| |
Collapse
|
12
|
Jiang X, Satapathy SC, Yang L, Wang SH, Zhang YD. A Survey on Artificial Intelligence in Chinese Sign Language Recognition. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2020. [DOI: 10.1007/s13369-020-04758-2] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
13
|
Yu Y, Chen X, Cao S, Zhang X, Chen X. Exploration of Chinese Sign Language Recognition Using Wearable Sensors Based on Deep Belief Net. IEEE J Biomed Health Inform 2019; 24:1310-1320. [PMID: 31536027 DOI: 10.1109/jbhi.2019.2941535] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
In this paper, deep belief net (DBN) was applied into the field of wearable-sensor based Chinese sign language (CSL) recognition. Eight subjects were involved in the study, and all of the subjects finished a five-day experiment performing CSL on a target word set consisting of 150 CSL subwords. During the experiment, surface electromyography (sEMG), accelerometer (ACC), and gyroscope (GYRO) signals were collected from the participants. In order to obtain the optimal structure of the network, three different sensor fusion strategies, including data-level fusion, feature-level fusion, and decision-level fusion, were explored. In addition, for the feature-level fusion strategy, two different feature sources, which are hand-crafted features and network generated features, and two different network structures, which are fully-connected net and DBN, were also compared. The result showed that feature level fusion could achieve the best recognition accuracy among the three fusion strategies, and feature-level fusion with network generated features and DBN could achieve the best recognition accuracy. The best recognition accuracy realized in this study was 95.1% for the user-dependent test and 88.2% for the user-independent test. The significance of the study is that it applied the deep learning method into the field of wearable sensors-based CSL recognition, and according to our knowledge it's the first study comparing human engineered features with the network generated features in the correspondent field. The results from the study shed lights on the method of using network-generated features during sensor fusion and CSL recognition.
Collapse
|
14
|
Non-Touch Sign Word Recognition Based on Dynamic Hand Gesture Using Hybrid Segmentation and CNN Feature Fusion. APPLIED SCIENCES-BASEL 2019. [DOI: 10.3390/app9183790] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Hand gesture-based sign language recognition is a prosperous application of human– computer interaction (HCI), where the deaf community, hard of hearing, and deaf family members communicate with the help of a computer device. To help the deaf community, this paper presents a non-touch sign word recognition system that translates the gesture of a sign word into text. However, the uncontrolled environment, perspective light diversity, and partial occlusion may greatly affect the reliability of hand gesture recognition. From this point of view, a hybrid segmentation technique including YCbCr and SkinMask segmentation is developed to identify the hand and extract the feature using the feature fusion of the convolutional neural network (CNN). YCbCr performs image conversion, binarization, erosion, and eventually filling the hole to obtain the segmented images. SkinMask images are obtained by matching the color of the hand. Finally, a multiclass SVM classifier is used to classify the hand gestures of a sign word. As a result, the sign of twenty common words is evaluated in real time, and the test results confirm that this system can not only obtain better-segmented images but also has a higher recognition rate than the conventional ones.
Collapse
|
15
|
An Recognition-Verification Mechanism for Real-Time Chinese Sign Language Recognition Based on Multi-Information Fusion. SENSORS 2019; 19:s19112495. [PMID: 31159240 PMCID: PMC6603597 DOI: 10.3390/s19112495] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/04/2019] [Revised: 05/18/2019] [Accepted: 05/27/2019] [Indexed: 11/30/2022]
Abstract
For online sign language recognition (SLR) based on inertial measurement unit (IMU) and a surface electromyography (sEMG) sensor, achieving high-accuracy is a major challenge. The traditional method for that is the segmentation–recognition mechanism, which has two key challenges: (1) it is difficult to design a highly robust segmentation method for online data with inconspicuous segmentation information; and (2) the diversity of input data will increase the burden of the classification. The recognition–verification mechanism was proposed to improve the performance of online SLR. In the recognition stage, we used sliding windows to pull the data, and applied a convolutional neural network (CNN) to classify the sign language signal. In the verification stage, the confidence was evaluated by the Siamese network to judge the correctness of the classification results. The accuracy and rapidity of the classification model were discussed for 86 categories of Chinese sign language. In the experiments for online SLR, the superiority of the recognition–verification mechanism compared to the segmentation–recognition mechanism was verified.
Collapse
|
16
|
He J, Luo H, Jia J, Yeow JTW, Jiang N. Wrist and Finger Gesture Recognition With Single-Element Ultrasound Signals: A Comparison With Single-Channel Surface Electromyogram. IEEE Trans Biomed Eng 2019; 66:1277-1284. [DOI: 10.1109/tbme.2018.2872593] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
17
|
Shull PB, Jiang S, Zhu Y, Zhu X. Hand Gesture Recognition and Finger Angle Estimation via Wrist-Worn Modified Barometric Pressure Sensing. IEEE Trans Neural Syst Rehabil Eng 2019; 27:724-732. [PMID: 30892217 DOI: 10.1109/tnsre.2019.2905658] [Citation(s) in RCA: 43] [Impact Index Per Article: 7.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
This paper presents a new approach to wearable hand gesture recognition and finger angle estimation based on the modified barometric pressure sensing. Barometric pressure sensors were encased and injected with VytaFlex rubber such that the rubber directly contacted the sensing element allowing pressure change detection when the encasing rubber was pressed. A wearable prototype consisting of an array of ten modified barometric pressure sensors around the wrist was developed and validated with experimental testing for three different hand gesture sets and finger flexion/extension trials for each of the five fingers. The overall hand gesture recognition classification accuracy was 94%. Further analysis revealed that the most important sensor location was the underside of the wrist and that when reducing the sensor number to only five optimally placed sensors, classification accuracy was still 90%. For continuous finger angle estimation, aggregate R2 values between actual and predicted angles were thumb: 0.81 ± 0.10, index finger: 0.85±0.06, middle finger: 0.77±0.08, ring finger: 0.77 ± 0.12, and pinkie finger: 0.75 ± 0.10, and the overall average was 0.79 ± 0.05. These results demonstrate that a modified barometric pressure wristband can be used to classify hand gestures and to estimate individual finger joint angles. This approach could serve to improve the clinical treatment for upper extremity deficiencies, such as for stroke rehabilitation, by providing objective patient motor control metrics to inform and aid physicians and therapists throughout the rehabilitation process.
Collapse
|
18
|
Wei W, Dai Q, Wong Y, Hu Y, Kankanhalli M, Geng W. Surface-Electromyography-Based Gesture Recognition by Multi-View Deep Learning. IEEE Trans Biomed Eng 2019; 66:2964-2973. [PMID: 30762526 DOI: 10.1109/tbme.2019.2899222] [Citation(s) in RCA: 84] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Gesture recognition using sparse multichannel surface electromyography (sEMG) is a challenging problem, and the solutions are far from optimal from the point of view of muscle-computer interface. In this paper, we address this problem from the context of multi-view deep learning. A novel multi-view convolutional neural network (CNN) framework is proposed by combining classical sEMG feature sets with a CNN-based deep learning model. The framework consists of two parts. In the first part, multi-view representations of sEMG are modeled in parallel by a multistream CNN, and a performance-based view construction strategy is proposed to choose the most discriminative views from classical feature sets for sEMG-based gesture recognition. In the second part, the learned multi-view deep features are fused through a view aggregation network composed of early and late fusion subnetworks, taking advantage of both early and late fusion of learned multi-view deep features. Evaluations on 11 sparse multichannel sEMG databases as well as five databases with both sEMG and inertial measurement unit data demonstrate that our multi-view framework outperforms single-view methods on both unimodal and multimodal sEMG data streams.
Collapse
|
19
|
Kim S, Kim J, Ahn S, Kim Y. Finger language recognition based on ensemble artificial neural network learning using armband EMG sensors. Technol Health Care 2018; 26:249-258. [PMID: 29710753 PMCID: PMC6005006 DOI: 10.3233/thc-174602] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
BACKGROUND Deaf people use sign or finger languages for communication, but these methods of communication are very specialized. For this reason, the deaf can suffer from social inequalities and financial losses due to their communication restrictions. OBJECTIVE In this study, we developed a finger language recognition algorithm based on an ensemble artificial neural network (E-ANN) using an armband system with 8-channel electromyography (EMG) sensors. METHODS The developed algorithm was composed of signal acquisition, filtering, segmentation, feature extraction and an E-ANN based classifier that was evaluated with the Korean finger language (14 consonants, 17 vowels and 7 numbers) in 17 subjects. E-ANN was categorized according to the number of classifiers (1 to 10) and size of training data (50 to 1500). The accuracy of the E-ANN-based classifier was obtained by 5-fold cross validation and compared with an artificial neural network (ANN)-based classifier. RESULTS AND CONCLUSIONS As the number of classifiers (1 to 8) and size of training data (50 to 300) increased, the average accuracy of the E-ANN-based classifier increased and the standard deviation decreased. The optimal E-ANN was composed with eight classifiers and 300 size of training data, and the accuracy of the E-ANN was significantly higher than that of the general ANN.
Collapse
Affiliation(s)
| | | | | | - Youngho Kim
- Corresponding author: Youngho Kim, Department of Biomedical Engineering, Yonsei University, Wonju, Korea. Tel.: +82 33 760 2492; E-mail: .
| |
Collapse
|
20
|
Motor Function Evaluation of Hemiplegic Upper-Extremities Using Data Fusion from Wearable Inertial and Surface EMG Sensors. SENSORS 2017; 17:s17030582. [PMID: 28335394 PMCID: PMC5375868 DOI: 10.3390/s17030582] [Citation(s) in RCA: 29] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/08/2016] [Revised: 02/06/2017] [Accepted: 02/23/2017] [Indexed: 11/25/2022]
Abstract
Quantitative evaluation of motor function is of great demand for monitoring clinical outcome of applied interventions and further guiding the establishment of therapeutic protocol. This study proposes a novel framework for evaluating upper limb motor function based on data fusion from inertial measurement units (IMUs) and surface electromyography (EMG) sensors. With wearable sensors worn on the tested upper limbs, subjects were asked to perform eleven straightforward, specifically designed canonical upper-limb functional tasks. A series of machine learning algorithms were applied to the recorded motion data to produce evaluation indicators, which is able to reflect the level of upper-limb motor function abnormality. Sixteen healthy subjects and eighteen stroke subjects with substantial hemiparesis were recruited in the experiment. The combined IMU and EMG data yielded superior performance over the IMU data alone and the EMG data alone, in terms of decreased normal data variation rate (NDVR) and improved determination coefficient (DC) from a regression analysis between the derived indicator and routine clinical assessment score. Three common unsupervised learning algorithms achieved comparable performance with NDVR around 10% and strong DC around 0.85. By contrast, the use of a supervised algorithm was able to dramatically decrease the NDVR to 6.55%. With the proposed framework, all the produced indicators demonstrated high agreement with the routine clinical assessment scale, indicating their capability of assessing upper-limb motor functions. This study offers a feasible solution to motor function assessment in an objective and quantitative manner, especially suitable for home and community use.
Collapse
|
21
|
A Wearable System for Recognizing American Sign Language in Real-Time Using IMU and Surface EMG Sensors. IEEE J Biomed Health Inform 2016; 20:1281-1290. [DOI: 10.1109/jbhi.2016.2598302] [Citation(s) in RCA: 135] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
22
|
Yang X, Chen X, Cao X, Wei S, Zhang X. Chinese Sign Language Recognition Based on an Optimized Tree-Structure Framework. IEEE J Biomed Health Inform 2016; 21:994-1004. [PMID: 27164613 DOI: 10.1109/jbhi.2016.2560907] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Chinese Sign Language (CSL) subword recognition based on surface electromyography (sEMG), accelerometer (ACC), and gyroscope (GYRO) sensors was explored in this paper. In order to fuse effectively the information of these three kinds of sensors, the classification abilities of sEMG, ACC, GYRO, and their combinations in three common sign components (one or two handed, hand orientation, and hand amplitude) were evaluated first and then an optimized tree-structure classification framework was proposed for CSL subword recognition. Eight subjects participated in this study and recognition experiments under different testing conditions were implemented on a target set consisting of 150 CSL subwords. The proposed optimized tree-structure classification framework based on sEMG, ACC, and GYRO obtained the best performance among seven different testing conditions with single sensor, paired-sensor fusion, and three-sensor fusion, and the overall recognition accuracies of 94.31% and 87.02% were obtained for 150 CSL subwords in a user-specific test and user-independent test, respectively. Our study could lay a basis for the implementation of large-vocabulary sign language recognition system based on sEMG, ACC, and GYRO sensors.
Collapse
|
23
|
Abstract
This paper focuses on the development of an upper limb rehabilitation training system designed for use by children with cerebral palsy (CP). It attempts to meet the requirements of in-home training by taking advantage of the combination of portable accelerometers (ACC) and surface electromyography (SEMG) sensors worn on the upper limb to capture functional movements. In the proposed system, the EMG-ACC acquisition device works essentially as wireless game controller, and three rehabilitation games were designed for improving upper limb motor function under a clinician's guidance. The games were developed on the Android platform based on a physical engine called Box2D. The results of a system performance test demonstrated that the developed games can respond to the upper limb actions within 210 ms. Positive questionnaire feedbacks from twenty CP subjects who participated in the game test verified both the feasibility and usability of the system. Results of a long-term game training conducted with three CP subjects demonstrated that CP patients could improve in their game performance through repetitive training, and persistent training was needed to improve and enhance the rehabilitation effect. According to our experimental results, the novel multi-feedback SEMG-ACC-based user interface improved the users' initiative and performance in rehabilitation training.
Collapse
|
24
|
A Component-Based Vocabulary-Extensible Sign Language Gesture Recognition Framework. SENSORS 2016; 16:s16040556. [PMID: 27104534 PMCID: PMC4851070 DOI: 10.3390/s16040556] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/03/2016] [Revised: 04/12/2016] [Accepted: 04/13/2016] [Indexed: 11/23/2022]
Abstract
Sign language recognition (SLR) can provide a helpful tool for the communication between the deaf and the external world. This paper proposed a component-based vocabulary extensible SLR framework using data from surface electromyographic (sEMG) sensors, accelerometers (ACC), and gyroscopes (GYRO). In this framework, a sign word was considered to be a combination of five common sign components, including hand shape, axis, orientation, rotation, and trajectory, and sign classification was implemented based on the recognition of five components. Especially, the proposed SLR framework consisted of two major parts. The first part was to obtain the component-based form of sign gestures and establish the code table of target sign gesture set using data from a reference subject. In the second part, which was designed for new users, component classifiers were trained using a training set suggested by the reference subject and the classification of unknown gestures was performed with a code matching method. Five subjects participated in this study and recognition experiments under different size of training sets were implemented on a target gesture set consisting of 110 frequently-used Chinese Sign Language (CSL) sign words. The experimental results demonstrated that the proposed framework can realize large-scale gesture set recognition with a small-scale training set. With the smallest training sets (containing about one-third gestures of the target gesture set) suggested by two reference subjects, (82.6 ± 13.2)% and (79.7 ± 13.4)% average recognition accuracy were obtained for 110 words respectively, and the average recognition accuracy climbed up to (88 ± 13.7)% and (86.3 ± 13.7)% when the training set included 50~60 gestures (about half of the target gesture set). The proposed framework can significantly reduce the user’s training burden in large-scale gesture recognition, which will facilitate the implementation of a practical SLR system.
Collapse
|
25
|
Su R, Chen X, Cao S, Zhang X. Random Forest-Based Recognition of Isolated Sign Language Subwords Using Data from Accelerometers and Surface Electromyographic Sensors. SENSORS 2016; 16:s16010100. [PMID: 26784195 PMCID: PMC4732133 DOI: 10.3390/s16010100] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/12/2015] [Revised: 01/10/2016] [Accepted: 01/12/2016] [Indexed: 12/05/2022]
Abstract
Sign language recognition (SLR) has been widely used for communication amongst the hearing-impaired and non-verbal community. This paper proposes an accurate and robust SLR framework using an improved decision tree as the base classifier of random forests. This framework was used to recognize Chinese sign language subwords using recordings from a pair of portable devices worn on both arms consisting of accelerometers (ACC) and surface electromyography (sEMG) sensors. The experimental results demonstrated the validity of the proposed random forest-based method for recognition of Chinese sign language (CSL) subwords. With the proposed method, 98.25% average accuracy was obtained for the classification of a list of 121 frequently used CSL subwords. Moreover, the random forests method demonstrated a superior performance in resisting the impact of bad training samples. When the proportion of bad samples in the training set reached 50%, the recognition error rate of the random forest-based method was only 10.67%, while that of a single decision tree adopted in our previous work was almost 27.5%. Our study offers a practical way of realizing a robust and wearable EMG-ACC-based SLR systems.
Collapse
Affiliation(s)
- Ruiliang Su
- Department of Electronic Science and Technology, University of Science and Technology of China, Hefei 230027, China.
| | - Xiang Chen
- Department of Electronic Science and Technology, University of Science and Technology of China, Hefei 230027, China.
| | - Shuai Cao
- Department of Electronic Science and Technology, University of Science and Technology of China, Hefei 230027, China.
| | - Xu Zhang
- Department of Electronic Science and Technology, University of Science and Technology of China, Hefei 230027, China.
| |
Collapse
|
26
|
|
27
|
Atzori M, Gijsberts A, Müller H, Caputo B. Classification of hand movements in amputated subjects by sEMG and accelerometers. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2015; 2014:3545-9. [PMID: 25570756 DOI: 10.1109/embc.2014.6944388] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Numerous recent studies have aimed to improve myoelectric control of prostheses. However, the majority of these studies is characterized by two problems that could be easily fulfilled with recent resources supplied by the scientific literature. First, the majority of these studies use only intact subjects, with the unproved assumption that the results apply equally to amputees. Second, usually only electromyography data are used, despite other sensors (e.g., accelerometers) being easy to include into a real life prosthesis control system. In this paper we analyze the mentioned problems by the classification of 40 hand movements in 5 amputated and 40 intact subjects, using both sEMG and accelerometry data and applying several different state of the art methods. The datasets come from the NinaPro database, which supplies publicly available sEMG data to develop and test machine learning algorithms for prosthetics. The number of subjects can seem small at first sight, but it is not considering the literature of the field (which has to face the difficulty of recruiting trans-radial hand amputated subjects). Our results indicate that the maximum average classification accuracy for amputated subjects is 61.14%, which is just 15.86% less than intact subjects, and they show that intact subjects results can be used as proxy measure for amputated subjects. Finally, our comparison shows that accelerometry as a modality is less affected by amputation than electromyography, suggesting that real life prosthetics performance may easily be improved by inclusion of accelerometers.
Collapse
|
28
|
A Novel Phonology- and Radical-Coded Chinese Sign Language Recognition Framework Using Accelerometer and Surface Electromyography Sensors. SENSORS 2015; 15:23303-24. [PMID: 26389907 PMCID: PMC4610461 DOI: 10.3390/s150923303] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/24/2015] [Revised: 09/04/2015] [Accepted: 09/08/2015] [Indexed: 11/17/2022]
Abstract
Sign language recognition (SLR) is an important communication tool between the deaf and the external world. It is highly necessary to develop a worldwide continuous and large-vocabulary-scale SLR system for practical usage. In this paper, we propose a novel phonology- and radical-coded Chinese SLR framework to demonstrate the feasibility of continuous SLR using accelerometer (ACC) and surface electromyography (sEMG) sensors. The continuous Chinese characters, consisting of coded sign gestures, are first segmented into active segments using EMG signals by means of moving average algorithm. Then, features of each component are extracted from both ACC and sEMG signals of active segments (i.e., palm orientation represented by the mean and variance of ACC signals, hand movement represented by the fixed-point ACC sequence, and hand shape represented by both the mean absolute value (MAV) and autoregressive model coefficients (ARs)). Afterwards, palm orientation is first classified, distinguishing “Palm Downward” sign gestures from “Palm Inward” ones. Only the “Palm Inward” gestures are sent for further hand movement and hand shape recognition by dynamic time warping (DTW) algorithm and hidden Markov models (HMM) respectively. Finally, component recognition results are integrated to identify one certain coded gesture. Experimental results demonstrate that the proposed SLR framework with a vocabulary scale of 223 characters can achieve an averaged recognition accuracy of 96.01% ± 0.83% for coded gesture recognition tasks and 92.73% ± 1.47% for character recognition tasks. Besides, it demonstrats that sEMG signals are rather consistent for a given hand shape independent of hand movements. Hence, the number of training samples will not be significantly increased when the vocabulary scale increases, since not only the number of the completely new proposed coded gestures is constant and limited, but also the transition movement which connects successive signs needs no training samples to model even though the same coded gesture performed in different characters. This work opens up a possible new way to realize a practical Chinese SLR system.
Collapse
|
29
|
Martin-Martin J, Cuesta-Vargas AI. Quantification of functional hand grip using electromyography and inertial sensor-derived accelerations: clinical implications. Biomed Eng Online 2014; 13:161. [PMID: 25496710 PMCID: PMC4273320 DOI: 10.1186/1475-925x-13-161] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2014] [Accepted: 12/04/2014] [Indexed: 11/10/2022] Open
Abstract
BACKGROUND Assessing hand injury is of great interest given the level of involvement of the hand with the environment. Knowing different assessment systems and their limitations generates new perspectives. The integration of digital systems (accelerometry and electromyography) as a tool to supplement functional assessment allows the clinician to know more about the motor component and its relation to movement. Therefore, the purpose of this study was the kinematic and electromyography analysis during functional hand movements. METHOD Ten subjects carried out six functional movements (terminal pinch, termino-lateral pinch, tripod pinch, power grip, extension grip and ball grip). Muscle activity (hand and forearm) was measured in real time using electromyograms, acquired with the Mega ME 6000, whilst acceleration was measured using the AcceleGlove. RESULTS Electrical activity and acceleration variables were recorded simultaneously during the carrying out of the functional movements. The acceleration outcome variables were the modular vectors of each finger of the hand and the palm. In the electromyography, the main variables were normalized by the mean and by the maximum muscle activity of the thenar region, hypothenar, first interosseous dorsal, wrist flexors, carpal flexors and wrist extensors. CONCLUSIONS Knowing muscle behavior allows the clinician to take a more direct approach in the treatment. Based on the results, the tripod grip shows greater kinetic activity and the middle finger is the most relevant in this regard. Ball grip involves most muscle activity, with the thenar region playing a fundamental role in hand activity. CLINICAL RELEVANCE Relating muscle activation, movements, individual load and displacement offers the possibility to proceed with rehabilitation by individual component.
Collapse
Affiliation(s)
- Jaime Martin-Martin
- />Departamento de Psiquiatría y Fisioterapia, Facultad de Ciencias de la Salud, Universidad de Malaga, Andalucia Tech, Instituto de Biomedicina de Malaga (IBIMA), Grupo de Clinimetria (AE-14), Malaga, Spain
| | - Antonio I Cuesta-Vargas
- />Departamento de Psiquiatría y Fisioterapia, Facultad de Ciencias de la Salud, Universidad de Malaga, Andalucia Tech, Instituto de Biomedicina de Malaga (IBIMA), Grupo de Clinimetria (AE-14), Malaga, Spain
- />School of Clinical Science, Faculty of Health Science, Queensland University Technology, Brisbane, Australia
| |
Collapse
|
30
|
Ramos-Garcia RI, Muth ER, Gowdy JN, Hoover AW. Improving the recognition of eating gestures using intergesture sequential dependencies. IEEE J Biomed Health Inform 2014; 19:825-31. [PMID: 24919205 DOI: 10.1109/jbhi.2014.2329137] [Citation(s) in RCA: 35] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
This paper considers the problem of recognizing eating gestures by tracking wrist motion. Eating gestures are activities commonly undertaken during the consumption of a meal, such as sipping a drink of liquid or using utensils to cut food. Each of these gestures causes a pattern of wrist motion that can be tracked to automatically identify the activity. Previous works have studied this problem at the level of a single gesture. In this paper, we demonstrate that individual gestures have sequential dependence. To study this, three types of classifiers were built: 1) a K-nearest neighbor classifier which uses no sequential context, 2) a hidden Markov model (HMM) which captures the sequential context of subgesture motions, and 3) HMMs that model intergesture sequential dependencies. We built first-order to sixth-order HMMs to evaluate the usefulness of increasing amounts of sequential dependence to aid recognition. On a dataset of 25 meals, we found that the baseline accuracies for the KNN and the subgesture HMM classifiers were 75.8% and 84.3%, respectively. Using HMMs that model intergesture sequential dependencies, we were able to increase accuracy to up to 96.5%. These results demonstrate that sequential dependencies exist between eating gestures and that they can be exploited to improve recognition accuracy.
Collapse
|