Startseite Real-time classification of hand movements as a basis for intuitive control of grasp neuroprostheses
Artikel Open Access

Real-time classification of hand movements as a basis for intuitive control of grasp neuroprostheses

  • Dmitry Amelin EMAIL logo , Ivan Potapov , Josep Cardona Audí , Andreas Kogut , Rüdiger Rupp , Roman Ruff und Klaus-Peter Hoffmann
Veröffentlicht/Copyright: 19. Oktober 2020
Veröffentlichen auch Sie bei De Gruyter Brill

Abstract

This paper reports on the evaluation of recurrent and convolutional neural networks as real-time grasp phase classifiers for future control of neuroprostheses for people with high spinal cord injury. A field-programmable gate array has been chosen as an implementation platform due to its form factor and ability to perform parallel computations, which are specific for the selected neural networks. Three different phases of two grasp patterns and the additional open hand pattern were predicted by means of surface Electromyography (EMG) signals (i.e. Seven classes in total). Across seven healthy subjects, CNN (Convolutional Neural Networks) and RNN (Recurrent Neural Networks) had a mean accuracy of 85.23% with a standard deviation of 4.77% and 112 µs per prediction and 83.30% with a standard deviation of 4.36% and 40 µs per prediction, respectively.

Introduction

According to a report from the World Health Organisation [1], each year the population of people with spinal cord injury (SCI) increases by up to 500,000. With grasp neuroprostheses based on functional electrical stimulation (FES), people with tetraplegia are able to perform basic grasping movements [2]. Although most potential users of grasp neuroprosthesis still have residual muscle activities in their upper extremities, established user interfaces do not take these into account. The objective of this work was to develop a method for classifying the user intent in real-time with respect to grasping, based on the electromyogram (EMG) of muscles relevant for grasping. In the future this method is intended to be used for more intuitive control of a grasp neuroprosthesis, helping to reduce the workload of end users.

Material and methods

Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN) were chosen as grasp classifiers [3]. To narrow the research focus, we evaluated the performance of Artificial Neural Networks (ANN) with a single convolutional layer in CNN and the long short-term memory (LSTM) layer in RNN models, which helps to overcome the following two problems: vanishing and exploding gradients [5]. The dimension of the input and output vectors corresponds to the number of EMG channels used and the number of prediction classes accordingly. We fed the neural networks with Root Mean Square (RMS) values of the EMG signals.

We used Python with Keras library to train (ANN) models by supervised learning. In order to reduce the chances of model overfitting, the dropout regularization method [4] was applied. Hyperparameters search was performed to find the best configuration by varying LSTM cells in recurrent layer and kernel number in convolutional one. Data from seven healthy subjects (age range 23–41 years; six males and one female) was used. ANN models were trained for each subject individually. The datasets were split into the training (75% of total), the test (25% of total), and the validation (20% of training data) sets. Each model was trained 3 times with the same configuration; given accuracy values are maximum for these trials.

A Field-Programmable Gate Array (FPGA) was selected as an implementation platform for the ANNs, because of its ability to perform mathematical operations in parallel and its form-factor which allows the end product to have wearable size. FPGA board Altera DE2-115 with Cyclone® IV EP4CE115 was used for classifiers implementation. The FPGA chip has 114,480 logic elements with 3,888 Kbits of embedded memory. CNN and RNN scalable architectures were developed on the FPGA with parameterized standalone modules. The trained model was transferred on the FPGA by extracting its weights and biases, converting them into the fixed-point format, and storing them in the embedded memory.

The used FPGA has some hardware restrictions, therefore some of the approximations were implemented as fixed-point (instead of floating-point) data representation and activation functions were substituted by lookup tables (LUT). For evaluation of the consequences of the approximations on classification accuracy, the ANNs were implemented both on Personal Computer (PC) hardware and FPGA.

As stated in [6] to reduce the memory and computation load in CNN implementation without affecting accuracy we removed the fully connected hidden layer and used a smaller number of kernels, which we defined during hyper parameters search.

Data acquisition setup

For this work, in total 90 repetitions of each lateral and cylindrical grasps were recorded per subject and analyzed at three different time points of the grasp execution: half-finished, completed, and completed with force. This results in seven classes: three different states of two grasp patterns and an additional open hand class.

The Leap Motion Controller was used to record finger motions during the grasp executions, which allowed identifying the phases of the grasps based on their relative position. To measure the force of the grasp, custom-designed handles (different for each lateral and cylindrical grasps) each equipped with one force sensor (Flexiforce A201) were used. Data from the sensors and IR camera was synchronized to allow for the automatic splitting measurements into proper data sets of the different grasp phases.

EMG was recorded with disposable Ag/AgCl electrodes (adhesive surface electrode 15 × 20 mm with snap connector, Spes Medica s.r.l.) with a sampling frequency of 1 kHz from the following muscles: extensor carpi radialis, extensor digitorum communis, abductor pollicis brevis, adductor pollicis, flexor pollicis longus, flexor digitorum superficialis. The RMS window length was 20 ms with a 10 ms overlap.

Results

Mean accuracy for FPGA-based CNN models was 85.23% (maximum 92.25%; minimum 78.37%; standard deviation 4.77%) and 83.30% for RNN (maximum 89.00%; minimum 76.73%; standard deviation 4.36%).

Almost similar prediction accuracies were achieved for floating-point PC and fixed-point (Q10.14) FPGA implementations (Table 1 and Table 2).

Table 1:

Accuracy of CNN, %.

SubjectPCFPGA
184.4984.40
278.3178.37
387.9087.78
492.2292.25
587.6887.72
680.6280.13
785.8485.98
  1. CNN, Convolutional Neural Networks.

Table 2:

Accuracy of RNN, %.

SubjectPCFPGA
185.0085.13
276.5976.73
386.9487.01
489.0589.00
584.5684.71
680.7580.80
779.5679.69
  1. RNN, Recurrent Neural Networks.

The single input calculation time is in the μs range for both FPGA ANN-implementations and therefore negligible in comparison to the feature extraction frame.

Based on the results of the performance evaluation, optimal configurations of 64 kernels and 16 LSTM cells were identified for CNN and RNN, respectively.

Table 3 shows that the CNN resulting model implemented on the FPGA has better accuracy than the RNN but requires more logic elements. Nevertheless, the biggest difference we can see in the memory requirements, due to RNN’s activation function implementation with LUT.

Table 3:

Performance and comparison of used resources between CNN and RNN models on the FPGA. Values in brackets correspond to the total used resources.

CNNRNN
Mean accuracy, %85.2383.30
Mean models match, %99.6799.33
Number of parameters26311591
Required memory, kB7.5 (75.4)52.6 (120.5)
Logic elements54276 (58370)48445 (52539)
Calculation time, µs11240
  1. CNN, Convolutional Neural Networks; RNN, Recurrent Neural Networks.

Conclusion

Most research on the use of ANNs for EMG parameter classification is focused on artificial hand prostheses [7], [8], [9]. In this work, we investigated CNN and RNN as potential classification method of grasp patterns and phases for future control of grasp neuroprostheses. The neural networks were trained to recognize three different states of two grasp patterns and the open hand pattern.

The difference in accuracy between PC and FPGA implementations is small. Moreover, it takes in the worst case 112 µs to process single data input, which is much faster than data transfer in some of the communication links, e.g. radiofrequency, and electromechanical delay of a neuroprosthesis [10]. This allows using FPGA implementations for real-time grasp classification and computer implementation for prototyping. On the other hand, achieved accuracy is not enough for medical use and, therefore, improvements have to be made. Among the future steps for enhancing accuracy are incorporating different features, besides RMS, as input to ANNs and potentially more complex ANNs architectures or other ANN types. Possible candidates are U-Net [11] for CNN and the Legendre Memory Unit [12] for RNN.

We used only three different phases of two grasp patterns, which increases the misclassification rate of the ANN’s and results in a relatively high amount of false-positives. As a future extension, we would like to include data from additional grasp patterns such as wrist flexion, wrist extension, pronation, and supination and evaluate the effect on ANN’s robustness and accuracy.


Corresponding author: Dmitry Amelin, Fraunhofer Institute for Biomedical Engineering, Joseph-von-Fraunhofer-Weg 1, 66280 Sulzbach, Germany, E-mail:

  1. Research funding: This research was supported by the German Federal Ministry of Education and Research (BMBF) within the Innovation Cluster INTAKT (16SV7637K).

  2. Author contributions: All authors have accepted responsibility for the entire content of this manuscript and approved its submission.

  3. Conflict of interest: Authors state no conflict of interest.

  4. Informed consent: Informed consent has been obtained from all individuals included in this study.

  5. Ethical approval: The research related to human use complies with all the relevant national regulations, institutional policies and was performed in accordance with the tenets of the Helsinki Declaration, and has been approved by the ethics committee of Heidelberg University (vote number S-016/2014).

References

1. Spinal cord injury, World Health Organization. [Accessed 10 Jan 2020] https://www.who.int/news-room/fact-sheets/detail/spinal-cord-injury.Suche in Google Scholar

2. Ho, CH, Triolo, RJ, Elias, AL, Kilgore, KL, DiMarco, AF, Bogie, K, et al. Functional electrical stimulation and spinal cord injury. Phys Med Rehabil Clin 2014;25:631–54. https://doi.org/10.1016/j.pmr.2014.05.001.Suche in Google Scholar

3. Aggarwal, CC. Neural Networks and Deep Learning: a Textbook. Switzerland: Springer International Publishing; 2018.10.1007/978-3-319-94463-0Suche in Google Scholar

4. Srivastava, N, Hinton, G, Krizhevsky, A, Sutskever, I, Salakhutdinov, R. Dropout: a simple way to prevent neural networks from over fitting. J Mach Learn Res 2014;15:1929–58.Suche in Google Scholar

5. Hochreiter, S, Schmidhuber, J. Long short-term memory. Neural Comput 1997;9:1735–80. https://doi.org/10.1162/neco.1997.9.8.1735.Suche in Google Scholar

6. Hailesellasie, M, Hasan, SR, Khalid, F, Wad, FA, Shafique, M. FPGA-based convolutional neural network architecture with reduced parameter requirements. In: 2018 IEEE International Symposium on Circuits and Systems (ISCAS). IEEE; 2018:1–5 p. https://doi.org/10.1109/ISCAS.2018.8351283.Suche in Google Scholar

7. Gandolla, M, Ferrante, S, Ferrigno, G, Baldassini, D, Molteni, F, Guanziroli, E, et al. Artificial neural network EMG classifier for functional hand grasp movements prediction. J Int Med Res 2017;45:1831–47. https://doi.org/10.1177/0300060516656689.Suche in Google Scholar

8. Zecca, M, Micera, S, Carrozza, MC, Dario, P. Control of multifunctional prosthetic hands by processing the electromyographic signal. Crit Rev Biomed Eng 2002;30:459–85. https://doi.org/10.1615/critrevbiomedeng.v30.i456.80.Suche in Google Scholar

9. Wojtczak, P, Amaral, TG, Dias, OP, Wolczowski, A, Kurzynski, M. Hand movement recognition based on biosignal analysis. Eng Appl Artif Intell 2009;22:608–15. https://doi.org/10.1016/j.engappai.2008.12.004.Suche in Google Scholar

10. Cavanagh, PR, Komi, PV. Electromechanical delay in human skeletal muscle under concentric and eccentric contractions. Eur J Appl Physiol 1979;42:159–63. https://doi.org/10.1007/bf00431022.Suche in Google Scholar

11. Ronneberger, O, Fischer, P, Brox, T. U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. Springer, Cham, Switzerland; 2015:234–41 p. https://doi.org/10.1007/978-3-319-24574-4_28.Suche in Google Scholar

12. Voelker, A, Kajić, I, Eliasmith, C. Legendre memory units: continuous-time representation in recurrent neural networks. In: Wallach, H, Larochelle, H, Beygelzimer, A, Alché-Buc, F, Fox, E, Garnett, R, editors. In advances in neural information processing systems 32: Curran Associates, Inc.; 2019:15570–9 pp.Suche in Google Scholar

Published Online: 2020-10-19

© 2020 Dmitry Amelin et al., published by De Gruyter, Berlin/Boston

This work is licensed under the Creative Commons Attribution 4.0 International License.

Heruntergeladen am 25.9.2025 von https://www.degruyterbrill.com/document/doi/10.1515/cdbme-2020-2011/html
Button zum nach oben scrollen