Paper ID: 2407.00014

Kinetic and Kinematic Sensors-free Approach for Estimation of Continuous Force and Gesture in sEMG Prosthetic Hands

Gang Liu, Zhenxiang Wang, Chuanmei Xi, Ziyang He, Shanshan Guo, Rui Zhang, Dezhong Yao

Regression-based sEMG prosthetic hands are widely used for their ability to provide continuous kinetic and kinematic parameters. However, establishing these models requires complex sensors systems to collect corresponding kinetic and kinematic data in synchronization with sEMG, which is cumbersome and user-unfriendly. This paper proposes a kinetic and kinematic sensors-free approach for controlling sEMG prosthetic hands, enabling continuous decoding and execution of three hand movements: individual finger flexion/extension, multiple finger flexion/extension, and fist opening/closing. This approach utilizes only two data points (-1 and 1), representing maximal finger flexion force label and extension force label respectively, and their corresponding sEMG data to establish a near-linear model based on sEMG data and labels. The model's output labels values are used to control the direction and magnitude of fingers forces, enabling the estimation of continuous gestures. To validate this approach, we conducted offline and online experiments using four models: Dendritic Net (DD), Linear Net (LN), Multi-Layer Perceptron (MLP), and Convolutional Neural Network (CNN). The offline analysis assessed each model's ability to classify finger force direction and interpolate intermediate force values, while online experiments evaluated real-time control performance in controlling gestures and accurately adjusting forces. Our results demonstrate that the DD and LN models provide excellent real-time control of finger forces and gestures, highlighting the practical potential of this sensors-free approach for prosthetic applications. This study significantly reduces the complexity of collecting kinetic and kinematic parameters in sEMG-based regression prosthetics, thus enhancing the usability and convenience of prosthetic hands.

Submitted: May 1, 2024