Post-stroke neurorehabilitation based on virtual therapies are performed completing repetitive exercises shown in visual electronic devices, whose content represents imaginary or daily life tasks. Currently, there are two ways of visualization of these task. 3D virtual environments are used to get a three dimensional space that represents the real world with a high level of detail, whose realism is determinated by the resolucion and fidelity of the objects of the task. Furthermore, 2D virtual environments are used to represent the tasks with a low degree of realism using techniques of bidimensional graphics. However, the type of visualization can influence the quality of perception of the task, affecting the patient's sensorimotor performance. The purpose of this paper was to evaluate if there were differences in patterns of kinematic movements when post-stroke patients performed a reach task viewing a virtual therapeutic game with two different type of visualization of virtual environment: 2D and 3D. Nine post-stroke patients have participated in the study receiving a virtual therapy assisted by PUPArm rehabilitation robot. Horizontal movements of the upper limb were performed to complete the aim of the tasks, which consist in reaching peripheral or perspective targets depending on the virtual environment shown. Various parameter types such as the maximum speed, reaction time, path length, or initial movement are analyzed from the data acquired objectively by the robotic device to evaluate the influence of the task visualization. At the end of the study, a usability survey was provided to each patient to analysis his/her satisfaction level. For all patients, the movement trajectories were enhanced when they completed the therapy. This fact suggests that patient's motor recovery was increased. Despite of the similarity in majority of the kinematic parameters, differences in reaction time and path length were higher using the 3D task. Regarding the success rates were very similar. In conclusion, the using of 2D environments in virtual therapy may be a more appropriate and comfortable way to perform tasks for upper limb rehabilitation of post-stroke patients, in terms of accuracy in order to effectuate optimal kinematic trajectories.
This paper presents a novel kinematic reconstruction of the human arm chain with five degrees of freedom and the estimation of the shoulder location during rehabilitation therapy assisted by end-effector robotic devices. This algorithm is based on the pseudoinverse of the Jacobian through the acceleration of the upper arm, measured using an accelerometer, and the orientation of the shoulder, estimated with a magnetic angular rate and gravity (MARG) device. The results show a high accuracy in terms of arm joints and shoulder movement with respect to the real arm measured through an optoelectronic system. Furthermore, the range of motion (ROM) of 50 healthy subjects is studied from two different trials, one trying to avoid shoulder movements and the second one forcing them. Moreover, the shoulder movement in the second trial is also estimated accurately. Besides the fact that the posture of the patient can be corrected during the exercise, the therapist could use the presented algorithm as an objective assessment tool. In conclusion, the joints’ estimation enables a better adjustment of the therapy, taking into account the needs of the patient, and consequently, the arm motion improves faster.
Assistive technologies help all persons with disabilities to improve their accessibility in all aspects of their life. The AIDE European project contributes to the improvement of current assistive technologies by developing and testing a modular and adaptive multimodal interface customizable to the individual needs of people with disabilities. This paper describes the computer vision algorithms part of the multimodal interface developed inside the AIDE European project. The main contribution of this computer vision part is the integration with the robotic system and with the other sensory systems (electrooculography (EOG) and electroencephalography (EEG)). The technical achievements solved herein are the algorithm for the selection of objects using the gaze, and especially the state-of-the-art algorithm for the efficient detection and pose estimation of textureless objects. These algorithms were tested in real conditions, and were thoroughly evaluated both qualitatively and quantitatively. The experimental results of the object selection algorithm were excellent (object selection over 90%) in less than 12 s. The detection and pose estimation algorithms evaluated using the LINEMOD database were similar to the state-of-the-art method, and were the most computationally efficient.
Conventional rehabilitation strategies for stroke survivors become difficult when voluntary movements are severely disturbed. Combining passive limb mobilization, robotic devices and EEG-based brain-computer interfaces (BCI) systems might improve treatment and clinical follow-up of these patients, but detailed knowledge of neurophysiological mechanisms involved in functional recovery, which might help for tailoring stroke treatment strategies, is lacking. Movement-related EEG changes (EEG event-related desynchronization (ERD) in [Formula: see text] and [Formula: see text] bands, an indicator of motor cortex activation traditionally used for BCI systems), were evaluated in a group of 23 paralyzed chronic stroke patients in two unilateral motor tasks alternating paretic and healthy hands ((i) passive movement, using a hand exoskeleton, and (ii) voluntary movement), and compared to nine healthy subjects. In tasks using unaffected hand, we observed an increase of contralesional hemisphere activation for stroke patients group. Unexpectedly, when using paralyzed hand, motor cortex activation was reduced or absent in severely affected group of patients, while patients with moderate motor deficit showed an activation greater than control group. Cortical activation was reduced or absent in damaged hemisphere of all the patients in both tasks. Significant differences related to severity of motor deficit were found in the time course of [Formula: see text]-[Formula: see text] bands power ratio in EEG of contralesional hemisphere while moving affected hand. These findings suggest the presence of different compensation mechanisms in contralesional hemisphere of stroke patients related to the grade of motor disability, that might turn quantitative EEG during a movement task, obtained from a BCI system controlling a robotic device included in a rehabilitation task, into a valuable tool for monitoring clinical progression, evaluating recovery, and tailoring treatment of stroke patients.
BackgroundAssistive technologies aim to increase quality of life, reduce dependence on care giver and on the long term care system. Several studies have demonstrated the effectiveness in the use of assistive technology for environment control and communication systems. The progress of brain-computer interfaces (BCI) research together with exoskeleton enable a person with motor impairment to interact with new elements in the environment. This paper aims to evaluate the environment control interface (ECI) developed under the AIDE project conditions, a multimodal interface able to analyze and extract relevant information from the environments as well as from the identification of residual abilities, behaviors, and intentions of the user.MethodsThis study evaluated the ECI in a simulated scenario using a two screen layout: one with the ECI and the other with a simulated home environment, developed for this purpose. The sensorimotor rhythms and the horizontal oculoversion, acquired through BCI2000, a multipurpose standard BCI platform, were used to online control the ECI after the user training and system calibration. Eight subjects with different neurological diseases and spinal cord injury participated in this study. The subjects performed simulated activities of daily living (ADLs), i.e. actions in the simulated environment as drink, switch on a lamp or raise the bed head, during ten minutes in two different modes, AIDE mode, using a prediction model, to recognize the user intention facilitating the scan, and Manual mode, without a prediction model.ResultsThe results show that the mean task time spent in the AIDE mode was less than in the Manual, i.e the users were able to perform more tasks in the AIDE mode during the same time. The results showed a statistically significant differences with p<0.001. Regarding the steps, i.e the number of abstraction levels crossed in the ECI to perform an ADL, the users performed one step in the 90% of the tasks using the AIDE mode and three steps, at least, were necessary in the Manual mode. The user’s intention prediction was performed through conditional random fields (CRF), with a global accuracy about 87%.ConclusionsThe environment analysis and the identification of the user’s behaviors can be used to predict the user intention opening a new paradigm in the design of the ECIs. Although the developed ECI was tested only in a simulated home environment, it can be easily adapted to a real environment increasing the user independence at home.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.