Exoskeleton with haptic sensors helps paralysed man to kick off World Cup

12 June 2014

share this:

VIDEO UPDATE 06/13 It’s June 2014 and all eyes are on Brazil. If you’re a football fan then June 12th is the day you’ve been waiting for, but eagle-eyed technophiles are likely to have noticed one very exciting addition to the opening ceremony.

Given that the World Cup is a celebration of all things football, it seems only right that this event should showcase some emerging technologies that could help a disabled patient to achieve a dream that so many have – taking the first kick at the World Cup.

Researchers work with volunteers to ensure that the exoskeleton is functional and comfortable. Source: Still from NBC Video

The Walk Again Project is a large collaborative project headed by Prof. Miguel Nicolelis of Duke University and featuring researchers in the USA, Brazil, Germany and Switzerland. It aims to give movement back to a paralysed person by means of an exoskeleton using a wide range of technologies.

An EEG cap transmits the intentions of the wearer to the exoskeletonSource: Still from NBC Video.

The exoskeleton is controlled through an EEG (Electroencephalography) cap placed over the head – the sensors on the cap read brain activity from the scalp. The cap then deciphers the intention of the user and transmits the instruction to the exoskeleton, which produces the gait movement using a set of hydraulic actuators. While EEG is less sensitive than implanting sensors on the brain, it has the advantage of being non-invasive, meaning that the teenager was not required to undergo surgery for this one-off event.

Despite being made from alloys and polymers to make it as light and flexible as possible, the exoskeleton weighs in at around 60kg.

For someone who is paraplegic, a major issue when using an exoskeleton is that the person is not able to feel their legs and therefore cannot know what position they are in. Think for a second how hard walking would be if you had absolutely no sensation in your legs – even if your muscles were able to do it, the physical stimuli that we all receive and process without active thought would not be there – so the mental struggle that you would feel – along with the lack of haptic feedback that able-bodied people take for granted – would probably stop you from being able to walk effectively.

Haptic feedback sleeves developed by doctoral students at LSRO, EPFL vibrate against the arm, giving the person wearing the exoskeleton vital information to help them know when their foot hits the ground.

This is where the haptic sleeve system comes in to play. The exoskeleton user wears sleeves created by a team lead by Prof. Hannes Bleuler at LSRO, EPFL and NCCR Robotics and stemming from the ongoing PhD work of Simon Gallo. The sleeves are embedded with vibrators that are spread over the forearm – these vibrators then give a haptic feedback to the wearer as they indicate not only if the foot is on the floor (from sensors on the foot), but which part of the foot is on the floor, giving a proprioceptive effect and allowing the user to walk upright. Thus the wearer can sense what feels like an artificial footprint on their arm as the sensors roll from heel to toe. It is for this part of the technology that the team have spent months training with volunteers, trying out different arrangements of actuators on the arms, trying epsilateral and contralateral arrangements and using virtual reality and an avatar to allow the user to see real time limb positioning.

The sleeves are soft and easy to wear; its sensors create vibrations that correspond to the section of the foot that is on the ground.

One of the major challenges of the project was the training process for the eight volunteers (males and females), to enable them to master control of an exoskeleton directly with their brain in a stressful environment like a stadium. The hopeful volunteers have trained in a realistic virtual environment mimicking the real stadium while receiving feedback from the tactile sleeves for the past 8 months with the final decision on who would take the kick made only last week.

The second worry for the project was that electromagnetic fields from cell phones carried by people in the stadium would block electronic signals in the exoskeleton – however testing in a comparable stadium showed that this would not be a problem.

The researchers hope that by using such a high profile event as the World Cup, the general public will gain a greater understanding and a greater tolerance of what wearable technologies can do, along with giving those with disabilities a hope for the future.

To keep up to date with NCCR Robotics, please sign up to our monthly infoletter and news bulletins, and follow us on Twitter @NCCRRobotics.

If you liked this article, you may also be interested in:

See all the latest robotics news on Robohub, or sign up for our weekly newsletter.

tags: , , , , , , , , ,

NCCR Robotics

Related posts :

Interview with Dautzenberg Roman: #IROS2023 Best Paper Award on Mobile Manipulation sponsored by OMRON Sinic X Corp.

The award-winning author describe their work on an aerial robot which can exert large forces onto walls.
19 November 2023, by

Robot Talk Episode 62 – Jorvon Moss

In the latest episode of the Robot Talk podcast, Claire chatted to Jorvon (Odd-Jayy) Moss from Digikey about making robots at home, and robot design and aesthetics.
17 November 2023, by

California is the robotics capital of the world

In California, robotics technology is a small fish in a much bigger technology pond, and that tends to conceal how important Californian companies are to the robotics revolution.
12 November 2023, by

Robot Talk Episode 61 – Masoumeh Mansouri

In the latest episode of the Robot Talk podcast, Claire chatted to Masoumeh (Iran) Mansouri from the University of Birmingham about culturally sensitive robots and planning in complex environments.
10 November 2023, by

The 5 levels of Sustainable Robotics

Robots can solve the UN SDGs and not just via the application area.
08 November 2023, by

Using language to give robots a better grasp of an open-ended world

By blending 2D images with foundation models to build 3D feature fields, a new MIT method helps robots understand and manipulate nearby objects with open-ended language prompts.
06 November 2023, by

©2021 - ROBOTS Association


©2021 - ROBOTS Association