Robohub.org
 

Insect-inspired flying robot handles collisions, goes where other robots can’t


by
30 October 2013



share this:
gimball
Gimball is a flying robot that survives collisions. It weighs just 370g for 34cm in diameter. Photo credit: A. Herzog, EPFL.

Generally, flying robots are programmed to avoid obstacles, which is far from easy in cluttered environments. At the Laboratory of Intelligent Systems, we think that flying robots should be able to physically interact with their surroundings. Take insects: they often collide with obstacles and continue flying afterwards. We thus designed GimBall, a flying robot that can collide with objects seamlessly. Thanks to a passively rotating spherical cage, it remains stable even after taking hits from all sides. This approach enables GimBall to fly in the most difficult places without complex sensors.

Presentation and demonstration of the GimBall in various environments.

The concept
The inner frame comprises everything a flying robot needs to stabilize in flight: a coaxial motor, two control surfaces, the battery, an IMU and control electronics. In case of collision, GimBall’s spherical protective frame prevents obstacles from touching the inner frame and can passively rotate thanks to a gimbal system (hence the name). This way, the contact force does not affect the orientation of the inner frame, whose center of mass is also carefully centered. This allows the inner frame to always stay upright, and maintain the robot in stable flight.
Animation_two_robots

A conventional platform will be strongly affected by collisions, because its orientation is disturbed. GimBall keeps the orientation of its inner frame stable thanks to the gimbal system. Note that this is a side-view of a collision in the vertical plane, but the principle remains the same from all angles, since the protective cage rotates around all 3 axes.

The story
We have been designing flying robots capable of interacting with the environment for a while now. Our previous prototype, the AirBurr, was equipped with a protective cage and legs for uprighting.

Presentation of an early prototype of the AirBurr, capable of active self-recovery.

This platform allowed us to realize that autonomous navigation was much simpler once you can afford collisions: advanced sensors are not required, and the obstacles can even be used to guide the robot. This idea of exploiting collisions was pushed further by integrating contact sensors to the structure itself. This way, the robot can be programmed to change direction once it detects a contact. We demonstrated that the robot could fly completely autonomously just based on tiny touch sensors (no other exteroceptive sensors whatsoever), as featured in the video below.

Demonstration of contact-based navigation with a recent version of the AirBurr, and presentation of a perching mechanism.

While the AirBurr worked well in the lab, where the environment is nicely flat and structured, we are now targeting much harder environments with the GimBall. For example, the robot could fly autonomously through a forest with a very simple strategy: fly straight. It made its way through the forest without any obstacle avoidance technique while experiencing multiple collisions.

We think that research on flying robots capable of interacting with the environment will enable new uses for UAVs. Currently, options for flight in cluttered spaces are limited, because obstacle avoidance methods cannot be trusted all the time. A robot that can survive collisions will be useful in cluttered disaster zones to find survivors or to inspect semi-collapsed buildings for example. Thanks to less stringent sensor requirements, these robots are lighter, faster and more robust than conventional platforms. Finally, they are very safe for operation close to people, since they won’t crash if you bump them and the propellers are enclosed in the protective cage.

You can find all relevant publications and more pictures or videos on our project webpage.

Coming to IROS next week? Comme see us at iRex, we’ll be running demos from Wednesday 6 to Sunday 9.

This research was supported by the Swiss National Science Foundation through the National Centre of Competence in Research (NCCR) Robotics.

If you liked this article, you may also be interested in:

See all the latest robotics news on Robohub, or sign up for our weekly newsletter.



tags: , , , , , , , ,


Adrien Briod is co-founder and CTO of Flyability.
Adrien Briod is co-founder and CTO of Flyability.


Subscribe to Robohub newsletter on substack



Related posts :

Robot Talk Episode 149 – Robot safety and security, with Krystal Mattich

  20 Mar 2026
In the latest episode of the Robot Talk podcast, Claire chatted to Krystal Mattich from Brain Corp about trustworthy autonomous robots in public spaces.

A multi-armed robot for assisting with agricultural tasks

  18 Mar 2026
How can a robot safely manipulate branches to reveal hidden flowers while remaining aware of interaction forces and minimizing damage?

Graphene-based sensor to improve robot touch

  16 Mar 2026
Multiscale-structured miniaturized 3D force sensors for improved robot touch.

Robot Talk Episode 148 – Ethical robot behaviour, with Alan Winfield

  13 Mar 2026
In the latest episode of the Robot Talk podcast, Claire chatted to Alan Winfield from the University of the West of England about developing new standards for ethics and transparency in robotics.

Coding for underwater robotics

  12 Mar 2026
Lincoln Laboratory intern Ivy Mahncke developed and tested algorithms to help human divers and robots navigate underwater.

Restoring surgeons’ sense of touch with robotic fingertips

  10 Mar 2026
Researchers are developing robotic “fingertips” that could give surgeons back their sense of touch during minimally invasive and robotic operations.

Robot Talk Episode 147 – Miniature living robots, with Maria Guix

  06 Mar 2026
In the latest episode of the Robot Talk podcast, Claire chatted to Maria Guix from the University of Barcelona about combining electronics and biology to create biohybrid robots with emergent properties.

Developing an optical tactile sensor for tracking head motion during radiotherapy: an interview with Bhoomika Gandhi

  05 Mar 2026
Bhoomika Gandhi discusses her work on an optical sensor for medical robotics applications.



Robohub is supported by:


Subscribe to Robohub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence