Robohub.org
 

Drone flight through narrow gaps using onboard sensing and computing

detection_fisheye

In this work, we address one of the main challenges towards autonomous drone flight in complex environments, which is flight through narrow gaps. One-day micro drones will be used to search and rescue people in the aftermath of an earthquake. In these situations, collapsed buildings cannot be accessed through conventional windows, so that small gaps may be the only way to get inside. What makes this problem challenging is that a gap can be very small, such that precise trajectory-following is required, and can have arbitrary orientations, such that the quadrotor cannot fly through it in near-hover conditions. This makes it necessary to execute an agile trajectory (i.e., with high velocity and angular accelerations) in order to align the vehicle to the gap orientation.

Previous works on aggressive flight through narrow gaps have focused solely on the control and planning problem and therefore used motion-capture systems for state estimation and external computing. Conversely, we focus on using only onboard sensors and computing. More specifically, we address the case where state estimation is done via gap detection through a single, forward-facing camera and show that this raises an interesting problem of coupled perception and planning: for the robot to localize with respect to the gap, a trajectory has be selected, which guarantees that the quadrotor always faces the gap (perception constraint) and has to be replanned multiple times during its execution to cope with the varying uncertainty of the state estimate. Furthermore, during the traverse, the quadrotor has to maximize the distance from the edges of the gap (geometric constraint) to avoid collisions and, at the same time, it has to be able to do so without relying on any visual feedback (when the robot is very close to the gap, this exits from the camera field of view). Finally, the trajectory has to be feasible with respect to the dynamic constraints of the vehicle. In order to recover and lock into stable hovering after passing the gap, we used the recovery procedure described in our former ICRA’15 paper and also described in a former Robohub article.

Our proposed trajectory generation approach is independent of the gap-detection algorithm being used; thus, to simplify the perception task, we used a gap with a simple black-and-white rectangular pattern. One technical aspect to point out is that, in order to allow the quadrotor actuators to quickly change the vehicle orientation that would allow always facing the window during the approach maneuver, the propellers had to be tilted by 15 degrees. Tilting the propellers provided three times more yaw-control action, while only losing 3% of the collective thrust.

quad_passing_clean

We successfully evaluated our approach with gap orientations of up to 45 degrees vertically and up to 30 horizontally. Our vehicle weighs 830 grams and has a thrust-to-weight ratio of 2.5. Our trajectory generation formulation handles trajectories up to 90-degree gap orientations although the quadrotor used in these experiments is too heavy and the motors saturate for more than 45-degree gap orientations. The vehicle reaches speeds of up to 3 meters per second and angular velocities of up to 400 degrees per second, with accelerations of up to 1.5 g. We can pass through gaps 1.5 times the size of the quadrotor, with only 10 centimeters of tolerance. Our method does not require any prior knowledge about the position and the orientation of the gap. No external infrastructure, such as a motion-capture system, is needed. This is the first time that such an aggressive maneuver through narrow gaps has been done by fusing gap detection from a single onboard camera and IMU.

Passing through narrow gaps is even challenging for human pilots! We invited two Swiss professional FPV drone-racing pilots to come to our lab and demo flight through the same gap we used in our experiments, using FPV glasses. It was not easy at all and they only managed after several attempts:

This work has been submitted to ICRA 2017: Aggressive Quadrotor Flight through Narrow Gaps with Onboard Sensing and Computing, by Davide Falanga, Elias Mueggler, Matthias Faessler, and Davide Scaramuzza.



tags: , ,


Davide Falanga is a PhD student at Prof. Scaramuzza's Robotics and Perception Group of the University of Zurich.
Davide Falanga is a PhD student at Prof. Scaramuzza's Robotics and Perception Group of the University of Zurich.

Elias Müggler is a PhD student at Prof. Scaramuzza's Robotics and Perception Group of the University of Zurich.
Elias Müggler is a PhD student at Prof. Scaramuzza's Robotics and Perception Group of the University of Zurich.

Davide Scaramuzza is Assistant Professor of Robotics at the University of Zurich.
Davide Scaramuzza is Assistant Professor of Robotics at the University of Zurich.

Matthias Fässler is a PhD student at Prof. Scaramuzza's Robotics and Perception Group of the University of Zurich.
Matthias Fässler is a PhD student at Prof. Scaramuzza's Robotics and Perception Group of the University of Zurich.





Related posts :



Robot Talk Episode 110 – Designing ethical robots, with Catherine Menon

  21 Feb 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Catherine Menon from the University of Hertfordshire about designing home assistance robots with ethics in mind.

Robot Talk Episode 109 – Building robots at home, with Dan Nicholson

  14 Feb 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Dan Nicholson from MakerForge.tech about creating open source robotics projects you can do at home.

Robot Talk Episode 108 – Giving robots the sense of touch, with Anuradha Ranasinghe

  07 Feb 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Anuradha Ranasinghe from Liverpool Hope University about haptic sensors for wearable tech and robotics.

Robot Talk Episode 107 – Animal-inspired robot movement, with Robert Siddall

  31 Jan 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Robert Siddall from the University of Surrey about novel robot designs inspired by the way real animals move.

Robot Talk Episode 106 – The future of intelligent systems, with Didem Gurdur Broo

  24 Jan 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Didem Gurdur Broo from Uppsala University about how to shape the future of robotics, autonomous vehicles, and industrial automation.

Robot Talk Episode 105 – Working with robots in industry, with Gianmarco Pisanelli 

  17 Jan 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Gianmarco Pisanelli from the Advanced Manufacturing Research Centre about how to promote the safe and intuitive use of robots in manufacturing.

Robot Talk Episode 104 – Robot swarms inspired by nature, with Kirstin Petersen

  10 Jan 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Kirstin Petersen from Cornell University about how robots can work together to achieve complex behaviours.

Robot Talk Episode 103 – Delivering medicine by drone, with Keenan Wyrobek

  20 Dec 2024
In the latest episode of the Robot Talk podcast, Claire chatted to Keenan Wyrobek from Zipline about drones for delivering life-saving medicine to remote locations.





Robohub is supported by:




Would you like to learn how to tell impactful stories about your robot or AI system?


scicomm
training the next generation of science communicators in robotics & AI


©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association