Learning tasks across different environments

27 July 2010

share this:

In the future, robots will be expected to learn a task and execute it in a variety of realistic situations. Reinforcement-learning and planning algorithms are exactly intended for that purpose. However, one of the main challenges is to make sure actions learned in one environment can be used in new and unforeseen situations in real time.

To address this challenge, Stolle et al. have imagined a series of algorithms which they demonstrate on complex tasks such as solving a marble maze or making Boston Dynamic’s Little Dog navigate over complex terrain (see video below).

The first ingredient of success relies on making robots learn what action to take based on local features, meaning features as viewed by the robot (e.g. “there is a wall to the right”). These local features can then be recognized in new environments when the robot is in similar situations. Instead, many existing algorithms use global information, for example by saying “perform this action in position (x,y,z)”. Changing the environment however would typically make these global policies useless.

The second ingredient makes robots build libraries containing sequences of actions (trajectories) that can bring a robot from its current state to an aimed goal. Robots then apply the actions from the trajectory nearest to their state to achieve a task. This strategy is interesting because it is not computationally expensive and does not require large amounts of fast memory.

Finally, don’t miss the following video of little-dog climbing over a fence. This special purpose behavior can be used in a variety of situations.

Sabine Hauert is President of Robohub and Associate Professor at the Bristol Robotics Laboratory
Sabine Hauert is President of Robohub and Associate Professor at the Bristol Robotics Laboratory

Related posts :

Meet the Oystamaran

Working directly with oyster farmers, MIT students are developing a robot that can flip heavy, floating bags of oysters, helping the shellfish to grow and stay healthy.
08 December 2021, by

Exploring ROS2 with a wheeled robot – #4 – Obstacle avoidance

In this post you’ll learn how to program a robot to avoid obstacles using ROS2 and C++. Up to the end of the post, the Dolly robot moves autonomously in a scene with many obstacles, simulated using Gazebo 11.
06 December 2021, by

Team builds first living robots that can reproduce

AI-designed Xenobots reveal entirely new form of biological self-replication—promising for regenerative medicine.
02 December 2021, by

Exploring ROS2 using wheeled Robot – #3 – Moving the robot

In this post you’ll learn how to publish to a ROS2 topic using ROS2 C++. We are moving the robot Dolly robot, simulated using Gazebo 11.
30 November 2021, by

An inventory of robotics roadmaps to better inform policy and investment

Silicon Valley Robotics in partnership with the Industrial Activities Board of the IEEE Robotics and Automation Society, is compiling an up to date resource list of various robotics, AIS and AI roadmaps, national or otherwise.
29 November 2021, by

Robots can be companions, caregivers, collaborators — and social influencers

People are hardwired to respond socially to technology that presents itself as even vaguely social. While this may sound like the beginnings of a Black Mirror episode, this tendency is precisely what allows us to enjoy social interactions with robots and place them in caregiver, collaborator or companion roles.
26 November 2021, by

©2021 - ROBOTS Association


©2021 - ROBOTS Association