Robohub.org
 

Social learning

by
29 August 2010



share this:

Robots are portrayed as tomorrows helpers, be it in schools, hospitals, workplaces or homes. Unfortunately, such robots won’t be truly useful out-of-the-box because of the complexity of real-world environments and tasks. Instead, they will need to learn how to interact with objects in their environment to produce a desired outcome (affordance learning).

For this purpose, robots can explore the world while using machine learning techniques to update their knowledge. However, the learning process is sometimes saturated with examples of objects, actions and effects that won’t help the robot in its purpose.

In these cases, humans or other social partners can help direct robot learning (social learning). Most studies have focussed on scenarios where a teacher demonstrates how to correctly do a task. The robot then imitates the teacher by reproducing the same actions to achieve the same goals.

This approach, while being very efficient, typically means that the teacher needs to take time to train the robot, which can be burdensome. Furthermore, the robot might be so specialized for the demonstrated scenario that it will have trouble performing tasks that slightly differ. In addition, imitation only works when the teacher and robot have similar motion constraints and morphologies.

Luckily, humans and animals use a large variety of mechanisms to learn from social partners. Tapping into this reservoir, Cakmak et al. propose mechanisms where:
– robots interact with the same objects as the social partner (stimulus enhancement)
– robots try to achieve the same effect on the same object as the social partner (emulation)
– robots reproduce the same action as the social partner (mimicking)

Experiments performed in simulation compare stimulus enhancement, emulation, mimicking, imitation and non-social learning in a large variety of situations. The results summarize which mechanisms are better suited for which scenarios in a series of very useful guidelines. Demonstrations with two robots, Jimmy and Jane, were done to validate the study. Don’t miss the excellent video below for a summary of the article.

In the future, Cakmak et al. will focus on combining learning approaches to harness the full potential of this rich set of mechanisms.



tags:


Sabine Hauert is President of Robohub and Associate Professor at the Bristol Robotics Laboratory
Sabine Hauert is President of Robohub and Associate Professor at the Bristol Robotics Laboratory





Related posts :



Robot Talk Episode 64 – Rav Chunilal

In the latest episode of the Robot Talk podcast, Claire chatted to Rav Chunilal from Sellafield all about robotics and AI for nuclear decommissioning.
31 December 2023, by

AI holidays 2023

Thanks to those that sent and suggested AI and robotics-themed holiday videos, images, and stories. Here’s a sample to get you into the spirit this season....
31 December 2023, by and

Faced with dwindling bee colonies, scientists are arming queens with robots and smart hives

By Farshad Arvin, Martin Stefanec, and Tomas Krajnik Be it the news or the dwindling number of creatures hitting your windscreens, it will not have evaded you that the insect world in bad shape. ...
31 December 2023, by

Robot Talk Episode 63 – Ayse Kucukyilmaz

In the latest episode of the Robot Talk podcast, Claire chatted to Ayse Kucukyilmaz from the University of Nottingham about collaboration, conflict and failure in human-robot interactions.
31 December 2023, by

Interview with Dautzenberg Roman: #IROS2023 Best Paper Award on Mobile Manipulation sponsored by OMRON Sinic X Corp.

The award-winning author describe their work on an aerial robot which can exert large forces onto walls.
19 November 2023, by

Robot Talk Episode 62 – Jorvon Moss

In the latest episode of the Robot Talk podcast, Claire chatted to Jorvon (Odd-Jayy) Moss from Digikey about making robots at home, and robot design and aesthetics.
17 November 2023, by





©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association