Manipulating objects is still a major challenge for robots in human-centered environments. To overcome this hurdle, Prats et al. propose to combine vision, force and tactile sensing to achieve robust and reliable manipulation with a robot arm fitted with a 3-finger hand (see video below).
Using three sensing modalities increases the robustness of the system, especially since each sensor taken alone has its shortcomings. For example, vision can be used to track a manipulated object and can therefor be used to control manipulation. However, vision is sometimes badly calibrated or occluded. Furthermore, forces applied to the robot arm can be measured to make sure the efforts are focussed in the right direction. However, if the robot does not have a good grip on the object it is manipulating, this might cause it to slip. Adding tactile sensing instead is useful to feel the object manipulated and readjust the position of the manipulator when errors occur.
To prove their point, Prats et al. test different combinations of all three sensor modalities on a tricky task for robots, opening a sliding door. In the end, it seems that a combination of vision, force and tactile sensing saves the day.