Beer-pouring robot programmed to anticipate human actions (w/ Video)
A robot in Cornell’s Personal Robotics Lab has learned to foresee human action in order to step in and offer a helping hand, or more accurately, roll in and offer a helping claw. (Excerpt from an article in Phys.Org)
Understanding when and where to pour a beer or knowing when to offer assistance opening a refrigerator door can be difficult for a robot because of the many variables it encounters while assessing the situation. A team from Cornell has created a solution.
Gazing intently with a Microsoft Kinect 3-D camera and using a database of 3D videos, the Cornell robot identifies the activities it sees, considers what uses are possible with the objects in the scene and determines how those uses fit with the activities. It then generates a set of possible continuations into the future – such as eating, drinking, cleaning, putting away – and finally chooses the most…
View original post 245 more words