Science & Technology - Posted by Byron Spice-Carnegie Mellon on Friday, December 14, 2012 15:31 - 1 Comment
How to turn robots into social butterflies
CARNEGIE MELLON (US) — New technology could teach robots how to behave at parties and other social settings.
Knowing what others are looking at is important because the answer identifies something of interest, or helps delineate social groupings.
Those insights someday will be essential for robots designed to interact with humans, so researchers at Carnegie Mellon University’s Robotics Institute have developed a method for detecting where people’s gazes intersect.
Straight from the Source
The researchers tested the method using groups of people with head-mounted video cameras. By noting where their gazes converged in three-dimensional space, the researchers could determine if they were listening to a single speaker, interacting as a group, or even following the bouncing ball in a ping-pong game.
The researchers’ algorithm for determining “social saliency” could ultimately be used to evaluate a variety of social cues, such as the expressions on people’s faces or body movements, or data from other types of visual or audio sensors.
“This really is just a first step toward analyzing the social signals of people,” says Hyun Soo Park, a Ph.D. student in mechanical engineering at Carnegie Mellon, who worked on the project with Yaser Sheikh, assistant research professor of robotics, and Eakta Jain of Texas Instruments, who was awarded a Ph.D. in robotics last spring.
“In the future, robots will need to interact organically with people and to do so they must understand their social environment, not just their physical environment.”
Though head-mounted cameras are still unusual, police officers, soldiers, search-and-rescue personnel, and even surgeons have begun to wear body-mounted cameras. Head-mounted systems, such as those integrated into eyeglass frames, are poised to become more common.
Even if person-mounted cameras don’t become ubiquitous, Sheikh notes that these cameras someday might be used routinely by people who work in cooperative teams with robots.
The technique was tested in three real-world settings: a meeting involving two work groups; a musical performance; and a party in which participants played pool and ping-pong and chatted in small groups.
The head-mounted cameras provided precise data about what people were looking at in social settings. The algorithm developed by the research team was able to automatically estimate the number and 3D position of “gaze concurrences”—positions where the gazes of multiple people intersected.
But the researchers were surprised by the level of detail they were able to detect. In the party setting, for instance, the algorithm didn’t just indicate that people were looking at the ping-pong table; the gaze concurrence video actually shows the flight of the ball as it bounces and is batted back and forth.
That finding suggests another possible application for monitoring gaze concurrence: player-level views of ball games. If basketball players all wore head-mounted cameras, for instance, it might be possible to reconstruct the game, not from the point of view of a single player, but from a collective view of the players as they all keep their eyes on the ball.
Another potential use is the study of social behavior, such as group dynamics and gender interactions, and research into behavioral disorders, such as autism.
The researchers reported their findings at the Neural Information Processing Systems Conference in Lake Tahoe, Nevada. The research was sponsored by the Samsung Global Research Outreach Program, Intel, and the National Science Foundation.
Source: Carnegie Mellon University