Publication | Closed Access
Human-robot interaction through real-time auditory and visual multiple-talker tracking
70
Citations
15
References
2002
Year
Unknown Venue
Human-robot Collaborative AssemblyEngineeringIntelligent SystemsCommunicationSpeech RecognitionHumanrobot CollaborationConversation AnalysisCompanion RobotMultimodal Human Computer InterfaceHealth SciencesAssistive TechnologyTalker TrackingHuman Agent InteractionHuman-robot InteractionSpeech CommunicationSpeech TechnologyEye TrackingSpeech ProcessingReal-time AuditorySpeech PerceptionRoboticsSpeech InterfaceVoice Interaction
Nakadai et al. (2001) have developed a real-time auditory and visual multiple-talker tracking technique. In this paper, this technique is applied to human-robot interaction including a receptionist robot and a companion robot at a party. The system includes face identification, speech recognition, focus-of-attention control, and sensorimotor task in tracking multiple talkers. The system is implemented on a upper-torso humanoid and the talker tracking is attained by distributed processing on three nodes connected by 100Base-TX network. The delay of tracking is 200 msec. Focus-of-attention is controlled by associating auditory and visual streams by using the sound source direction and talker position as a clue. Once an association is established, the humanoid keeps its face to the direction of the associated talker.
| Year | Citations | |
|---|---|---|
Page 1
Page 1