Publication | Open Access
What speakers do and what addressees look at
140
Citations
63
References
2006
Year
MusicFixation TargetPerceptionCommunicationAttentionSocial SciencesSpeech RecognitionVisual CognitionSpeaker DiarizationGesture ProcessingGesture StudiesCognitive ScienceGestures Draw FixationsEye ContactSpeech CommunicationEye TrackingHuman InteractionSpeech ProcessingHuman-computer InteractionSpeaker RecognitionArtsAffect PerceptionAudio InterfaceDisplay SizeNonverbal Communication
The study examines whether listeners visually attend to speakers’ gestures and how social context and display size influence this attention. The authors compare a live face‑to‑face interaction with two video conditions. Listeners fixated the speaker’s face more than gestures, with only a minority of gestures drawing gaze; gesture holds and speakers’ self‑gaze reliably attracted attention across all conditions, while speaker‑fixated gestures received less attention in video settings, indicating a social effect on gaze‑following, and the study validates a video‑based paradigm for future cross‑modal research.
This study investigates whether addressees visually attend to speakers’ gestures in interaction and whether attention is modulated by changes in social setting and display size. We compare a live face-to-face setting to two video conditions. In all conditions, the face dominates as a fixation target and only a minority of gestures draw fixations. The social and size parameters affect gaze mainly when combined and in the opposite direction from the predicted with fewer gestures fixated on video than live. Gestural holds and speakers’ gaze at their own gestures reliably attract addressees’ fixations in all conditions. The attraction force of holds is unaffected by changes in social and size parameters, suggesting a bottom-up response, whereas speaker-fixated gestures draw significantly less attention in both video conditions, suggesting a social effect for overt gaze-following and visual joint attention. The study provides and validates a video-based paradigm enabling further experimental but ecologically valid explorations of cross-modal information processing.
| Year | Citations | |
|---|---|---|
Page 1
Page 1