Skip to main content
eScholarship
Open Access Publications from the University of California

Semantic and Visual Features Drive the Intrinsic Memorability of Co-Speech Gestures

Abstract

Co-speech gestures that teachers spontaneously produce during explanations benefit students' learning by enhancing memory (Church et al., 2007). However, it remains unclear whether certain gestures are intrinsically more memorable, and if so, owing to what semantic and visual features. We created 360 10-second audiovisual stimuli by recording 20 actors producing natural, unscripted explanations of Piagetian conservation problems. For each audiovisual stimulus, two trained experimenters extracted high-level semantic and low-level visual/acoustic features in speech and gesture. We then tested online participants' memories using a between-subjects study-test paradigm in three different conditions: audiovisual (gesture+speech stimuli), visual-only (gesture-only version of the same stimuli), and audio-only (speech-only version of the same stimuli). We found that participants consistently remembered certain gesture, gesture+speech, and speech stimuli better than others. Focusing on the visual-only (gesture-only) condition, we discovered that both semantic (speech and gesture meaningfulness) and visual (number of hands used) features make co-speech gestures memorable.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View