Premium
Signers and Co‐speech Gesturers Adopt Similar Strategies for Portraying Viewpoint in Narratives
Author(s) -
QuintoPozos David,
Parrill Fey
Publication year - 2015
Publication title -
topics in cognitive science
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.191
H-Index - 56
eISSN - 1756-8765
pISSN - 1756-8757
DOI - 10.1111/tops.12120
Subject(s) - narrative , linguistics , indirect speech , sociology , psychology , communication , computer science , cognitive science , philosophy
Gestural viewpoint research suggests that several dimensions determine which perspective a narrator takes, including properties of the event described. Events can evoke gestures from the point of view of a character ( CVPT ), an observer ( OVPT ), or both perspectives. CVPT and OVPT gestures have been compared to constructed action ( CA ) and classifiers ( CL ) in signed languages. We ask how CA and CL , as represented in ASL productions, compare to previous results for CVPT and OVPT from English‐speaking co‐speech gesturers. Ten ASL signers described cartoon stimuli from Parrill (2010). Events shown by Parrill to elicit a particular gestural strategy ( CVPT , OVPT , both) were coded for signers' instances of CA and CL . CA was divided into three categories: CA ‐torso, CA ‐affect, and CA ‐handling. Signers used CA ‐handling the most when gesturers used CVPT exclusively. Additionally, signers used CL the most when gesturers used OVPT exclusively and CL the least when gesturers used CVPT exclusively.