Visual speech information for face recognition |
| |
Authors: | Rosenblum Lawrence D Yakel Deborah A Baseer Naser Panchal Anjani Nodarse Brynn C Niehus Ryan P |
| |
Affiliation: | Department of Psychology, University of California, Riverside 92521, USA. rosenblu@citrus.ucr.edu |
| |
Abstract: | Two experiments test whether isolated visible speech movements can be used for face matching. Visible speech information was isolated with a point-light methodology. Participants were asked to match articulating point-light faces to a fully illuminated articulating face in an XAB task. The first experiment tested single-frame static face stimuli as a control. The results revealed that the participants were significantly better at matching the dynamic face stimuli than the static ones. Experiment 2 tested whether the observed dynamic advantage was based on the movement itself or on the fact that the dynamic stimuli consisted of many more static and ordered frames. For this purpose, frame rate was reduced, and the frames were shown in a random order, a correct order with incorrect relative timing, or a correct order with correct relative timing. The results revealed better matching performance with the correctly ordered and timed frame stimuli, suggesting that matches were based on the actual movement itself. These findings suggest that speaker-specific visible articulatory style can provide information for face matching. |
| |
Keywords: | |
本文献已被 PubMed 等数据库收录! |
|