首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
The visual system has been suggested to integrate different views of an object in motion. We investigated differences in the way moving and static objects are represented by testing for priming effects to previously seen ("known") and novel object views. We showed priming effects for moving objects across image changes (e.g., mirror reversals, changes in size, and changes in polarity) but not over temporal delays. The opposite pattern of results was observed for objects presented statically; that is, static objects were primed over temporal delays but not across image changes. These results suggest that representations for moving objects are: (1) updated continuously across image changes, whereas static object representations generalize only across similar images, and (2) more short-lived than static object representations. These results suggest two distinct representational mechanisms: a static object mechanism rather spatially refined and permanent, possibly suited for visual recognition, and a motion-based object mechanism more temporary and less spatially refined, possibly suited for visual guidance of motor actions.  相似文献   

2.
Harman KL  Humphrey GK 《Perception》1999,28(5):601-615
When we look at an object as we move or the object moves, our visual system is presented with a sequence of different views of the object. It has been suggested that such regular temporal sequences of views of objects contain information that can aid in the process of representing and recognising objects. We examined whether seeing a series of perspective views of objects in sequence led to more efficient recognition than seeing the same views of objects but presented in a random order. Participants studied images of 20 novel three-dimensional objects rotating in depth under one of two study conditions. In one study condition, participants viewed an ordered sequence of views of objects that was assumed to mimic important aspects of how we normally encounter objects. In the other study condition, participants were presented the same object views, but in a random order. It was expected that studying a regular sequence of views would lead to more efficient recognition than studying a random presentation of object views. Although subsequent recognition accuracy was equal for the two groups, differences in reaction time between the two study groups resulted. Specifically, the random study group responded reliably faster than the sequence study group. Some possible encoding differences between the two groups are discussed.  相似文献   

3.
Liu T 《Perception》2007,36(9):1320-1333
How humans recognize objects remains a contentious issue in current research on high-level vision. Here, I test the proposal by Wallis and Bülthoff (1999 Trends in Cognitive Sciences 3 22-31) suggesting that object representations can be learned through temporal association of multiple views of the same object. Participants first studied image sequences of novel, three-dimensional objects in a study block. On each trial, the images were from either an orderly sequence of depth-rotated views of the same object (SS), a scrambled sequence of those views (SR), or a sequence of different objects (RR). Recognition memory was assessed in a following test block. A within-object advantage was consistently observed --greater accuracy in the SR than the RR condition in all four experiments, greater accuracy in the SS than the RR condition in two experiments. Furthermore, spatiotemporal coherence did not produce better recognition than temporal coherence alone (similar or less accuracy in the SS compared to the SR condition). These results suggest that the visual system can use temporal regularity to build invariant object representations, via the temporal-association mechanism.  相似文献   

4.
Learning to recognize objects appears to depend critically on extended observation of appearance over time. Specifically, temporal association between dissimilar views of an object has been proposed as a tool for learning invariant representations for recognition. We examined heretofore untested aspects of the temporal association hypothesis using a familiar dynamic object, the human body. Specifically, we examined the role of appearance prediction (temporal asymmetry) in temporal association. In our task, observers performed a change detection task using upright and inverted images of a walking body either with or without previous exposure to a motion stimulus depicting an upright walker. Observers who were exposed to the dynamic stimulus were further divided into two groups dependent on whether the observed motion depicted forward or backward walking. We find that the effect of the motion stimulus on sensitivity is highly dependent on whether the observed motion is consistent with past experience.  相似文献   

5.
Objects are best recognized from so-called “canonical” views. The characteristics of canonical views of arbitrary objects have been qualitatively described using a variety of different criteria, but little is known regarding how these views might be acquired during object learning. We address this issue, in part, by examining the role of object motion in the selection of preferred views of novel objects. Specifically, we adopt a modeling approach to investigate whether or not the sequence of views seen during initial exposure to an object contributes to observers’ preferences for particular images in the sequence. In two experiments, we exposed observers to short sequences depicting rigidly rotating novel objects and subsequently collected subjective ratings of view canonicality (Experiment 1) and recall rates for individual views (Experiment 2). Given these two operational definitions of view canonicality, we attempted to fit both sets of behavioral data with a computational model incorporating 3-D shape information (object foreshortening), as well as information relevant to the temporal order of views presented during training (the rate of change for object foreshortening). Both sets of ratings were reasonably well predicted using only 3-D shape; the inclusion of terms that capture sequence order improved model performance significantly.  相似文献   

6.
Natural variability between instances of unfamiliar faces can make it difficult to reconcile two images as the same person. Yet for familiar faces, effortless recognition occurs even with considerable variability between images. To explore how stable face representations develop, we employed incidental learning in the form of a face sorting task. In each trial, multiple images of two facial identities were sorted into two corresponding piles. Following the sort, participants showed evidence of having learnt the faces performing more accurately on a matching task with seen than with unseen identities. Furthermore, ventral temporal event-related potentials were more negative in the N250 time range for previously seen than for previously unseen identities. These effects appear to demonstrate some degree of abstraction, rather than simple picture learning, as the neurophysiological and behavioural effects were observed with novel images of the previously seen identities. The results provide evidence of the development of facial representations, allowing a window onto natural mechanisms of face learning.  相似文献   

7.
Z Kourtzi  M Shiffrar 《Acta psychologica》1999,102(2-3):265-292
Depth rotations can reveal new object parts and result in poor recognition of "static" objects (Biederman & Gerhardstein, 1993). Recent studies have suggested that multiple object views can be associated through temporal contiguity and similarity (Edelman & Weinshall, 1991; Lawson, Humphreys & Watson, 1994; Wallis, 1996). Motion may also play an important role in object recognition since observers recognize novel views of objects rotating in the picture plane more readily than novel views of statically re-oriented objects (Kourtzi & Shiffrar, 1997). The series of experiments presented here investigated how different views of a depth-rotated object might be linked together even when these views do not share the same parts. The results suggest that depth rotated object views can be linked more readily with motion than with temporal sequence alone to yield priming of novel views of 3D objects that fall in between "known" views. Motion can also enhance path specific view linkage when visible object parts differ across views. Such results suggest that object representations depend on motion processes.  相似文献   

8.
We used fMRI to directly compare activation in two cortical regions previously identified as relevant to real-world scene processing: retrosplenial cortex and a region of posterior parahippocampal cortex functionally defined as the parahippocampal place area (PPA). We compared activation in these regions to full views of scenes from a global perspective, close-up views of sub-regions from the same scene category, and single objects highly diagnostic of that scene category. Faces were included as a control condition. Activation in parahippocampal place area was greatest for full scene views that explicitly included the 3D spatial structure of the environment, with progressively less activation for close-up views of local scene regions containing diagnostic objects but less explicitly depicting 3D scene geometry, followed by single scene-diagnostic objects. Faces did not activate parahippocampal place area. In contrast, activation in retrosplenial cortex was greatest for full scene views, and did not differ among close-up views, diagnostic objects, and faces. The results showed that parahippocampal place area responds in a graded fashion as images become more completely scene-like and include more explicit 3D structure, whereas retrosplenial cortex responds in a step-wise manner to the presence of a complete scene. These results suggest scene processing areas are particularly sensitive to the 3D geometric structure that distinguishes scenes from other types of complex and meaningful visual stimuli.  相似文献   

9.
When novel and familiar faces are viewed simultaneously, humans and monkeys show a preference for looking at the novel face. The facial features attended to in familiar and novel faces, were determined by analyzing the visual exploration patterns, or scanpaths, of four monkeys performing a visual paired comparison task. In this task, the viewer was first familiarized with an image and then it was presented simultaneously with a novel and the familiar image. A looking preference for the novel image indicated that the viewer recognized the familiar image and hence differentiates between the familiar and the novel images. Scanpaths and relative looking preference were compared for four types of images: (1) familiar and novel objects, (2) familiar and novel monkey faces with neutral expressions, (3) familiar and novel inverted monkey faces, and (4) faces from the same monkey with different facial expressions. Looking time was significantly longer for the novel face, whether it was neutral, expressing an emotion, or inverted. Monkeys did not show a preference, or an aversion, for looking at aggressive or affiliative facial expressions. The analysis of scanpaths indicated that the eyes were the most explored facial feature in all faces. When faces expressed emotions such as a fear grimace, then monkeys scanned features of the face, which contributed to the uniqueness of the expression. Inverted facial images were scanned similarly to upright images. Precise measurement of eye movements during the visual paired comparison task, allowed a novel and more quantitative assessment of the perceptual processes involved the spontaneous visual exploration of faces and facial expressions. These studies indicate that non-human primates carry out the visual analysis of complex images such as faces in a characteristic and quantifiable manner.  相似文献   

10.
The influence of motion information and temporal associations on recognition of non-familiar faces was investigated using two groups which performed a face recognition task. One group was presented with regular temporal sequences of face views designed to produce the impression of motion of the face rotating in depth, the other group with random sequences of the same views. In one condition, participants viewed the sequences of the views in rapid succession with a negligible interstimulus interval (ISI). This condition was characterized by three different presentation times. In another condition, participants were presented a sequence with a 1-sec. ISI among the views. That regular sequences of views with a negligible ISI and a shorter presentation time were hypothesized to give rise to better recognition, related to a stronger impression of face rotation. Analysis of data from 45 participants showed a shorter presentation time was associated with significantly better accuracy on the recognition task; however, differences between performances associated with regular and random sequences were not significant.  相似文献   

11.
Viewpoint-dependent recognition of familiar faces   总被引:5,自引:0,他引:5  
Troje NF  Kersten D 《Perception》1999,28(4):483-487
The question whether object representations in the human brain are object-centered or viewer-centered has motivated a variety of experiments with divergent results. A key issue concerns the visual recognition of objects seen from novel views. If recognition performance depends on whether a particular view has been seen before, it can be interpreted as evidence for a viewer-centered representation. Earlier experiments used unfamiliar objects to provide the experimenter with complete control over the observer's previous experience with the object. In this study, we tested whether human recognition shows viewpoint dependence for the highly familiar faces of well-known colleagues and for the observer's own face. We found that observers are poorer at recognizing their own profile, whereas there is no difference in response time between frontal and profile views of other faces. This result shows that extensive experience and familiarity with one's own face is not sufficient to produce viewpoint invariance. Our result provides strong evidence for viewer-centered representations in human visual recognition even for highly familiar objects.  相似文献   

12.
Despite intense research on the visual communication of domestic dogs, their cognitive capacities have not yet been explored by eye tracking. The aim of the current study was to expand knowledge on the visual cognition of dogs using contact-free eye movement tracking under conditions where social cueing and associative learning were ruled out. We examined whether dogs spontaneously look at actual objects within pictures and can differentiate between pictures according to their novelty or categorical information content. Eye movements of six domestic dogs were tracked during presentation of digital color images of human faces, dog faces, toys, and alphabetic characters. We found that dogs focused their attention on the informative regions of the images without any task-specific pre-training and their gazing behavior depended on the image category. Dogs preferred the facial images of conspecifics over other categories and fixated on a familiar image longer than on novel stimuli regardless of the category. Dogs’ attraction to conspecifics over human faces and inanimate objects might reflect their natural interest, but further studies are needed to establish whether dogs possess picture object recognition. Contact-free eye movement tracking is a promising method for the broader exploration of processes underlying special socio-cognitive skills in dogs previously found in behavioral studies.  相似文献   

13.
Arcimboldo images induce the perception of faces when shown upright despite the fact that only nonfacial objects such as vegetables and fruits are painted. In the current study, we examined whether infants recognize a face in the Arcimboldo images by using the preferential looking technique and near-infrared spectroscopy (NIRS). In the first experiment, we measured looking preference between upright and inverted Arcimboldo images among 5- and 6-month-olds and 7- and 8-month-olds. We hypothesized that if infants perceive the Arcimboldo images as faces, they would prefer the upright images to the inverted ones. We found that only 7- and 8-month-olds significantly preferred upright images, suggesting that they could perceive the Arcimboldo images as faces. In the second experiment, we measured hemodynamic responses using NIRS. Based on the behavioral data, we hypothesized that 7- and 8-month-olds would show different neural activity for upright and inverted Arcimboldo images, as do adults. Therefore, we measured hemodynamic responses in 7- and 8-month-olds while they were looking at upright and inverted Arcimboldo images. Their responses were then compared with the baseline activation during the presentation of individual vegetables. We found that the concentration of oxyhemoglobin increased in the left temporal area during the presentation of the upright images compared with the baseline during the presentation of vegetables. The results of the two experiments suggest that (a) the ability to recognize the upright Arcimboldo images as faces develops at around 7 or 8 months of age and (b) processing of the upright Arcimboldo images is related to the left temporal area of the brain.  相似文献   

14.
ABSTRACT— Representations of individual faces evolve with experience to support progressively more robust recognition. Knowledge of three-dimensional face structure is required to predict an image of a face as illumination and viewpoint change. Robust recognition across such transformations can be achieved with representations based on multiple two-dimensional views, three-dimensional structure, or both. We used face-identity adaptation in a familiarization paradigm to address a long-standing controversy concerning the role of two-dimensional versus three-dimensional information in face representations. We reasoned that if three-dimensional information is coded in the representations of familiar faces, then learning a new face using images generated by one three-dimensional transformation should enhance the robustness of the representation to another type of three-dimensional transformation. Familiarization with multiple views of faces enhanced the transfer of face-identity adaptation effects across changes in illumination by compensating for a generalization cost at a novel test viewpoint. This finding demonstrates a role for three-dimensional information in representations of familiar faces.  相似文献   

15.
When faces are learned from rotating view sequences, novel views may be recognized by matching them with an integrated representation of the sequence or with individual views. An integrated-representation process should benefit from short view durations, and thus from the inclusion of views in a short temporal window, allowing the distribution of attention over the entire sequence. A view-matching process should benefit from long view durations, allowing the attention to focus on each view. In a sequential comparison task, we tested the recognition of learned and novel interpolated and extrapolated views after learning faces from rapid and slow sequences (240 ms or 960 ms for each view). We found a superiority of rapid over slow sequences, in favour of the integrated-representation hypothesis. In addition, the recognition pattern for the different viewpoints in the sequence depended on the absence or presence of extrapolated views, showing a bias of the distribution of attention.  相似文献   

16.
The ability to recognize identity despite within-person variability in appearance is likely a face-specific skill and shaped by experience. Ensemble coding – the automatic extraction of the average of a stimulus array – has been proposed as a mechanism underlying face learning (allowing one to recognize novel instances of a newly learned face). We investigated whether ensemble encoding, like face learning and recognition, is refined by experience by testing participants with upright own-race faces and two categories of faces with which they lacked experience: other-race faces (Experiment 1) and inverted faces (Experiment 2). Participants viewed four images of an unfamiliar identity and then were asked whether a test image of that same identity had been in the study array. Each test image was a matching exemplar (from the array), matching average (the average of the images in the array), non-matching exemplar (a novel image of the same identity), or non-matching average (an average of four different images of the same identity). Adults showed comparable ensemble coding for all three categories (i.e., reported that matching averages had been present more than non-matching averages), providing evidence that this early stage of face learning is not shaped by face-specific experience.  相似文献   

17.
Implicit memory refers to nonconscious retrieval of past experience demonstrated by facilitation in test performance on tasks that do not require intentional recollection of previous experiences. Explicit memory, in contrast, refers to the conscious retrieval of prior information, as demonstrated during standard recall and recognition tasks. In this experiment, positron emission tomographic (PET) measurements of regional cerebral blood flow (CBF), a marker of local neuronal activity, were used to identify and contrast brain regions that participate in the perception, implicit memory, and explicit memory for structurally possible and impossible visual objects. Ten CBF images were acquired in 16 normal women as they made possible/impossible and old/new recognition decisions about previously studied (old) and nonstudied (new) structurally possible and impossible objects. As reported previously, object decisions for familiar possible objects were associated with increased CBF in the vicinity of the left inferior temporal and fusiform gyri and recognition memory for familiar possible objects was associated with increased CBF in the vicinity of the right hippocampus. In this report, we provide more extensive analyses of the roles of the inferior temporal cortex, the hippocampus, the parahippocampus, and the pulvinar in encoding and retrieval operations. Additionally, patterns of CBF increases and decreases provide information regarding the neural structures involved in implicit and explicit memory.  相似文献   

18.
Recent research suggests that repetition priming (RP) for unfamiliar faces is highly view dependent and is eliminated when the viewpoint of target faces changes between study and test. The current research examined whether increased familiarity with novel faces from a single viewpoint at study would support RP from an alternative viewpoint at test. Participants passively viewed novel face images from a single viewpoint at study (i.e., either front or three-quarters), with half of the images seen once and half seen on five occasions. During a sex classification task at test, participants were faster to respond to face images seen from the same view as that at study than they were to previously unseen distractor faces for both single exposure faces and faces seen on five occasions (i.e., standard RP). When, however, face images at test were shown from a different viewpoint from that at study, RP only occurred for faces viewed on five occasions.  相似文献   

19.
Recent research suggests that repetition priming (RP) for unfamiliar faces is highly view dependent and is eliminated when the viewpoint of target faces changes between study and test. The current research examined whether increased familiarity with novel faces from a single viewpoint at study would support RP from an alternative viewpoint at test. Participants passively viewed novel face images from a single viewpoint at study (i.e., either front or three-quarters), with half of the images seen once and half seen on five occasions. During a sex classification task at test, participants were faster to respond to face images seen from the same view as that at study than they were to previously unseen distractor faces for both single exposure faces and faces seen on five occasions (i.e., standard RP). When, however, face images at test were shown from a different viewpoint from that at study, RP only occurred for faces viewed on five occasions.  相似文献   

20.
Evidence from neurophysiological and psychological studies is coming together to shed light on how we represent and recognize objects. This review describes evidence supporting two major hypotheses: the first is that objects are represented in a mosaic-like form in which objects are encoded by combinations of complex, reusable features, rather than two-dimensional templates, or three-dimensional models. The second hypothesis is that transform-invariant representations of objects are learnt through experience, and that this learning is affected by the temporal sequence in which different views of the objects are seen, as well as by their physical appearance.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号