首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 93 毫秒
1.
Humans and macaques are more sensitive to differences in nonaccidental image properties, such as straight vs. curved contours, than to differences in metric properties, such as degree of curvature [Biederman, I., Bar, M., 1999. One-shot viewpoint invariance in matching novel objects. Vis. Res. 39, 2885-2899; Kayaert, G., Biederman, I., Vogels, R., 2003. Shape tuning in macaque inferior temporal cortex. J. Neurosci. 23, 3016-3027; Kayaert, G., Biederman, I., Vogels, R., 2005. Representation of regular and irregular shapes in macaque inferotemporal cortex. Cereb. Cortex 15, 1308-1321]. This differential sensitivity allows facile recognition when the object is viewed at an orientation in depth not previously experienced. In Experiment 1, we trained pigeons to discriminate grayscale, shaded images of four shapes. Pigeons made more confusion errors to shapes that shared more nonaccidental properties. Although the images in that experiment were not well controlled for incidental changes in metric properties, the same results were apparent with better controlled stimuli in Experiment 2: pigeons trained to discriminate a target shape from a metrically changed shape and a nonaccidentally changed shape committed more confusion errors to the metrically changed shape, suggesting that they perceived it to be more similar to the target shape. Humans trained with similar stimuli and procedure exhibited the same tendency to make more errors to the metrically changed shape. These results document the greater saliency of nonaccidental differences for shape recognition and discrimination in a non-primate species and suggest that nonaccidental sensitivity may be characteristic of all shape-discriminating species.  相似文献   

2.

Background

How do people sustain a visual representation of the environment? Currently, many researchers argue that a single visual working memory system sustains non-spatial object information such as colors and shapes. However, previous studies tested visual working memory for two-dimensional objects only. In consequence, the nature of visual working memory for three-dimensional (3D) object representation remains unknown.

Methodology/Principal Findings

Here, I show that when sustaining information about 3D objects, visual working memory clearly divides into two separate, specialized memory systems, rather than one system, as was previously thought. One memory system gradually accumulates sensory information, forming an increasingly precise view-dependent representation of the scene over the course of several seconds. A second memory system sustains view-invariant representations of 3D objects. The view-dependent memory system has a storage capacity of 3–4 representations and the view-invariant memory system has a storage capacity of 1–2 representations. These systems can operate independently from one another and do not compete for working memory storage resources.

Conclusions/Significance

These results provide evidence that visual working memory sustains object information in two separate, specialized memory systems. One memory system sustains view-dependent representations of the scene, akin to the view-specific representations that guide place recognition during navigation in humans, rodents and insects. The second memory system sustains view-invariant representations of 3D objects, akin to the object-based representations that underlie object cognition.  相似文献   

3.
Can nonhuman animals attend to visual stimuli as whole, coherent objects? We investigated this question by adapting for use with pigeons a task in which human participants must report whether two visual attributes belong to the same object (one-object trial) or to different objects (two-object trial). We trained pigeons to discriminate a pair of differently colored shapes that had two targets either on a single object or on two different objects. Each target equally often appeared on the one-object and two-object stimuli; therefore, a specific target location could not serve as a discriminative cue. The pigeons learned to report whether the two target dots were located on a single object or on two different objects; follow-up tests demonstrated that this ability was not entirely based on memorization of the dot patterns and locations. Additional tests disclosed predominate stimulus control by the color, but not by the shape of the two objects. These findings suggest that human psychophysical methods are readily applicable to the study of object discrimination by nonhuman animals.  相似文献   

4.
Humans can recognize an object within a fraction of a second, even if there are no clues about what kind of object it might be. Recent findings have identified functional properties of extrastriate regions in the ventral visual pathway that are involved in the representation and perception of objects and faces. The functional properties of these regions, and the correlation between the activation of these regions and visual recognition, indicate that the lateral and ventral occipito-temporal areas are important in perceiving and recognizing objects and faces.  相似文献   

5.
Although pictures are widely used as stimuli in cognitive experiments with both humans and animals, the question of how subjects interpret pictures receives less attention. Gaining a better understanding of this is especially important when working with avian subjects, as their visual anatomy and processing is different from that of humans, and even differs from one avian species to another. Successful testing for picture recognition in birds has been carried out mainly with pigeons, but no such research has been explicitly performed with ‘brainy’ birds like parrots, despite the fact that these have been the subject of exciting cognitive research. This study tested kea (Nestor notabilis) mountain parrots for picture–object recognition using a procedure which required the transfer of a learned discrimination task between pictures and objects. Kea successfully showed both picture‐to‐object and object‐to‐picture transfer and performed at a comparable level when pictures were displayed on a touch screen or as printed photographs.  相似文献   

6.
The processes underlying object recognition are fundamental for the understanding of visual perception. Humans can recognize many objects rapidly even in complex scenes, a task that still presents major challenges for computer vision systems. A common experimental demonstration of this ability is the rapid animal detection protocol, where human participants earliest responses to report the presence/absence of animals in natural scenes are observed at 250–270 ms latencies. One of the hypotheses to account for such speed is that people would not actually recognize an animal per se, but rather base their decision on global scene statistics. These global statistics (also referred to as spatial envelope or gist) have been shown to be computationally easy to process and could thus be used as a proxy for coarse object recognition. Here, using a saccadic choice task, which allows us to investigate a previously inaccessible temporal window of visual processing, we showed that animal – but not vehicle – detection clearly precedes scene categorization. This asynchrony is in addition validated by a late contextual modulation of animal detection, starting simultaneously with the availability of scene category. Interestingly, the advantage for animal over scene categorization is in opposition to the results of simulations using standard computational models. Taken together, these results challenge the idea that rapid animal detection might be based on early access of global scene statistics, and rather suggests a process based on the extraction of specific local complex features that might be hardwired in the visual system.  相似文献   

7.
A variety of similarities between visual and haptic object recognition suggests that the two modalities may share common representations. However, it is unclear whether such common representations preserve low-level perceptual features or whether transfer between vision and haptics is mediated by high-level, abstract representations. Two experiments used a sequential shape-matching task to examine the effects of size changes on unimodal and crossmodal visual and haptic object recognition. Participants felt or saw 3D plastic models of familiar objects. The two objects presented on a trial were either the same size or different sizes and were the same shape or different but similar shapes. Participants were told to ignore size changes and to match on shape alone. In Experiment 1, size changes on same-shape trials impaired performance similarly for both visual-to-visual and haptic-to-haptic shape matching. In Experiment 2, size changes impaired performance on both visual-to-haptic and haptic-to-visual shape matching and there was no interaction between the cost of size changes and direction of transfer. Together the unimodal and crossmodal matching results suggest that the same, size-specific perceptual representations underlie both visual and haptic object recognition, and indicate that crossmodal memory for objects must be at least partly based on common perceptual representations.  相似文献   

8.
An object in the peripheral visual field is more difficult to recognize when surrounded by other objects. This phenomenon is called “crowding”. Crowding places a fundamental constraint on human vision that limits performance on numerous tasks. It has been suggested that crowding results from spatial feature integration necessary for object recognition. However, in the absence of convincing models, this theory has remained controversial. Here, we present a quantitative and physiologically plausible model for spatial integration of orientation signals, based on the principles of population coding. Using simulations, we demonstrate that this model coherently accounts for fundamental properties of crowding, including critical spacing, “compulsory averaging”, and a foveal-peripheral anisotropy. Moreover, we show that the model predicts increased responses to correlated visual stimuli. Altogether, these results suggest that crowding has little immediate bearing on object recognition but is a by-product of a general, elementary integration mechanism in early vision aimed at improving signal quality.  相似文献   

9.
Photographs, especially of humans, are widely used as stimuli in behavioural research with pigeons. Despite their abundant use, it is not clear to what extent pigeons perceive photographs as representing three-dimensional objects. To address this question, we trained 16 pigeons to identify individual, real-life humans. This discrimination depended primarily on visual cues from the heads of the persons. Subsequently, the pigeons were shown photographs of these individuals to test for transfer to a two-dimensional representation. Successful identification of a three-dimensional person did not facilitate learning of the corresponding photographs. These results demonstrate limitations of cross-recognition of complex objects and their photographs in pigeons.  相似文献   

10.
Weakly electric fish orient at night in complete darkness by employing their active electrolocation system. They emit short electric signals and perceive the consequences of these emissions with epidermal electroreceptors. Objects are detected by analyzing the electric images which they project onto the animal's electroreceptive skin surface. This process corresponds to similar processes during vision, where visual images are cast onto the retinas of eyes. Behavioral experiments have shown that electric fish can measure the distance of objects during active electrolocation, thus possessing three-dimensional depth perception of their surroundings. The fundamental mechanism for distance determination differs from stereopsis used during vision by two-eyed animals, but resembles some supplementary mechanisms for distance deduction in humans. Weakly electric fish can also perceive the three-dimensional shape of objects. The fish can learn to identify certain objects and discriminate them from all other objects. In addition, they spontaneously categorize objects according to their shapes and not according to object size or material properties. There is good evidence that some fundamental types of perceptional invariances during visual object recognition in humans are also found in electric fish during active electrolocation. These include size invariance (maybe including size constancy), rotational invariance, and translational invariance. The mechanisms of shape detection during electrolocation are still unknown, and their discoveries require additional experiments.  相似文献   

11.
Recognizing depth-rotated objects: a review of recent research and theory   总被引:1,自引:0,他引:1  
Biederman I 《Spatial Vision》2000,13(2-3):241-253
  相似文献   

12.
On the use of size functions for shape analysis   总被引:1,自引:0,他引:1  
According to a recent mathematical theory a shape can be represented by size functions, which convey information on both the topological and metric properties of the viewed shape. In this paper the relevance of the theory of size functions to visual perception is investigated. An algorithm for the computation of the size functions is presented, and many theoretical properties of the theory are demonstrated on real images. It is shown that the representation of shape in terms of size functions (1) can be tailored to suit the invariance of the problem at hand and (2) is stable against small qualitative and quantitative changes of the viewed shape. A distance between size functions is used as a measure of similarity between the representations of two different shapes. The results obtained indicate that size functions are likely to be very useful for object recognition. In particular, they seem to be well suited for the recognition of natural and articulated objects.  相似文献   

13.
Color constancy is the term given to the ability to recognize the color of objects correctly under different conditions of illumination. For this purpose the visual system must determine the character of the illumination, introduce a correction for it into the spectal composition of the light received from the object, and hence recreate the true color of its surface. Behavioral experiments on fish showed that they possess constant color vision of objects. Electrophysiological experiments on ganglion cells of the color type showed that the simplest mechanisms of correction for illumination are found at the retinal level. An investigation of model algorithms providing for color constancy showed thatthe presence of color vision makes it much easier to recognize the three-dimensional form of objects. This fact compels a reexamination of established views regarding the place and role of color vision in functions of the animal visual system as a whole.Institute for Problems in Information Transmission, Academy of Sciences of the USSR, Moscow. Translated from Neirofiziologiya, Vol. 7, No. 1, pp. 21–26, January–February, 1975.  相似文献   

14.
Humans can effectively and swiftly recognize objects in complex natural scenes. This outstanding ability has motivated many computational object recognition models. Most of these models try to emulate the behavior of this remarkable system. The human visual system hierarchically recognizes objects in several processing stages. Along these stages a set of features with increasing complexity is extracted by different parts of visual system. Elementary features like bars and edges are processed in earlier levels of visual pathway and as far as one goes upper in this pathway more complex features will be spotted. It is an important interrogation in the field of visual processing to see which features of an object are selected and represented by the visual cortex. To address this issue, we extended a hierarchical model, which is motivated by biology, for different object recognition tasks. In this model, a set of object parts, named patches, extracted in the intermediate stages. These object parts are used for training procedure in the model and have an important role in object recognition. These patches are selected indiscriminately from different positions of an image and this can lead to the extraction of non-discriminating patches which eventually may reduce the performance. In the proposed model we used an evolutionary algorithm approach to select a set of informative patches. Our reported results indicate that these patches are more informative than usual random patches. We demonstrate the strength of the proposed model on a range of object recognition tasks. The proposed model outperforms the original model in diverse object recognition tasks. It can be seen from the experiments that selected features are generally particular parts of target images. Our results suggest that selected features which are parts of target objects provide an efficient set for robust object recognition.  相似文献   

15.
Mechanisms of explicit object recognition are often difficult to investigate and require stimuli with controlled features whose expression can be manipulated in a precise quantitative fashion. Here, we developed a novel method (called "Dots"), for generating visual stimuli, which is based on the progressive deformation of a regular lattice of dots, driven by local contour information from images of objects. By applying progressively larger deformation to the lattice, the latter conveys progressively more information about the target object. Stimuli generated with the presented method enable a precise control of object-related information content while preserving low-level image statistics, globally, and affecting them only little, locally. We show that such stimuli are useful for investigating object recognition under a naturalistic setting--free visual exploration--enabling a clear dissociation between object detection and explicit recognition. Using the introduced stimuli, we show that top-down modulation induced by previous exposure to target objects can greatly influence perceptual decisions, lowering perceptual thresholds not only for object recognition but also for object detection (visual hysteresis). Visual hysteresis is target-specific, its expression and magnitude depending on the identity of individual objects. Relying on the particular features of dot stimuli and on eye-tracking measurements, we further demonstrate that top-down processes guide visual exploration, controlling how visual information is integrated by successive fixations. Prior knowledge about objects can guide saccades/fixations to sample locations that are supposed to be highly informative, even when the actual information is missing from those locations in the stimulus. The duration of individual fixations is modulated by the novelty and difficulty of the stimulus, likely reflecting cognitive demand.  相似文献   

16.
Behavioral testing has revealed that pigeons may use the same visual information sources as humans to discriminate between three-dimensional shapes.  相似文献   

17.
Humans routinely complete partly occluded objects to recognize the whole objects. However, a number of studies using geometrical figures and even conspecific images have shown that pigeons fail to do so. In the present study, we tested whether pigeons complete partially occluded objects in a situation simulating a natural feeding context. In Experiment 1, we trained pigeons to peck at any photograph of food and not to peck at any containing a non-food object. At test, we presented both photos of food partly occluded by pigeon's feather and photos simply truncated at the same part. We predicted that if the pigeons perceptually completed the occluded portion, then they would discriminate the photos of occluded food better than the truncated photos. The result was that the pigeons pecked at the truncated photos earlier than the occluded photos. Placing the occluder next to all of the stimuli in Experiment 2 or substituting indented lozenge for the feather in Experiment 3 did not affect the results. Thus, even in a simulated ecologically significant situation, pigeons continued to not show evidence of perceptual completion.  相似文献   

18.
Neurons in the inferior temporal cortex (IT), an area crucially involved in visual object recognition in monkeys, show the visual response properties and anatomical/chemical nature which are distinct from those in the cortical areas that feed visual inputs to the IT. Earlier physiological studies showed that IT neurons have large receptive fields covering the center and contralateral (often bilateral) visual fields, stimulus selectivity for images of complex objects or shapes, and translation invariance of the stimulus selectivity. Recent studies have revealed new aspects of their properties such as invariant selectivity for shapes despite drastic changes in various physical attributes of stimuli, latent excitatory inputs masked by stimulus-specific GABAergic inhibition, selectivity for binocular disparity and 3-dimensional surface structures, profound effects of learning on the stimulus selectivity, and columnar clustering of neurons with similarstimulus selectivity for shapes and other object features. Another line of research using histological techniques have revealed that pyramidal neurons in the IT are larger in the size of dendritic arbors, in the number of dendritic branches and spines, and in the size and distribution of horizontal axonal arbors than those in the earlier areas, allowing them to integrate a larger population of afferents and process more diverse inputs. The concentration of several neurochemicals including those related to synaptic transmission or plasticity changes systematically towards the IT along the occipitotemporal pathway. Many of the characteristics of IT neurons parallel or explain certain aspects of visual object perception, although the behavioral relevance has yet to be addressed experimentally.  相似文献   

19.
In this article we review current literature on cross-modal recognition and present new findings from our studies on object and scene recognition. Specifically, we address the questions of what is the nature of the representation underlying each sensory system that facilitates convergence across the senses and how perception is modified by the interaction of the senses. In the first set of our experiments, the recognition of unfamiliar objects within and across the visual and haptic modalities was investigated under conditions of changes in orientation (0 degrees or 180 degrees ). An orientation change increased recognition errors within each modality but this effect was reduced across modalities. Our results suggest that cross-modal object representations of objects are mediated by surface-dependent representations. In a second series of experiments, we investigated how spatial information is integrated across modalities and viewpoint using scenes of familiar, 3D objects as stimuli. We found that scene recognition performance was less efficient when there was either a change in modality, or in orientation, between learning and test. Furthermore, haptic learning was selectively disrupted by a verbal interpolation task. Our findings are discussed with reference to separate spatial encoding of visual and haptic scenes. We conclude by discussing a number of constraints under which cross-modal integration is optimal for object recognition. These constraints include the nature of the task, and the amount of spatial and temporal congruency of information across the modalities.  相似文献   

20.
Rainer G  Miller EK 《Neuron》2000,27(1):179-189
The perception and recognition of objects are improved by experience. Here, we show that monkeys' ability to recognize degraded objects was improved by several days of practice with these objects. This improvement was reflected in the activity of neurons in the prefrontal (PF) cortex, a brain region critical for a wide range of visual behaviors. Familiar objects activated fewer neurons than did novel objects, but these neurons were more narrowly tuned, and the object representation was more resistant to the effects of degradation, after experience. These results demonstrate a neural correlate of visual learning in the PF cortex of adult monkeys.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号