首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Image motion is a primary source of visual information about the world. However, before this information can be used the visual system must determine the spatio-temporal displacements of the features in the dynamic retinal image, which originate from objects moving in space. This is known as the motion correspondence problem. We investigated whether cross-cue matching constraints contribute to the solution of this problem, which would be consistent with physiological reports that many directionally selective cells in the visual cortex also respond to additional visual cues. We measured the maximum displacement limit (Dmax) for two-frame apparent motion sequences. Dmax increases as the number of elements in such sequences decreases. However, in our displays the total number of elements was kept constant while the number of a subset of elements, defined by a difference in contrast polarity, binocular disparity or colour, was varied. Dmax increased as the number of elements distinguished by a particular cue was decreased. Dmax was affected by contrast polarity for all observers, but only some observers were influenced by binocular disparity and others by colour information. These results demonstrate that the human visual system exploits local, cross-cue matching constraints in the solution of the motion correspondence problem.  相似文献   

2.
Both dorsal and ventral cortical visual streams contain neurons sensitive to binocular disparities, but the two streams may underlie different aspects of stereoscopic vision. Here we investigate stereopsis in the neurological patient D.F., whose ventral stream, specifically lateral occipital cortex, has been damaged bilaterally, causing profound visual form agnosia. Despite her severe damage to cortical visual areas, we report that DF''s stereo vision is strikingly unimpaired. She is better than many control observers at using binocular disparity to judge whether an isolated object appears near or far, and to resolve ambiguous structure-from-motion. DF is, however, poor at using relative disparity between features at different locations across the visual field. This may stem from a difficulty in identifying the surface boundaries where relative disparity is available. We suggest that the ventral processing stream may play a critical role in enabling healthy observers to extract fine depth information from relative disparities within one surface or between surfaces located in different parts of the visual field.  相似文献   

3.
Richards (1985) showed that veridical three-dimensional shape may be recovered from the integration of binocular disparity and retinal motion information, but proposed that this integration may only occur for horizontal retinal motion. Psychophysical evidence supporting the combination of stereo and motion information is limited to the case of horizontal motion (Johnston et al., 1994), and has been criticised on the grounds of potential object boundary cues to shape present in the stimuli. We investigated whether veridical shape can be recovered under more general conditions. Observers viewed cylinders that were defined by binocular disparity, two-frame motion or a combination of disparity and motion, presented at simulated distances of 30 cm, 90 cm or 150 cm. Horizontally and vertically oriented cylinders were rotated about vertical and horizontal axes. When rotation was about the cylinder's own axis, no boundary cues to shape were introduced. Settings were biased for the disparity and two-frame motion stimuli, while more veridical shape judgements were made under all conditions for combined cue stimuli. These results demonstrate that the improved perception of three-dimensional shape in these stimuli is not a consequence of the presence of object boundary cues, and that the combination of disparity and motion is not restricted to horizontal image motion.  相似文献   

4.
It is shown that existing processing schemes of 3D motion perception such as interocular velocity difference, changing disparity over time, as well as joint encoding of motion and disparity, do not offer a general solution to the inverse optics problem of local binocular 3D motion. Instead we suggest that local velocity constraints in combination with binocular disparity and other depth cues provide a more flexible framework for the solution of the inverse problem. In the context of the aperture problem we derive predictions from two plausible default strategies: (1) the vector normal prefers slow motion in 3D whereas (2) the cyclopean average is based on slow motion in 2D. Predicting perceived motion directions for ambiguous line motion provides an opportunity to distinguish between these strategies of 3D motion processing. Our theoretical results suggest that velocity constraints and disparity from feature tracking are needed to solve the inverse problem of 3D motion perception. It seems plausible that motion and disparity input is processed in parallel and integrated late in the visual processing hierarchy.  相似文献   

5.
Relative binocular disparity cannot tell us the absolute 3D shape of an object, nor the 3D trajectory of its motion, unless the visual system has independent access to how far away the object is at any moment. Indeed, as the viewing distance is changed, the same disparate retinal motions will correspond to very different real 3D trajectories. In this paper we were interested in whether binocular 3D motion detection is affected by viewing distance. A visual search task was used, in which the observer is asked to detect a target dot, moving in 3D, amidst 3D stationary distractor dots. We found that distance does not affect detection performance. Motion-in-depth is consistently harder to detect than the equivalent lateral motion, for all viewing distances. For a constant retinal motion with both lateral and motion-in-depth components, detection performance is constant despite variations in viewing distance that produce large changes in the direction of the 3D trajectory. We conclude that binocular 3D motion detection relies on retinal, not absolute, visual signals.  相似文献   

6.
Vision plays a major role in many spiders, being involved in prey hunting, orientation or substrate choice, among others. In Misumena vatia, which experiences morphological color changes, vision has been reported to be involved in substrate color matching. Electrophysiological evidence reveals that at least two types of photoreceptors are present in this species, but these data are not backed up by morphological evidence. This work analyzes the functional structure of the eyes of this spider and relates it to its color-changing abilities. A broad superposition of the visual field of the different eyes was observed, even between binocular regions of principal and secondary eyes. The frontal space is simultaneously analyzed by four eyes. This superposition supports the integration of the visual information provided by the different eye types. The mobile retina of the principal eyes of this spider is organized in three layers of three different types of rhabdoms. The third and deepest layer is composed by just one large rhabdom surrounded by dark screening pigments that limit the light entry. The three pairs of secondary eyes have all a single layer of rhabdoms. Our findings provide strong support for an involvement of the visual system in color matching in this spider.  相似文献   

7.
The visual cortex is able to extract disparity information through the use of binocular cells. This process is reflected by the Disparity Energy Model, which describes the role and functioning of simple and complex binocular neuron populations, and how they are able to extract disparity. This model uses explicit cell parameters to mathematically determine preferred cell disparities, like spatial frequencies, orientations, binocular phases and receptive field positions. However, the brain cannot access such explicit cell parameters; it must rely on cell responses. In this article, we implemented a trained binocular neuronal population, which encodes disparity information implicitly. This allows the population to learn how to decode disparities, in a similar way to how our visual system could have developed this ability during evolution. At the same time, responses of monocular simple and complex cells can also encode line and edge information, which is useful for refining disparities at object borders. The brain should then be able, starting from a low-level disparity draft, to integrate all information, including colour and viewpoint perspective, in order to propagate better estimates to higher cortical areas.  相似文献   

8.
In most respects, the response properties of cells in the secondary visual cortex of the newborn lamb were indistinguishable from those in the adult. The cells were sharply selective to orientation; the orientation preferences were the same in each eye, and they varied systematically as the electrode penetrated the cortex. The receptive-field organization did not differ noticeably from that in adults, and complex, hypercomplex, and a few simple cells were all observed. The ocular dominance distribution was similar to that in the adult. Most importantly, binocular cells were found with disparate receptive fields even in newborn, visually inexperienced animals. As in the adult, the disparities were largely horizontal, and they appeared to be arranged in columns. Many of the cells responded preferentially to a binocular stimulus at a particular disparity setting (often approximately zero), but unlike those in the adult almost all the binocular cells in the newborn lamb would also respond monocularly, and the enhancement at the optimal disparity was less than in the adult. The full development of binocular selectivity took several weeks, and was blocked by binocular deprivation. We conclude that the basic wiring of stereoscopic mechanisms is innate, but the development of mature binocular interaction may depend on an adaptive process which makes use of the visual information received during binocular stimulation.  相似文献   

9.
视差检测:简单细胞、复杂细胞及能量模型   总被引:2,自引:0,他引:2  
立体视觉信息的处理在于皮层双眼性细胞的活动.皮层中简单细胞对视差的编码方式被认为有两种:位置差(position shift)和相位差(phase shift),但简单细胞并不适合作为视差检测器.对一些复杂细胞的视差响应特性的生理研究,发现复杂细胞是一种比较适合的视差检测器.模型的研究提出基于这类简单细胞的复杂细胞能量模型,可以很好的检测视差,并可以较好的解释一些生理现象.  相似文献   

10.
In the information processing procedure of stereo vision, the uniqueness constraint has been used as one of the constraints to solve the "correspondence problem". While the uniqueness constraint is valid in most cases, whether it is still valid in some particular stimulus configuration (such as Panum's limiting case) has been a problem of widespread debate for a long time. To investigate the problem, we adopted the Panum's limiting case as its basic stimulus configuration, and delved into the phenomenon of binocular fusion from two distinct aspects: visual direction and orientation disparity. The results show that in Panum's limiting case binocular fusion does not comply with the rules governing regular binocular fusion as far as visual direction and orientation disparity are concerned. This indicates that double fusion does not happen in Panum's limiting case and that the uniqueness constraint is still valid.  相似文献   

11.
Although many sources of three-dimensional information have been isolated and demonstrated to contribute independently, to depth vision in animal studies, it is not clear whether these distinct cues are perceived to be perceptually equivalent. Such ability is observed in humans and would seem to be advantageous for animals as well in coping with the often co-varying (or ambiguous) information about the layout of physical space. We introduce the expression primary-depth-cue equivalence to refer to the ability to perceive mutually consistent information about differences in depth from either stereopsis or motion-parallax. We found that owls trained to detect relative depth as a perceptual category (objects versus holes) when specified by binocular disparity alone (stereopsis), immediately transferred this discrimination to novel stimuli where the equivalent depth categories were available only through differences in motion information produced by head movements (observer-produced motion-parallax). Motion-parallax discrimination did occur under monocular viewing conditions and reliable performance depended heavily on the amplitude of side-to-side head movements. The presence of primary-depth-cue equivalence in the visual system of the owl provides further conformation of the hypothesis that neural systems evolved to detect differences in either disparity or motion information are likely to share similar processing mechanisms.  相似文献   

12.
Tsao DY  Conway BR  Livingstone MS 《Neuron》2003,38(1):103-114
Binocular simple cells in primary visual cortex (V1) are the first cells along the mammalian visual pathway to receive input from both eyes. Two models of how binocular simple cells could extract disparity information have been put forward. The phase-shift model proposes that the receptive fields in the two eyes have different subunit organizations, while the position-shift model proposes that they have different overall locations. In five fixating macaque monkeys, we recorded from 30 disparity-tuned simple cells that showed selectivity to the disparity in a random dot stereogram. High-resolution maps of the left and right eye receptive fields indicated that both phase and position shifts were common. Single cells usually showed a combination of the two, and the optimum disparity was best correlated with the sum of receptive field phase and position shift.  相似文献   

13.
Presenting the eyes with spatially mismatched images causes a phenomenon known as binocular rivalry-a fluctuation of awareness whereby each eye's image alternately determines perception. Binocular rivalry is used to study interocular conflict resolution and the formation of conscious awareness from retinal images. Although the spatial determinants of rivalry have been well-characterized, the temporal determinants are still largely unstudied. We confirm a previous observation that conflicting images do not need to be presented continuously or simultaneously to elicit binocular rivalry. This process has a temporal limit of about 350 ms, which is an order of magnitude larger than the visual system's temporal resolution. We characterize this temporal limit of binocular rivalry by showing that it is independent of low-level information such as interocular timing differences, contrast-reversals, stimulus energy, and eye-of-origin information. This suggests the temporal factors maintaining rivalry relate more to higher-level form information, than to low-level visual information. Systematically comparing the role of form and motion-the processing of which may be assigned to ventral and dorsal visual pathways, respectively-reveals that this temporal limit is determined by form conflict rather than motion conflict. Together, our findings demonstrate that binocular conflict resolution depends on temporally coarse form-based processing, possibly originating in the ventral visual pathway.  相似文献   

14.
In the information processing procedure of stereo vision, the uniqueness constraint has been used as one of the constraints to solve the “correspondence problem”. While the uniqueness constraint is valid in most cases, whether it is still valid in some particular stimulus configuration (such as Panum’s limiting case) has been a problem of widespread debate for a long time. To investigate the problem, we adopted the Panum’s limiting case as its basic stimulus configuration, and delved into the phenomenon of binocular fusion from two distinct aspects: visual direction and orientation disparity. The results show that in Panum’s limiting case binocular fusion does not comply with the rules governing regular binocular fusion as far as visual direction and orientation disparity are concerned. This indicates that double fusion does not happen in Panum’s limiting case and that the uniqueness constraint is still valid.  相似文献   

15.
In the information processing procedure of stereo vision, the uniqueness constraint has been used as one of the constraints to solve the “correspondence problem”. While the uniqueness constraint is valid in most cases, whether it is still valid in some particular stimulus configuration (such as Panum’s limiting case) has been a problem of widespread debate for a long time. To investigate the problem, we adopted the Panum’s limiting case as its basic stimulus configuration, and delved into the phenomenon of binocular fusion from two distinct aspects: visual direction and orientation disparity. The results show that in Panum’s limiting case binocular fusion does not comply with the rules governing regular binocular fusion as far as visual direction and orientation disparity are concerned. This indicates that double fusion does not happen in Panum’s limiting case and that the uniqueness constraint is still valid.  相似文献   

16.
Primary visual cortex is often viewed as a “cyclopean retina”, performing the initial encoding of binocular disparities between left and right images. Because the eyes are set apart horizontally in the head, binocular disparities are predominantly horizontal. Yet, especially in the visual periphery, a range of non-zero vertical disparities do occur and can influence perception. It has therefore been assumed that primary visual cortex must contain neurons tuned to a range of vertical disparities. Here, I show that this is not necessarily the case. Many disparity-selective neurons are most sensitive to changes in disparity orthogonal to their preferred orientation. That is, the disparity tuning surfaces, mapping their response to different two-dimensional (2D) disparities, are elongated along the cell''s preferred orientation. Because of this, even if a neuron''s optimal 2D disparity has zero vertical component, the neuron will still respond best to a non-zero vertical disparity when probed with a sub-optimal horizontal disparity. This property can be used to decode 2D disparity, even allowing for realistic levels of neuronal noise. Even if all V1 neurons at a particular retinotopic location are tuned to the expected vertical disparity there (for example, zero at the fovea), the brain could still decode the magnitude and sign of departures from that expected value. This provides an intriguing counter-example to the common wisdom that, in order for a neuronal population to encode a quantity, its members must be tuned to a range of values of that quantity. It demonstrates that populations of disparity-selective neurons encode much richer information than previously appreciated. It suggests a possible strategy for the brain to extract rarely-occurring stimulus values, while concentrating neuronal resources on the most commonly-occurring situations.  相似文献   

17.
Binocular correspondence must be determined if disparity is to be used to provide information about three-dimensional shape. The current study investigated whether knowledge of the statistical distribution of disparities in the natural environment is employed in this process. A simple model, which produces distributions of distances similar to those found in the natural environment, was used to predict the distribution of disparities in natural images. This model predicts that crossed disparities will be more likely as (i) stimulus elevation decreases below fixation and (ii) fixation distance increases. To determine whether these factors influence binocular correspondence for human observers, ambiguous stereograms were presented to observers, as stimulus elevation and fixation distance were manipulated. Clear biases were observed in the depth perceived in these stereograms, which were more likely to be seen as closer than fixation (i) for stimuli presented below fixation and (ii) as fixation distance increased. These results suggest that binocular correspondence is determined in a manner consistent with the distributions of disparities expected in natural scenes.  相似文献   

18.
Stereo "3D" depth perception requires the visual system to extract binocular disparities between the two eyes' images. Several current models of this process, based on the known physiology of primary visual cortex (V1), do this by computing a piecewise-frontoparallel local cross-correlation between the left and right eye's images. The size of the "window" within which detectors examine the local cross-correlation corresponds to the receptive field size of V1 neurons. This basic model has successfully captured many aspects of human depth perception. In particular, it accounts for the low human stereoresolution for sinusoidal depth corrugations, suggesting that the limit on stereoresolution may be set in primary visual cortex. An important feature of the model, reflecting a key property of V1 neurons, is that the initial disparity encoding is performed by detectors tuned to locally uniform patches of disparity. Such detectors respond better to square-wave depth corrugations, since these are locally flat, than to sinusoidal corrugations which are slanted almost everywhere. Consequently, for any given window size, current models predict better performance for square-wave disparity corrugations than for sine-wave corrugations at high amplitudes. We have recently shown that this prediction is not borne out: humans perform no better with square-wave than with sine-wave corrugations, even at high amplitudes. The failure of this prediction raised the question of whether stereoresolution may actually be set at later stages of cortical processing, perhaps involving neurons tuned to disparity slant or curvature. Here we extend the local cross-correlation model to include existing physiological and psychophysical evidence indicating that larger disparities are detected by neurons with larger receptive fields (a size/disparity correlation). We show that this simple modification succeeds in reconciling the model with human results, confirming that stereoresolution for disparity gratings may indeed be limited by the size of receptive fields in primary visual cortex.  相似文献   

19.
Perceived depth is conveyed by multiple cues, including binocular disparity and luminance shading. Depth perception from luminance shading information depends on the perceptual assumption for the incident light, which has been shown to default to a diffuse illumination assumption. We focus on the case of sinusoidally corrugated surfaces to ask how shading and disparity cues combine defined by the joint luminance gradients and intrinsic disparity modulation that would occur in viewing the physical corrugation of a uniform surface under diffuse illumination. Such surfaces were simulated with a sinusoidal luminance modulation (0.26 or 1.8 cy/deg, contrast 20%-80%) modulated either in-phase or in opposite phase with a sinusoidal disparity of the same corrugation frequency, with disparity amplitudes ranging from 0’-20’. The observers’ task was to adjust the binocular disparity of a comparison random-dot stereogram surface to match the perceived depth of the joint luminance/disparity-modulated corrugation target. Regardless of target spatial frequency, the perceived target depth increased with the luminance contrast and depended on luminance phase but was largely unaffected by the luminance disparity modulation. These results validate the idea that human observers can use the diffuse illumination assumption to perceive depth from luminance gradients alone without making an assumption of light direction. For depth judgments with combined cues, the observers gave much greater weighting to the luminance shading than to the disparity modulation of the targets. The results were not well-fit by a Bayesian cue-combination model weighted in proportion to the variance of the measurements for each cue in isolation. Instead, they suggest that the visual system uses disjunctive mechanisms to process these two types of information rather than combining them according to their likelihood ratios.  相似文献   

20.
As we move through the world, our eyes acquire a sequence of images. The information from this sequence is sufficient to determine the structure of a three-dimensional scene, up to a scale factor determined by the distance that the eyes have moved. Previous evidence shows that the human visual system accounts for the distance the observer has walked and the separation of the eyes when judging the scale, shape, and distance of objects. However, in an immersive virtual-reality environment, observers failed to notice when a scene expanded or contracted, despite having consistent information about scale from both distance walked and binocular vision. This failure led to large errors in judging the size of objects. The pattern of errors cannot be explained by assuming a visual reconstruction of the scene with an incorrect estimate of interocular separation or distance walked. Instead, it is consistent with a Bayesian model of cue integration in which the efficacy of motion and disparity cues is greater at near viewing distances. Our results imply that observers are more willing to adjust their estimate of interocular separation or distance walked than to accept that the scene has changed in size.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号