Matches in SemOpenAlex for { <https://semopenalex.org/work/W4237562161> ?p ?o ?g. }
Showing items 1 to 58 of
58
with 100 items per page.
- W4237562161 endingPage "R850" @default.
- W4237562161 startingPage "R845" @default.
- W4237562161 abstract "“When I began to see with two eyes, my visual world completely transformed. Trees looked totally different. Consider a leafless tree in winter. Its outer branches enclose and capture a volume of space through which the inner branches permeate. I had no concept of this. Oh, I could infer which branches were in front of others using monocular cues such as object occlusion, but I could not perceive this. Trees to me looked somewhat like a drawing. Before my vision changed I would not have said that the tree looked flat, but I had no idea just how round a tree's canopy really is… When I began to see with two eyes, everything looked crisper and much better outlined. As another formerly stereoblind person wrote to me, ‘Everything has edges!’ The world was not only flatter but less detailed and textured with my monocular vision.” The above quotation, from an email contribution by Susan Barry to an on-line discussion of the benefits of stereopsis, eloquently captures many of the perceptual, aesthetic, and even philosophical aspects of ‘stereopsis’, a word derived from Greek that translates literally as: στɛρɛoσ, ‘solid’ and oπισ, ‘power of sight’. Because Susan Barry acquired this capacity as an adult, she was made vividly aware of what most of us take for granted, namely, a direct sense of the three-dimensionality of the visual world. The quotation also hints at some other critical features of binocular vision that are not related to depth perception per se. For example, the statement that “Everything has edges!” attests to the fact that seeing the world twice, as it were, from two slightly different perspectives, makes a number of important visual computations, such as the detection of edges, more robust. In this primer, we shall mainly deal with stereopsis as narrowly defined, that is, as the use of differences in the images projected onto the retinas of the two eyes — so-called ‘binocular disparity’ — to reconstruct the third visual dimension of depth. We will first describe the basic geometry of binocular disparity before discussing how the brain uses this information to compute depth. We would be remiss, however, if we did not clarify two vital distinctions that place this peculiar capacity within the larger context of vision. The first is that binocular comparisons also serve many other important functions in vision, apart from the computation of depth. The second is that there are many other cues to visual depth that do not require binocular comparisons. Thus, after discussing stereopsis proper, we will briefly treat these two issues. Because the two eyes are separated horizontally, they see the same visual scene from two slightly different vantage points. When we ‘look at’ a particular feature in space, such as the black dot on the arrow in Figure 1A, what we are doing is aiming the fovea — the tiny retinal region of highest visual acuity — of each eye at that feature. This act defines the ‘plane of fixation’. If we start at this point on the arrow and move a fixed distance along the gray circle, the projections of the new point in the two eyes will move by exactly the same distance on each retina. These points of projection on the two retinas are defined as ‘corresponding points’. The geometric horopter is the collection of all such points in the image that project to corresponding points on the retina. (To an observer these points would all appear to lie at roughly the same depth as the point of fixation. The collection of points that appear at exactly the same depth is referred to as the “empirical horopter” and is slightly flatter than the geometric horopter.) Now consider parts of an object that lie either in front of or behind the plane of fixation, such as the head and tail, respectively, of the arrow in Figure 1B. These features will not project to corresponding points on the two retinas. That is, the retinal distance from f to h on the right retina will not be equal to the distance from f′ to h′ on the left retina. This difference [(h − f) − (h′ − f′)] is called ‘binocular disparity’, and it is the basis for stereopsis. The projection lines for retinal points of near features cross in front of the horopter and thus produce what are referred to as crossed disparities (by convention, these are assigned negative values), and, conversely, far features, such as the tail of the arrow in Figure 1B, produce uncrossed disparities, which are assigned positive values. If the retinal disparities produced by a feature fall within a limited range (about 0.5° in the vicinity of the fovea; slightly more at greater eccentricity), it will be seen as a single ‘fused’ feature that is offset in depth. Beyond this range, which is known as ‘Panum's fusional area’, the feature will appear double — the observer is aware of each retinal image independently, commonly known as ‘double vision’ — even though the visual system can still extract meaningful depth information up to binocular disparities of several degrees, depending on the size of the feature. Note that the above discussion takes for granted that one knows which feature in the right eye to match with the same feature in the left eye. In fact, this can be a difficult thing to figure out, and it is known in computational circles as the “‘correspondence problem’.” The problem is quite formidable if one is trying to match single pixels in the two retinal images (because there are a huge number of possible matches), but becomes much more manageable if one uses local features — such as short line segments (see Figure 3) — to do the matching. In addition, it is likely that higher order image properties, such as the global shape of an object, are used to constrain the local feature matches. There is another type of stereopsis known as ‘da Vinci stereopsis’ (after Leonardo da Vinci who first described it), which involves the use of unpaired regions of the images in the two eyes that occur as the result of occlusion. Basically, because the eye cannot see around objects, a foreground object in the shape of a square will create a vertical strip of the background on the left side that is visible to the left eye only, and a similar strip on the right side that is visible to the right eye only (Figure 1C). This type of stereopsis is also useful for depth perception, in particular, for defining occlusive relationships, but will not be discussed further here. To distinguish the two, the more common form of stereopsis, treated here, is sometimes referred to as ‘Wheatstone stereopsis’ after its discoverer Charles Wheatstone. Because the basic data that the brain has to work with are binocular disparities, the smallest depth difference that can be distinguished is ultimately set by the minimum retinal disparity that can be resolved. In human observers, this difference approaches 5 arc-seconds (0.0014 degrees) at the fovea — a truly remarkable precision given that the spacing between cones in the fovea is on the order of 30 arc-seconds. The binocular disparity produced by a given depth difference varies as the inverse square of the viewing distance, which means that, at close viewing distances, exceedingly small depth differences can be discriminated — as tiny as 25 μm (about the width of a fine human hair). Even at relatively large viewing distances, stereopsis can play a role provided the depth dimension is large enough. For example, at a distance of 100 meters, the minimum resolvable depth is about 4 meters. The raw data for stereopsis are binocular disparities, but in order to actually calculate the depth of an image feature, the signals from the two retinas must be compared by a neural circuit. This occurs in the visual cortex. In order to understand how the brain computes stereoscopic depth, we must first review some general properties of neurons in the visual pathways. A given neuron generates electrical impulses (called ‘action potentials’ or ‘spikes’) in response to visual stimuli that occur within a small region of the visual field. This region is the neuron's receptive field. The output cells of each eye, the retinal ganglion cells, project to the lateral geniculate nucleus (LGN) of the thalamus, which, in turn, projects to the primary visual cortex (V1). It is in V1 where single neurons that receive inputs from both eyes are first observed — a necessary, but not sufficient, condition for stereopsis. The key property for such a neuron is that it is sensitive to specific differences between the receptive fields in the two eyes. This can be tested for a given neuron by presenting the same visual stimulus independently to each eye while systematically varying the position of the stimulus within each receptive field. Visually-evoked spikes are counted and plotted as a function of the difference in image position between the two eyes, that is, the binocular disparity. Many neurons in V1 are remarkably sensitive to the binocular disparity, showing high firing rates for some disparities and lower rates for others. This type of firing rate modulation, shown graphically as a ‘disparity tuning curve’, is the signature of binocular disparity-selective neurons, which comprise the biological substrate for stereopsis. Disparity tuning curves occur in many shapes, which have been traditionally categorized by the location of the peak firing response (Figure 2). What is the underlying mechanism behind this diversity of disparity tuning functions? While this question is still being actively explored, several key findings have contributed to an elegant model for the computation of binocular disparity. The key mechanisms behind this model depend on the internal shape of each neuron's interocular receptive field. Many cells in the recipient layers of the primary visual cortex have receptive fields which can be mapped out using small spots of light. These so-called simple cells will fire action potentials in response to a small spot of light appearing in some parts of their receptive fields (ON regions), but are inhibited when the same spot appears in other regions of the receptive field and only respond upon stimulus disappearance (OFF regions). This method of receptive field mapping, pioneered in the visual cortex by David Hubel and Torsten Wiesel, has revealed that most simple cells have receptive fields consisting of elongated regions of ON excitation flanked by elongated regions of OFF inhibition (Figure 3A; or the inverse, OFF regions flanked by ON). These receptive field shapes are well described by a simple mathematical function (Figure 3B), which is useful for the kinds of models discussed below. Neurons with such receptive fields are optimized for the detection of lines or edges of a particular orientation. The specificity of stimulus position and orientation required to elicit responses from V1 simple cells suggests that binocular disparity could be coded by introducing small differences between the left and right receptive fields. That is, one could simply introduce a spatial offset between the retinal positions of each receptive field to capture binocular disparity cues, so-called ‘position disparity’. Alternatively, both receptive fields could share the exact same retinal position but contain differences in the relative positions of the excitatory and inhibitory regions within each receptive field, known as ‘shape disparity’. The majority of orientation-selective cells in the primary visual cortex have more complicated functional properties, however. While simple cells require that a given stimulus be placed in a specific excitatory region within its receptive field, most neurons in V1 respond optimally to oriented lines appearing anywhere within their receptive fields. This position invariance is a hallmark of ‘complex cells’, and it is believed to arise by the combination inputs from multiple simple cells. Not surprisingly, the disparity tuning properties of complex cells are also more nuanced: complex cells can respond to their preferred binocular disparities no matter where they may be presented within their receptive fields. A more elaborate ‘energy model’ has been developed to explain the responses of complex cells. This model describes disparity-sensitive complex cells as receiving inputs from multiple simple cells, each having the same binocular disparity and orientation preference but differing in the relative position of the peak excitatory domain within each receptive field (Figure 3). In the simplest case, four simple cells with the same elongation axis and phase disparities would have their spatial frequency component shifted by one quarter of a cycle. In reality, many more simple cells — on the order of several hundred — are likely to serve as inputs to complex cells. The disparity energy model accounts quite well for the observed responses of disparity-sensitive cells. In turn, the diversity of tuning curves emerging from this mechanism provides a neural toolkit for stereopsis. How these disparity-sensitive cells contribute to the actual perception of depth has been explored using a variety of techniques, including the analysis of correlation between single-neuron activity and perceptual decisions and electrical activation of disparity-sensitive cell clusters. The most informative approaches have used monkeys trained to report the depth of a visual stimulus (Figure 4A). Investigators then record the activity of disparity-selective neurons while the animal is performing the task. Using this approach, it has been shown that single neurons are not only very sensitive to the depth signals in the visual stimulus, but they can be highly predictive of the choices that the animal will make on a given trial. This predictive power, or ‘choice probability’, constitutes strong evidence that the signals from these neurons are actually being used by the monkey's brain to solve the perceptual task. Further evidence for a causal role in perception is that electrically activating small clusters of similarly tuned binocular disparity-selective cells can directly bias a monkey's decisions in favor of the preferred disparity represented by those neurons (Figure 4B,C). Stereopsis is certainly a good reason for having two eyes, but it is not the only one. Some of the other benefits of binocularity have nothing to do with making binocular comparisons. For example, given how important vision is for many animals, it is clearly useful to have a spare eye in the event that one becomes damaged. Another potential advantage is that, if the two eyes are placed on opposite sides of the head, the combined fields of view can effectively cover the entire 360° of the visual field. This ability to ‘have eyes in the back of one's head’ might be particularly useful if one is being stalked and, indeed, is frequently found in prey, such as gazelles. Of course, this arrangement precludes stereopsis, or at least limits it to a very narrow strip where there is overlap of the monocular visual fields. This trade-off appears to play out in the animal kingdom as strategies adopted by the hunter — forward facing eyes with a large degree of binocular overlap and good stereopsis — versus the hunted — lateral eyes for panoramic vision. The hunter with mobile eyes can also make use of the angular displacement required to aim the two foveae on an object in order to compute its absolute distance. This range-finding ability would obviously be useful during encounters in which the predator must pounce on or strike at its prey. Another, often overlooked, binocular advantage is that combining the outputs of n independent detectors decreases noise by a factor of √n, so using two eyes improves the visual system's signal-to-noise ratio 1.4-fold. At very low light levels, such as those experienced by our nocturnal mammalian ancestors, this might have proven critical for survival. Finally, there is a rather large category of visual computations that make use of binocular disparity information, but that are not used for determining depth per se. Perhaps the most important of these is ‘breaking camouflage’. Imagine an animal with a coat that is perfectly matched in color and texture to that of its surroundings (the visual background against which it is viewed). As long as this animal remains stationary, it would be invisible to a single eye, but a mechanism exquisitely sensitive to binocular disparity would cause the animal's outline to ‘pop out’ from the background, because the animal and the background are at different depths. This use of disparity processing for figure-ground segregation was vividly demonstrated by the random dot stereograms created by Bela Julesz and can be seen in the once popular ‘Magic Eye’ images (Figure 5A). In these cases, the figure is visible only when viewed stereoscopically, so it is not so much a matter of perceiving the depth metrically but rather of grouping the relevant regions of the image so that a shape can be recognized. A related role is the determination of border ownership, which has profound effects on the linking of features that may be partially occluded. In this case, binocular comparisons aid in distinguishing real object boundaries from those that are created by occlusion. This effect was dramatically demonstrated by Ken Nakayama and colleagues, and can be appreciated even in a non-stereo version of their task: The face in Figure 5B is much more difficult to recognize when it appears in front of the occluding strips rather than behind them. A host of visual stimuli are perceived differently depending on the relative depth they are assigned. For example, the same small bright patch on a curved surface is interpreted differently depending on its relative binocular disparity: as a reflection of the light source from a shiny surface if it appears behind the surface or as paint if it is at the same depth as the surface. In sum, binocular disparity processing is not just for depth perception, but rather permeates many aspects of the act of seeing. A significant proportion of the human population — anywhere from 5% to 10% — lacks stereoscopic depth perception, just as Susan Barry did prior to her therapy. Given the discussion thus far, we might expect these people to be severely visually impaired, but, remarkably, they are not. Some, in fact, have managed to succeed at careers that would seem to place a high demand on good depth perception, such as surgery, professional skiing and dentistry. The reason this is possible is that there are a number of other methods by which depth information can be obtained. One of the most useful is to move the head in order to sample multiple views over time, producing ‘motion parallax’. When one moves one's head in this manner — a number of animal species and some monocular humans use this technique of ‘head bobbing’ — objects nearer than the point of fixation will move in a direction opposite to the direction of head motion and those farther away will move in the same direction. The speed of the motion indicates the magnitude of the depth difference. When tested under similar conditions, depth discrimination from motion parallax can be almost as good as that from stereopsis, differing by only a factor of two or so. In addition to motion parallax, there are a number of other ‘pictorial’ cues to depth. These are the cues that artists exploit to create the illusion of depth on a flat canvas. They include linear perspective (the tendency for parallel lines to converge as they recede towards the horizon), size and texture gradients (objects shrink and textures become finer with distance), occlusion (an object that partially occludes another is perceived as being nearer), atmospheric effects (distant objects become partially blurred as the result of particulates in the air, such as water vapor or pollution), and luminance (brighter objects appear closer). While none of these cues can produce the fine depth discrimination achievable by stereopsis, they provide an overall sense of the three-dimensional spatial structure of the environment, and, when combined with motion parallax, they can almost completely compensate for a lack of stereopsis. This appears to be especially true in individuals who have lacked stereopsis from a very young age and thus had to learn to navigate and perform tasks using only monocular cues. Thus, while not essential for vision, binocular disparity processing is remarkably useful for a host of visual computations, only one of which is the determination of depth, or stereopsis. It is also an excellent example of the amazing precision that neural circuits can achieve — differences in relative retinal position can be detected that are nearly a factor of ten smaller than the spacing between photoreceptors, perhaps the most dramatic example of ‘hyperacuity’. Because stereopsis is a highly derived measure — requiring the precise comparison of signals from the two eyes — it has also provided an excellent test-bed for models of how neurons compute and how these computations are related to perception. We thank Greg DeAngelis, Marge Livingstone, Ken Nakayama, and members of the Born lab for helpful comments." @default.
- W4237562161 created "2022-05-12" @default.
- W4237562161 creator A5081465951 @default.
- W4237562161 creator A5089263610 @default.
- W4237562161 date "2008-09-01" @default.
- W4237562161 modified "2023-10-18" @default.
- W4237562161 title "Stereopsis" @default.
- W4237562161 cites W1647993270 @default.
- W4237562161 cites W1971261228 @default.
- W4237562161 cites W2011610744 @default.
- W4237562161 cites W2011839065 @default.
- W4237562161 cites W2066065672 @default.
- W4237562161 cites W2093540385 @default.
- W4237562161 cites W2125838369 @default.
- W4237562161 cites W4244575955 @default.
- W4237562161 doi "https://doi.org/10.1016/j.cub.2008.07.006" @default.
- W4237562161 hasPubMedId "https://pubmed.ncbi.nlm.nih.gov/18812074" @default.
- W4237562161 hasPublicationYear "2008" @default.
- W4237562161 type Work @default.
- W4237562161 citedByCount "29" @default.
- W4237562161 countsByYear W42375621612012 @default.
- W4237562161 countsByYear W42375621612013 @default.
- W4237562161 countsByYear W42375621612014 @default.
- W4237562161 countsByYear W42375621612016 @default.
- W4237562161 countsByYear W42375621612017 @default.
- W4237562161 countsByYear W42375621612018 @default.
- W4237562161 countsByYear W42375621612019 @default.
- W4237562161 countsByYear W42375621612020 @default.
- W4237562161 countsByYear W42375621612021 @default.
- W4237562161 countsByYear W42375621612022 @default.
- W4237562161 countsByYear W42375621612023 @default.
- W4237562161 crossrefType "journal-article" @default.
- W4237562161 hasAuthorship W4237562161A5081465951 @default.
- W4237562161 hasAuthorship W4237562161A5089263610 @default.
- W4237562161 hasBestOaLocation W42375621611 @default.
- W4237562161 hasConcept C86803240 @default.
- W4237562161 hasConceptScore W4237562161C86803240 @default.
- W4237562161 hasIssue "18" @default.
- W4237562161 hasLocation W42375621611 @default.
- W4237562161 hasLocation W42375621612 @default.
- W4237562161 hasOpenAccess W4237562161 @default.
- W4237562161 hasPrimaryLocation W42375621611 @default.
- W4237562161 hasRelatedWork W1641042124 @default.
- W4237562161 hasRelatedWork W1990804418 @default.
- W4237562161 hasRelatedWork W1993764875 @default.
- W4237562161 hasRelatedWork W2013243191 @default.
- W4237562161 hasRelatedWork W2051339581 @default.
- W4237562161 hasRelatedWork W2082860237 @default.
- W4237562161 hasRelatedWork W2117258802 @default.
- W4237562161 hasRelatedWork W2130076355 @default.
- W4237562161 hasRelatedWork W2151865869 @default.
- W4237562161 hasRelatedWork W4234157524 @default.
- W4237562161 hasVolume "18" @default.
- W4237562161 isParatext "false" @default.
- W4237562161 isRetracted "false" @default.
- W4237562161 workType "article" @default.