Hubbry Logo
Visual systemVisual systemMain
Open search
Visual system
Community hub
Visual system
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Visual system
Visual system
from Wikipedia
Visual system
The visual system includes the eyes, the connecting pathways through to the visual cortex and other parts of the brain (human system shown).
The eye is the sensory organ of the visual system. The iris, pupil, and sclera are visible
Identifiers
FMA7191
Anatomical terminology

The visual system is the physiological basis of visual perception (the ability to detect and process light). The system detects, transduces and interprets information concerning light within the visible range to construct an image and build a mental model of the surrounding environment. The visual system is associated with the eye and functionally divided into the optical system (including cornea and lens) and the neural system (including the retina and visual cortex).

The visual system performs a number of complex tasks based on the image forming functionality of the eye, including the formation of monocular images, the neural mechanisms underlying stereopsis and assessment of distances to (depth perception) and between objects, motion perception, pattern recognition, accurate motor coordination under visual guidance, and colour vision. Together, these facilitate higher order tasks, such as object identification. The neuropsychological side of visual information processing is known as visual perception, an abnormality of which is called visual impairment, and a complete absence of which is called blindness. The visual system also has several non-image forming visual functions, independent of visual perception, including the pupillary light reflex and circadian photoentrainment.

This article describes the human visual system, which is representative of mammalian vision, and to a lesser extent the vertebrate visual system.

System overview

[edit]
This diagram linearly (unless otherwise mentioned) tracks the projections of all known structures that allow for vision to their relevant endpoints in the human brain. Click to enlarge the image.
Representation of optic pathways from each of the 4 quadrants of view for both eyes simultaneously

Optical

[edit]

Together, the cornea and lens refract light into a small image and shine it on the retina. The retina transduces this image into electrical pulses using rods and cones. The optic nerve then carries these pulses through the optic canal. Upon reaching the optic chiasm the nerve fibers decussate (left becomes right). The fibers then branch and terminate in three places.[1][2][3][4][5][6][7]

Neural

[edit]

Most of the optic nerve fibers end in the lateral geniculate nucleus (LGN). Before the LGN forwards the pulses to V1 of the visual cortex (primary) it gauges the range of objects and tags every major object with a velocity tag. These tags predict object movement.

The LGN also sends some fibers to V2 and V3.[8][9][10][11][12]

V1 performs edge-detection to understand spatial organization (initially, 40 milliseconds in, focusing on even small spatial and color changes. Then, 100 milliseconds in, upon receiving the translated LGN, V2, and V3 info, also begins focusing on global organization). V1 also creates a bottom-up saliency map to guide attention or gaze shift.[13]

V2 both forwards (direct and via pulvinar) pulses to V1 and receives them. Pulvinar is responsible for saccade and visual attention. V2 serves much the same function as V1, however, it also handles illusory contours, determining depth by comparing left and right pulses (2D images), and foreground distinguishment. V2 connects to V1 - V5.

V3 helps process 'global motion' (direction and speed) of objects. V3 connects to V1 (weak), V2, and the inferior temporal cortex.[14][15]

V4 recognizes simple shapes, and gets input from V1 (strong), V2, V3, LGN, and pulvinar.[16] V5's outputs include V4 and its surrounding area, and eye-movement motor cortices (frontal eye-field and lateral intraparietal area).

V5's functionality is similar to that of the other V's, however, it integrates local object motion into global motion on a complex level. V6 works in conjunction with V5 on motion analysis. V5 analyzes self-motion, whereas V6 analyzes motion of objects relative to the background. V6's primary input is V1, with V5 additions. V6 houses the topographical map for vision. V6 outputs to the region directly around it (V6A). V6A has direct connections to arm-moving cortices, including the premotor cortex.[17][18]

The inferior temporal gyrus recognizes complex shapes, objects, and faces or, in conjunction with the hippocampus, creates new memories.[19] The pretectal area is seven unique nuclei. Anterior, posterior and medial pretectal nuclei inhibit pain (indirectly), aid in REM, and aid the accommodation reflex, respectively.[20] The Edinger-Westphal nucleus moderates pupil dilation and aids (since it provides parasympathetic fibers) in convergence of the eyes and lens adjustment.[21] Nuclei of the optic tract are involved in smooth pursuit eye movement and the accommodation reflex, as well as REM.

The suprachiasmatic nucleus is the region of the hypothalamus that halts production of melatonin (indirectly) at first light.[22]

Structure

[edit]
The human eye (horizontal section)
The image projected onto the retina is inverted due to the optics of the eye.

These are components of the visual pathway, also called the optic pathway,[23] that can be divided into anterior and posterior visual pathways. The anterior visual pathway refers to structures involved in vision before the lateral geniculate nucleus. The posterior visual pathway refers to structures after this point.

Eye

[edit]

Light entering the eye is refracted as it passes through the cornea. It then passes through the pupil (controlled by the iris) and is further refracted by the lens. The cornea and lens act together as a compound lens to project an inverted image onto the retina.

S. Ramón y Cajal, Structure of the Mammalian Retina, 1900

Retina

[edit]

The retina consists of many photoreceptor cells which contain particular protein molecules called opsins. In humans, two types of opsins are involved in conscious vision: rod opsins and cone opsins. (A third type, melanopsin in some retinal ganglion cells (RGC), part of the body clock mechanism, is probably not involved in conscious vision, as these RGC do not project to the lateral geniculate nucleus but to the pretectal olivary nucleus.[24]) An opsin absorbs a photon (a particle of light) and transmits a signal to the cell through a signal transduction pathway, resulting in hyper-polarization of the photoreceptor.

Rods and cones differ in function. Rods are found primarily in the periphery of the retina and are used to see at low levels of light. Each human eye contains 120 million rods. Cones are found primarily in the center (or fovea) of the retina.[25] There are three types of cones that differ in the wavelengths of light they absorb; they are usually called short or blue, middle or green, and long or red. Cones mediate day vision and can distinguish color and other features of the visual world at medium and high light levels. Cones are larger and much less numerous than rods (there are 6-7 million of them in each human eye).[25]

In the retina, the photoreceptors synapse directly onto bipolar cells, which in turn synapse onto ganglion cells of the outermost layer, which then conduct action potentials to the brain. A significant amount of visual processing arises from the patterns of communication between neurons in the retina. About 130 million photo-receptors absorb light, yet roughly 1.2 million axons of ganglion cells transmit information from the retina to the brain. The processing in the retina includes the formation of center-surround receptive fields of bipolar and ganglion cells in the retina, as well as convergence and divergence from photoreceptor to bipolar cell. In addition, other neurons in the retina, particularly horizontal and amacrine cells, transmit information laterally (from a neuron in one layer to an adjacent neuron in the same layer), resulting in more complex receptive fields that can be either indifferent to color and sensitive to motion or sensitive to color and indifferent to motion.[26]

Mechanism of generating visual signals
[edit]

The retina adapts to change in light through the use of the rods. In the dark, the chromophore retinal has a bent shape called cis-retinal (referring to a cis conformation in one of the double bonds). When light interacts with the retinal, it changes conformation to a straight form called trans-retinal and breaks away from the opsin. This is called bleaching because the purified rhodopsin changes from violet to colorless in the light. At baseline in the dark, the rhodopsin absorbs no light and releases glutamate, which inhibits the bipolar cell. This inhibits the release of neurotransmitters from the bipolar cells to the ganglion cell. When there is light present, glutamate secretion ceases, thus no longer inhibiting the bipolar cell from releasing neurotransmitters to the ganglion cell and therefore an image can be detected.[27][28]

The final result of all this processing is five different populations of ganglion cells that send visual (image-forming and non-image-forming) information to the brain:[26]

  1. M cells, with large center-surround receptive fields that are sensitive to depth, indifferent to color, and rapidly adapt to a stimulus;
  2. P cells, with smaller center-surround receptive fields that are sensitive to color and shape;
  3. K cells, with very large center-only receptive fields that are sensitive to color and indifferent to shape or depth;
  4. another population that is intrinsically photosensitive; and
  5. a final population that is used for eye movements.[26]

A 2006 University of Pennsylvania study calculated the approximate bandwidth of human retinas to be about 8,960 kilobits per second, whereas guinea pig retinas transfer at about 875 kilobits.[29]

In 2007 Zaidi and co-researchers on both sides of the Atlantic studying patients without rods and cones, discovered that the novel photoreceptive ganglion cell in humans also has a role in conscious and unconscious visual perception.[30] The peak spectral sensitivity was 481 nm. This shows that there are two pathways for vision in the retina – one based on classic photoreceptors (rods and cones) and the other, newly discovered, based on photo-receptive ganglion cells which act as rudimentary visual brightness detectors.

Photochemistry

[edit]

The functioning of a camera is often compared with the workings of the eye, mostly since both focus light from external objects in the field of view onto a light-sensitive medium. In the case of the camera, this medium is film or an electronic sensor; in the case of the eye, it is an array of visual receptors. With this simple geometrical similarity, based on the laws of optics, the eye functions as a transducer, as does a CCD camera.

In the visual system, retinal, technically called retinene1 or "retinaldehyde", is a light-sensitive molecule found in the rods and cones of the retina. Retinal is the fundamental structure involved in the transduction of light into visual signals, i.e. nerve impulses in the ocular system of the central nervous system. In the presence of light, the retinal molecule changes configuration and as a result, a nerve impulse is generated.[26]

Optic nerve

[edit]
Information flow from the eyes (top), crossing at the optic chiasma, joining left and right eye information in the optic tract, and layering left and right visual stimuli in the lateral geniculate nucleus. V1 in red at bottom of image. (1543 image from Andreas Vesalius' Fabrica)

The information about the image via the eye is transmitted to the brain along the optic nerve. Different populations of ganglion cells in the retina send information to the brain through the optic nerve. About 90% of the axons in the optic nerve go to the lateral geniculate nucleus in the thalamus. These axons originate from the M, P, and K ganglion cells in the retina, see above. This parallel processing is important for reconstructing the visual world; each type of information will go through a different route to perception. Another population sends information to the superior colliculus in the midbrain, which assists in controlling eye movements (saccades)[31] as well as other motor responses.

A final population of photosensitive ganglion cells, containing melanopsin for photosensitivity, sends information via the retinohypothalamic tract to the pretectum (pupillary reflex), to several structures involved in the control of circadian rhythms and sleep such as the suprachiasmatic nucleus (the biological clock), and to the ventrolateral preoptic nucleus (a region involved in sleep regulation).[32] A recently discovered role for photoreceptive ganglion cells is that they mediate conscious and unconscious vision – acting as rudimentary visual brightness detectors as shown in rodless coneless eyes.[30]

Optic chiasm

[edit]

The optic nerves from both eyes meet and cross at the optic chiasm,[33][34] at the base of the hypothalamus of the brain. At this point, the information coming from both eyes is combined and then splits according to the visual field. The corresponding halves of the field of view (right and left) are sent to the left and right halves of the brain, respectively, to be processed. That is, the right side of primary visual cortex deals with the left half of the field of view from both eyes, and similarly for the left brain.[31] A small region in the center of the field of view is processed redundantly by both halves of the brain.

Optic tract

[edit]

Information from the right visual field (now on the left side of the brain) travels in the left optic tract. Information from the left visual field travels in the right optic tract. Each optic tract terminates in the lateral geniculate nucleus (LGN) in the thalamus.

Six layers in the LGN

Lateral geniculate nucleus

[edit]

The lateral geniculate nucleus (LGN) is a sensory relay nucleus in the thalamus of the brain. The LGN consists of six layers in humans and other primates starting from catarrhines, including cercopithecidae and apes. Layers 1, 4, and 6 correspond to information from the contralateral (crossed) fibers of the nasal retina (temporal visual field); layers 2, 3, and 5 correspond to information from the ipsilateral (uncrossed) fibers of the temporal retina (nasal visual field).

Layer one contains M cells, which correspond to the M (magnocellular) cells of the optic nerve of the opposite eye and are concerned with depth or motion. Layers four and six of the LGN also connect to the opposite eye, but to the P cells (color and edges) of the optic nerve. By contrast, layers two, three and five of the LGN connect to the M cells and P (parvocellular) cells of the optic nerve for the same side of the brain as its respective LGN.

Spread out, the six layers of the LGN are the area of a credit card and about three times its thickness. The LGN is rolled up into two ellipsoids about the size and shape of two small birds' eggs. In between the six layers are smaller cells that receive information from the K cells (color) in the retina. The neurons of the LGN then relay the visual image to the primary visual cortex (V1) which is located at the back of the brain (posterior end) in the occipital lobe in and close to the calcarine sulcus. The LGN is not just a simple relay station, but it is also a center for processing; it receives reciprocal input from the cortical and subcortical layers and reciprocal innervation from the visual cortex.[26]

Scheme of the optic tract with image being decomposed on the way, up to simple cortical cells (simplified)

Optic radiation

[edit]

The optic radiations, one on each side of the brain, carry information from the thalamic lateral geniculate nucleus to layer 4 of the visual cortex. The P layer neurons of the LGN relay to V1 layer 4C β. The M layer neurons relay to V1 layer 4C α. The K layer neurons in the LGN relay to large neurons called blobs in layers 2 and 3 of V1.[26]

There is a direct correspondence from an angular position in the visual field of the eye, all the way through the optic tract to a nerve position in V1 up to V4, i.e. the primary visual areas. After that, the visual pathway is roughly separated into a ventral and dorsal pathway.

Visual cortex

[edit]
Visual cortex:
V1; V2; V3; V4; V5 (also called MT)

The visual cortex is responsible for processing the visual image. It lies at the rear of the brain (highlighted in the image), above the cerebellum. The region that receives information directly from the LGN is called the primary visual cortex (also called V1 and striate cortex). It creates a bottom-up saliency map of the visual field to guide attention or eye gaze to salient visual locations.[35][clarification needed] Hence selection of visual input information by attention starts at V1[36] along the visual pathway.

Visual information then flows through a cortical hierarchy. These areas include V2, V3, V4 and area V5/MT. (The exact connectivity depends on the species of the animal.) These secondary visual areas (collectively termed the extrastriate visual cortex) process a wide variety of visual primitives. Neurons in V1 and V2 respond selectively to bars of specific orientations, or combinations of bars. These are believed to support edge and corner detection. Similarly, basic information about color and motion is processed here.[37]

Heider, et al. (2002) found that neurons involving V1, V2, and V3 can detect stereoscopic illusory contours; they found that stereoscopic stimuli subtending up to 8° can activate these neurons.[38]

Visual cortex is active even during resting state fMRI.

Visual association cortex

[edit]

As visual information passes forward through the visual hierarchy, the complexity of the neural representations increases. Whereas a V1 neuron may respond selectively to a line segment of a particular orientation in a particular retinotopic location, neurons in the lateral occipital complex respond selectively to a complete object (e.g., a figure drawing), and neurons in the visual association cortex may respond selectively to human faces, or to a particular object.

Along with this increasing complexity of neural representation may come a level of specialization of processing into two distinct pathways: the dorsal stream and the ventral stream (the Two Streams hypothesis,[39] first proposed by Ungerleider and Mishkin in 1982). The dorsal stream, commonly referred to as the "where" stream, is involved in spatial attention (covert and overt), and communicates with regions that control eye movements and hand movements. More recently, this area has been called the "how" stream to emphasize its role in guiding behaviors to spatial locations. The ventral stream, commonly referred to as the "what" stream, is involved in the recognition, identification and categorization of visual stimuli.

Intraparietal sulcus (red)

However, there is still much debate about the degree of specialization within these two pathways, since they are in fact heavily interconnected.[40]

Horace Barlow proposed the efficient coding hypothesis in 1961 as a theoretical model of sensory coding in the brain.[41] Limitations in the applicability of this theory in the primary visual cortex (V1) motivated the V1 Saliency Hypothesis that V1 creates a bottom-up saliency map to guide attention exogenously.[35] With attentional selection as a center stage, vision is seen as composed of encoding, selection, and decoding stages.[42]

The default mode network is a network of brain regions that are active when an individual is awake and at rest. The visual system's default mode can be monitored during resting state fMRI: Fox, et al. (2005) found that "the human brain is intrinsically organized into dynamic, anticorrelated functional networks",[43] in which the visual system switches from resting state to attention.

In the parietal lobe, the lateral and ventral intraparietal cortex are involved in visual attention and saccadic eye movements. These regions are in the intraparietal sulcus (marked in red in the adjacent image).

Development

[edit]

Infancy

[edit]

Newborn infants have limited color perception.[44] One study found that 74% of newborns can distinguish red, 36% green, 25% yellow, and 14% blue. After one month, performance "improved somewhat."[45] Infant's eyes do not have the ability to accommodate. Pediatricians are able to perform non-verbal testing to assess visual acuity of a newborn, detect nearsightedness and astigmatism, and evaluate the eye teaming and alignment. Visual acuity improves from about 20/400 at birth to approximately 20/25 at 6 months of age. This happens because the nerve cells in the retina and brain that control vision are not fully developed.

Childhood and adolescence

[edit]

Depth perception, focus, tracking and other aspects of vision continue to develop throughout early and middle childhood. From recent studies in the United States and Australia there is some evidence that the amount of time school aged children spend outdoors, in natural light, may have some impact on whether they develop myopia. The condition tends to get somewhat worse through childhood and adolescence, but stabilizes in adulthood. More prominent myopia (nearsightedness) and astigmatism are thought to be inherited. Children with this condition may need to wear glasses.

Adulthood

[edit]

Vision is often one of the first senses affected by aging. A number of changes occur with aging:

  • Over time, the lens becomes yellowed and may eventually become brown, a condition known as brunescence or brunescent cataract. Although many factors contribute to yellowing, lifetime exposure to ultraviolet light and aging are two main causes.
  • The vitreous humor naturally degenerates over time, causing it to liquefy, contract, and eventually separate from the posterior of the retina. This process leaves deposits of protein structures and other cellular waste within the vitrous humor, visualizing as eye floaters.
  • Presbyopia develops as the lens gradually loses its natural flexibility, reducing the eye's ability to focus.
  • While a healthy adult pupil typically has a size range of 2–8 mm, with age the range gets smaller, trending towards a moderately small diameter.
  • Tear production typically declines with age. However, there are a number of age-related conditions that can cause excessive tearing.

Other functions

[edit]

Balance

[edit]

Along with proprioception and vestibular function, the visual system plays an important role in the ability of an individual to control balance and maintain an upright posture. When these three conditions are isolated and balance is tested, it has been found that vision is the most significant contributor to balance, playing a bigger role than either of the two other intrinsic mechanisms.[46] The clarity with which an individual can see his environment, as well as the size of the visual field, the susceptibility of the individual to light and glare, and poor depth perception play important roles in providing a feedback loop to the brain on the body's movement through the environment. Anything that affects any of these variables can have a negative effect on balance and maintaining posture.[47] This effect has been seen in research involving elderly subjects when compared to young controls,[48] in glaucoma patients compared to age matched controls,[49] cataract patients pre and post surgery,[50] and even something as simple as wearing safety goggles.[51] Monocular vision (one eyed vision) has also been shown to negatively impact balance, which was seen in the previously referenced cataract and glaucoma studies,[49][50] as well as in healthy children and adults.[52]

According to Pollock et al. (2010) stroke is the main cause of specific visual impairment, most frequently visual field loss (homonymous hemianopia, a visual field defect). Nevertheless, evidence for the efficacy of cost-effective interventions aimed at these visual field defects is still inconsistent.[53]

Clinical significance

[edit]
Visual pathway lesions
From top to bottom:
1. Complete loss of vision, right eye
2. Bitemporal hemianopia
3. Homonymous hemianopsia
4. Quadrantanopia
5&6. Quadrantanopia with macular sparing

Proper function of the visual system is required for sensing, processing, and understanding the surrounding environment. Difficulty in sensing, processing and understanding light input has the potential to adversely impact an individual's ability to communicate, learn and effectively complete routine tasks on a daily basis.

In children, early diagnosis and treatment of impaired visual system function is an important factor in ensuring that key social, academic and speech/language developmental milestones are met.

Cataract is clouding of the lens, which in turn affects vision. Although it may be accompanied by yellowing, clouding and yellowing can occur separately. This is typically a result of ageing, disease, or drug use.

Presbyopia is a visual condition that causes farsightedness. The eye's lens becomes too inflexible to accommodate to normal reading distance, focus tending to remain fixed at long distance.

Glaucoma is a type of blindness that begins at the edge of the visual field and progresses inward. It may result in tunnel vision. This typically involves the outer layers of the optic nerve, sometimes as a result of buildup of fluid and excessive pressure in the eye.[54]

Scotoma is a type of blindness that produces a small blind spot in the visual field typically caused by injury in the primary visual cortex.

Homonymous hemianopia is a type of blindness that destroys one entire side of the visual field typically caused by injury in the primary visual cortex.

Quadrantanopia is a type of blindness that destroys only a part of the visual field typically caused by partial injury in the primary visual cortex. This is very similar to homonymous hemianopia, but to a lesser degree.

Prosopagnosia, or face blindness, is a brain disorder that produces an inability to recognize faces. This disorder often arises after damage to the fusiform face area.

Visual agnosia, or visual-form agnosia, is a brain disorder that produces an inability to recognize objects. This disorder often arises after damage to the ventral stream.

Other animals

[edit]

Different species are able to see different parts of the light spectrum; for example, bees can see into the ultraviolet,[55] while pit vipers can accurately target prey with their pit organs, which are sensitive to infrared radiation.[56] The mantis shrimp possesses arguably the most complex visual system of any species. The eye of the mantis shrimp holds 16 color receptive cones, whereas humans only have three. The variety of cones enables them to perceive an enhanced array of colors as a mechanism for mate selection, avoidance of predators, and detection of prey.[57] Swordfish also possess an impressive visual system. The eye of a swordfish can generate heat to better cope with detecting their prey at depths of 2000 feet.[58] Certain one-celled microorganisms, the warnowiid dinoflagellates have eye-like ocelloids, with analogous structures for the lens and retina of the multi-cellular eye.[59] The armored shell of the chiton Acanthopleura granulata is also covered with hundreds of aragonite crystalline eyes, named ocelli, which can form images.[60]

Many fan worms, such as Acromegalomma interruptum which live in tubes on the sea floor of the Great Barrier Reef, have evolved compound eyes on their tentacles, which they use to detect encroaching movement. If movement is detected, the fan worms will rapidly withdraw their tentacles. Bok, et al., have discovered opsins and G proteins in the fan worm's eyes, which were previously only seen in simple ciliary photoreceptors in the brains of some invertebrates, as opposed to the rhabdomeric receptors in the eyes of most invertebrates.[61]

Only higher primate Old World (African) monkeys and apes (macaques, apes, orangutans) have the same kind of three-cone photoreceptor color vision humans have, while lower primate New World (South American) monkeys (spider monkeys, squirrel monkeys, cebus monkeys) have a two-cone photoreceptor kind of color vision.[62]

Biologists have determined that humans have extremely good vision compared to the overwhelming majority of animals, particularly in daylight, surpassed only by a few large species of predatory birds.[63][64] Other animals such as dogs are thought to rely more on senses other than vision, which in turn may be better developed than in humans.[65][66]

History

[edit]

In the second half of the 19th century, many motifs of the nervous system were identified such as the neuron doctrine and brain localization, which related to the neuron being the basic unit of the nervous system and functional localisation in the brain, respectively. These would become tenets of the fledgling neuroscience and would support further understanding of the visual system.

The notion that the cerebral cortex is divided into functionally distinct cortices now known to be responsible for capacities such as touch (somatosensory cortex), movement (motor cortex), and vision (visual cortex), was first proposed by Franz Joseph Gall in 1810.[67] Evidence for functionally distinct areas of the brain (and, specifically, of the cerebral cortex) mounted throughout the 19th century with discoveries by Paul Broca of the language center (1861), and Gustav Fritsch and Eduard Hitzig of the motor cortex (1871).[67][68] Based on selective damage to parts of the brain and the functional effects of the resulting lesions, David Ferrier proposed that visual function was localized to the parietal lobe of the brain in 1876.[68] In 1881, Hermann Munk more accurately located vision in the occipital lobe, where the primary visual cortex is now known to be.[68]

In 2014, a textbook "Understanding vision: theory, models, and data" [42] illustrates how to link neurobiological data and visual behavior/psychological data through theoretical principles and computational models.

See also

[edit]

References

[edit]

Further reading

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
The visual system is the sensory apparatus responsible for detecting and processing to enable vision, encompassing the eyes, neural pathways, and regions that convert into perceptual experiences such as color, form, motion, and depth. It begins with photoreception in the and extends through interconnected neural structures to higher cortical areas, allowing organisms to interpret their environment. In humans, the visual system relies on the eye as its primary organ, where light enters through the and is focused by the crystalline lens onto the —a thin, multilayered neural tissue lining the back of the eye. The contains approximately 120 million rod photoreceptors for low-light ( and 6 million photoreceptors for high-acuity, color (, with cones concentrated in the for sharp central vision. Phototransduction occurs when light activates photopigments like in rods or opsins in (sensitive to long-, medium-, and short-wavelength light for , , and , respectively), triggering a cascade that hyperpolarizes photoreceptors and modulates release to bipolar and cells. cells, numbering about 1 million per eye, integrate signals from the and form axons that constitute the , transmitting action potentials to the . The afferent visual pathway routes these signals from each through the —where nasal fibers cross to the contralateral side—into the optic tracts, synapsing in the (LGN) of the , a six-layered structure that relays information via the optic radiations to the primary (V1, 17) in the . The LGN organizes inputs into magnocellular (M) layers for motion and depth processing and parvocellular (P) layers for color and fine detail, preserving retinotopic mapping throughout the pathway. Beyond V1, visual information disperses to extrastriate areas like V2, V4 (color processing), and V5/MT (motion detection), enabling complex functions such as , for depth, and attentional modulation. Accessory components include efferent pathways for and pupillary reflexes: the oculomotor (III), trochlear (IV), and abducens (VI) coordinate conjugate gaze, while parasympathetic fibers from the Edinger-Westphal nucleus constrict pupils in response to light via the optic nerve's afferent limb. Sympathetic innervation dilates pupils for low-light adaptation. Disruptions in this system, such as lesions at the chiasm causing bitemporal hemianopia, underscore its precision, with the overall architecture supporting trichromatic evolved from ancestral around 30 million years ago.

Overview

Optical principles

The visual system begins with the optical components of the eye, which refract incoming light to form a focused on the . Light rays from an object enter the eye primarily through the , the transparent anterior surface that accounts for approximately two-thirds of the eye's total refractive power, about 43 diopters (D) in the relaxed state. The rays then pass through the aqueous humor, a clear fluid in the anterior chamber between the and the lens, which maintains and contributes minimally to while nourishing avascular tissues like the . Next, the crystalline lens, with a relaxed power of around 20 D, further bends the light, followed by transmission through the vitreous humor, a gel-like substance filling the posterior chamber that helps maintain the eye's shape and allows undistorted passage of light to the . Together, these elements create an optical system with a total power of about 60 D and a of roughly 17 mm, projecting an inverted, onto the in the emmetropic (normal) eye, where parallel rays from distant objects converge precisely on the fovea. To focus on objects at varying distances, the eye employs accommodation, the dynamic adjustment of the lens's curvature via contraction of the ciliary muscle, which relaxes the zonular fibers and allows the lens to thicken and increase its power by 8–12 D. In young adults, this process enables a near point of accommodation at about 10 cm, corresponding to approximately 10 D of accommodative amplitude, allowing clear vision from infinity to this minimum distance. As aging progresses, the lens loses elasticity, leading to presbyopia, which typically begins around age 40 and reduces accommodative amplitude to less than 2 D by age 50, shifting the near point farther away and necessitating corrective lenses for close work. The eye's optics are not perfect and exhibit aberrations that degrade image quality. Spherical aberration occurs because peripheral rays refract more strongly than central rays, with the cornea contributing positive aberration and the lens negative, partially balancing in youth but worsening with age. Chromatic aberration arises from the lens and cornea refracting shorter wavelengths (e.g., blue light) more than longer ones (e.g., red), causing color fringing; the eye compensates partly through the fovea's insensitivity to blue and by constricting the pupil, which reduces the effective aperture and minimizes both spherical and chromatic aberrations by limiting peripheral ray entry. Deviations from result in refractive errors, quantified in diopters as the reciprocal of the in meters (D=1fD = \frac{1}{f}, where ff is in meters). In (nearsightedness), the eyeball is elongated or the refractive power excessive, causing distant rays to focus anterior to the ; correction requires a concave lens with negative power (e.g., -3 D for a of -0.33 m). () involves a shorter eyeball or insufficient power, focusing rays posterior to the , corrected by a convex positive lens (e.g., +2 D). stems from irregular curvature of the or lens, creating different refractive powers in principal meridians (e.g., +1.00 D sphere and -0.50 D cylinder at 90°), blurring images in specific orientations unless compensated by cylindrical lenses.

Neural organization

The visual system's neural organization exhibits a hierarchical structure that processes visual information from the through successive stages to the . At the retinal level, photoreceptors (rods and cones) detect light and transmit signals via bipolar and amacrine cells to retinal cells, whose axons form the . These fibers converge at the , where nasal retinal fibers from each eye cross to the contralateral side, ensuring that information is bilaterally represented in the . Post-chiasm, the optic tract carries these signals primarily to the (LGN) of the , a key relay station organized into six layers that maintain retinotopic mapping of the . From the LGN, geniculocalcarine radiations project to the primary (V1, or striate cortex) in the , where initial feature extraction occurs, before signals diverge to higher extrastriate areas for advanced processing. The retinogeniculate pathway is divided into parallel magnocellular (M) and parvocellular (P) streams, which segregate early in the retina and are preserved through the LGN to V1. The M pathway, originating from large parasol ganglion cells, conveys low-spatial-resolution, high-contrast, motion-sensitive signals through the ventral LGN layers (1 and 2), supporting detection of fast-moving or low-contrast stimuli. In contrast, the P pathway, from smaller midget ganglion cells, transmits high-spatial-resolution, color-opponent information via the dorsal LGN layers (3-6), enabling fine detail and chromatic discrimination. These streams partially converge in V1 but maintain functional separation, with additional koniocellular (K) projections handling blue-yellow color signals through the LGN's interlaminar regions. Extrastriate projections from V1 fan out to secondary areas like V2, V3, and beyond, integrating inputs for complex visual analysis. Beyond V1, the visual system employs two major parallel processing streams: the ventral "what" pathway and the dorsal "where" pathway. The ventral stream extends from V1 through the occipitotemporal cortex to the inferotemporal lobe, specializing in , form , and visual identification by analyzing invariant features like and color. The dorsal stream, projecting from V1 to the parietal cortex, focuses on spatial localization, motion guidance, and visuomotor coordination, facilitating actions such as reaching and grasping by computing egocentric representations. These streams originate post-V1 and interact bidirectionally, allowing flexible integration of perceptual and action-oriented processing. A hallmark of this organization is neural convergence, which enhances signal efficiency and properties. The human contains approximately 126 million photoreceptors that onto about 1 million cells, resulting in a convergence ratio of roughly 126:1 and reducing while amplifying sensitivity in . cell are circular and center-surround organized, with M cells featuring large fields for broad motion detection and P cells having smaller fields for precise acuity; this structure is preserved retinotopically through the LGN and V1, enabling a compressed yet spatially mapped representation of the visual world.

Anatomy

Eye structure

The human eye is a roughly spherical organ with an axial length of approximately 24 mm in adults, consisting of external protective layers and internal chambers that maintain its structure and optical function. External structures protect and interface with the environment. The eyelids are movable folds of skin and muscle that close reflexively to shield the eye from injury, distribute tears across the surface, and prevent corneal drying. The conjunctiva is a thin, vascularized mucous membrane that lines the inner surfaces of the eyelids (palpebral conjunctiva) and covers the anterior sclera (bulbar conjunctiva), producing mucus to lubricate the ocular surface and serving as a barrier against pathogens. The sclera forms the opaque, fibrous outer coat of the eye, comprising tough collagen fibers that provide structural integrity and rigidity to the posterior five-sixths of the globe, extending from the cornea to the optic nerve. Anteriorly, the sclera transitions into the cornea, a transparent, avascular dome-shaped structure with a horizontal diameter of about 11.7 mm in adults, which contributes significantly to the eye's refractive power by bending incoming light rays. Internally, the eye is divided into fluid-filled chambers that support its shape and nourishment. The anterior chamber lies between the and the iris, while the posterior chamber is a narrow space between the iris and the lens; both are filled with aqueous humor, a clear fluid produced by the processes to maintain and provide nutrients to avascular tissues like the and lens. The larger posterior chamber, or vitreous chamber, occupies the space between the lens and the , filled with the gel-like vitreous humor that helps maintain the eye's spherical form and transmits light. The lens is a biconvex, transparent structure suspended within the posterior chamber by zonular fibers (zonules of Zinn) that anchor it to the , allowing for accommodation. It consists primarily of elongated fiber cells filled with high concentrations of soluble proteins, which ensure optical clarity and gradation for focusing light. Vascular supply to the eye arises mainly from branches of the . The , a highly vascularized layer between the and , nourishes the outer eye structures through its dense capillary network supplied by short posterior ciliary arteries. The inner and optic nerve head receive blood via the central retinal artery and corresponding vein, which enter and exit through the .

Retinal layers and cells

The is a multilayered neural tissue lining the posterior inner surface of the eye, consisting of ten distinct layers that facilitate the conversion of into neural signals. These layers, organized from the innermost (vitreous-facing) to the outermost, include: the inner limiting membrane, a thin formed by the footplates of Müller glial cells that separates the from the vitreous humor; the fiber layer, comprising unmyelinated axons of retinal ganglion cells that converge to form the ; the ganglion cell layer, containing the cell bodies of these ganglion cells; the inner plexiform layer, where synapses occur between bipolar and processes and ganglion cell dendrites; the inner nuclear layer, housing the nuclei of bipolar cells, horizontal cells, and ; the outer plexiform layer, the site of ribbon synapses between photoreceptor terminals and the dendrites of bipolar and horizontal cells; the outer nuclear layer, consisting of the nuclei of rod and photoreceptors; the external limiting membrane, a fenestrated layer of adherens junctions between Müller cell processes and photoreceptors; the photoreceptor layer, including the inner and outer segments of rods and cones where phototransduction occurs; and the , a single layer of cuboidal cells that absorbs , recycles photopigments, and forms part of the blood- barrier. Photoreceptor cells are the primary light-detecting elements, located in the photoreceptor layer, with and cones differing in distribution, sensitivity, and function. , numbering approximately 120 million per , are specialized for scotopic (low-light) vision and exhibit peak sensitivity at 498 nm, enabling detection in dim conditions but without color discrimination. Cones, totaling about 6 million, support photopic (bright-light) vision and color , with three subtypes: long-wavelength-sensitive (L-cones) peaking at around 564 nm for light, medium-wavelength-sensitive (M-cones) at 534 nm for , and short-wavelength-sensitive (S-cones) at 420 nm for . These photoreceptors hyperpolarize in response to , initiating . Beyond photoreceptors, the retina features several neuronal and glial cell types that process and relay visual information. Bipolar cells, numbering around 10 million, form direct synaptic connections with photoreceptors and transmit graded potentials to cells, with subtypes specialized for on/off responses or specific cone inputs. Horizontal cells provide to enhance contrast by feedback to photoreceptors and feedforward to bipolar cells. Amacrine cells, diverse in morphology and use, modulate bipolar- synapses for temporal and spatial refinement, including direction selectivity. ganglion cells, about 1 million in total, integrate inputs in the inner plexiform layer and generate action potentials that travel via their axons in the ; subtypes like and parasol cells correspond to parvocellular and magnocellular pathways. Müller glial cells span all layers, offering structural support, metabolic aid, and ion homeostasis while contributing to the inner and external limiting membranes. The , a specialized depression in the macula lutea approximately 1.5 mm in , optimizes high-acuity vision by featuring a cone-only region with minimal overlying layers, allowing direct light access to a of approximately 150,000–200,000 cones per mm² (or ~15–20 cones per 100 µm²). Its central portion includes an avascular zone about 0.5 mm in , ensuring unobstructed capture without vascular interference. Retinal blood supply is dual: the inner layers (from inner limiting to outer plexiform layer) receive oxygenated blood via the central retinal artery, a branch of the that enters through the and forms superficial and deep capillary networks; the outer layers (photoreceptors and pigment epithelium) are nourished by the choriocapillaris of the , supplied by short and , supporting high metabolic demands through diffusion.

Central visual pathways

The central visual pathways begin with the formation of the , which consists of axons from cells that converge at the and exit the eye. This nerve contains approximately 1.2 million myelinated and unmyelinated fibers in humans, transmitting visual signals from the toward the . Myelination of these axons occurs primarily after the , beginning in the optic tract and progressing retrogradely toward the chiasm during late and early postnatal development. At the , located at the base of the anterior to the , the optic nerves from both eyes partially decussate. Fibers originating from the nasal of each eye cross to the contralateral side, while temporal retinal fibers remain uncrossed, resulting in approximately 50% of fibers decussating in humans to enable . This partial crossing ensures that each receives input from the contralateral , integrating monocular signals for and . Post-chiasm, the bundled axons form the optic tracts, which carry segregated visual information to multiple targets, including the (LGN) of the , the for orienting responses, and the pretectum for pupillary light reflexes. The optic tracts maintain a retinotopic organization, preserving the spatial mapping of the from the . The LGN serves as the primary thalamic station for visual signals en route to the cortex, featuring a layered structure with six distinct laminae in . Layers 1 and 2 comprise magnocellular (M) cells, which process low-spatial-frequency, motion-sensitive information, while layers 3 through 6 contain parvocellular (P) cells specialized for high-spatial-frequency, color-opponent signals. These layers exhibit strict retinotopic organization, with upper layers representing the inferior and lower layers the superior field, and alternate for ipsilateral and contralateral inputs. Local within the LGN modulate activity, providing inhibitory feedback to refine signal transmission. Anatomical variations in these pathways can occur, notably in albinism, where reduced during development leads to excessive at the , resulting in abnormal routing of over 90% of fibers to the contralateral hemisphere and disrupted binocular representation. Such anomalies highlight the role of pigmentation in guiding axonal during embryogenesis.

Visual cortex regions

The , located primarily in the , comprises a hierarchy of specialized regions that process visual information relayed from the (LGN) of the . These regions form a retinotopic , where adjacent neurons respond to adjacent parts of the , enabling precise spatial mapping. The primary visual cortex (V1) serves as the initial cortical entry point, followed by secondary areas like V2, V3, V4, and V5/MT, which handle increasingly complex features such as form, color, and motion. Higher-order regions in the temporal and parietal lobes integrate these signals for and spatial awareness, respectively. The primary visual cortex, also known as V1 or the striate cortex, corresponds to Brodmann area 17 and is situated along the in the . It receives direct thalamocortical inputs from the LGN, primarily terminating in layer 4. V1 exhibits a precise retinotopic map of the , with disproportionate representation of the fovea—a phenomenon called cortical magnification—where central vision occupies a larger cortical area due to higher acuity demands. This magnification factor can exceed 10 times that of , underscoring V1's role in fine-grained . Adjacent to V1, area V2 processes more integrated features, including contours and simple forms, while maintaining a retinotopic organization divided into thin and thick stripes for color and disparity processing, respectively. Area V3 (or VP in some nomenclatures) extends this by handling global form, via , and coarse color information. Area V4, located anteriorly in the ventral stream, specializes in and object shape invariance, enabling perception of hues independent of illumination changes. In the dorsal stream, area V5 (also MT) is dedicated to motion processing, with neurons selectively responsive to direction and speed, crucial for tracking moving objects. Beyond these early extrastriate areas, the inferotemporal cortex in the ventral pathway supports advanced , where neurons respond to complex shapes and faces, achieving viewpoint-invariant identification through hierarchical feature integration. In contrast, parietal regions, such as the , contribute to spatial attention by modulating visual salience and directing gaze, facilitating the selection of relevant stimuli in cluttered scenes. Cytoarchitecturally, V1 is distinguished by its layered structure, particularly layer , which is subdivided into 4Ca and 4Cb. Layer 4Ca receives magnocellular (magno) inputs from the LGN, conveying low-spatial-frequency information for motion and , while 4Cb receives parvocellular (parvo) inputs for high-resolution color and form details. These segregated inputs preserve parallel processing streams from the onward. Hemispheric specialization in the reflects asymmetric processing: the left hemisphere excels at detailed, local feature analysis, such as fine textures and letters, whereas the right hemisphere prioritizes global, holistic configurations, like overall scene layout. This arises from interhemispheric differences in connectivity and sizes, influencing tasks from reading to .

Function

Phototransduction

Phototransduction is the process by which photoreceptor cells in the convert light energy into electrical signals through a series of biochemical reactions. In the outer segments of and cones, light absorption by visual pigments initiates a G-protein-coupled cascade that ultimately modulates . This mechanism enables the detection of photons across a wide range of intensities and wavelengths, with specialized for low-light vision and cones for color and high-acuity tasks. In rod photoreceptors, the visual pigment consists of the protein bound to the chromophore . Upon absorption of a , 11-cis-retinal isomerizes to all-trans-retinal, inducing a conformational change in to its active form, metarhodopsin II (R*). This activated catalyzes the exchange of GDP for GTP on the G-protein , activating approximately 20 molecules per R*. The activated then stimulates (PDE), which hydrolyzes (cGMP) to 5'-GMP, rapidly reducing cytosolic cGMP levels. In the dark, high cGMP concentrations keep cation channels open, allowing Na⁺ and Ca²⁺ influx that depolarizes the rod to approximately -40 mV. Light-induced cGMP decline closes these channels, halting the "dark current" and hyperpolarizing the membrane to about -70 mV, which decreases glutamate release at the . This amplification—one activates approximately 20 transducin molecules—enhances sensitivity. Cone photoreceptors employ similar mechanisms but with distinct photopigments called iodopsins, each comprising variants (short-, medium-, and long-wavelength sensitive) covalently linked to 11-cis-retinal derivatives from . These pigments exhibit faster response kinetics and recovery times than , enabling cones to adapt quickly to varying light levels. The and PDE cascade in cones mirrors that in , leading to comparable cGMP-gated channel closure and hyperpolarization, though with lower amplification suited for brighter conditions. Dark and light adaptation maintain phototransduction sensitivity across illumination changes. In darkness, guanylate cyclase (GC) synthesizes cGMP to reopen channels, while light reduces Ca²⁺ influx through closed channels. This Ca²⁺ decline activates guanylate cyclase-activating proteins (GCAPs), which stimulate GC to restore cGMP levels and terminate the response. Calcium feedback via GCAPs provides gain control, compressing the response range in bright light and extending it in dim conditions. Rhodopsin deactivation by phosphorylation and arrestin binding further regulates adaptation. Rods display peak around 500 nm in the range, following the absorption spectrum of , while cones peak at approximately 420 nm (S-cones), 530 nm (M-cones), and 560 nm (L-cones). The quantum efficiency of , the probability that an absorbed triggers a detectable response, is approximately 0.67 in rods, contributing to their single-photon detection capability.

Visual signal processing

Visual signal processing begins in the , where retinal ganglion cells encode spatial patterns of light into action potentials that are transmitted via the . These cells exhibit center-surround receptive fields, characterized by a central region that responds oppositely to a surrounding annular region, enabling contrast detection. For instance, ON-center ganglion cells increase firing when light stimulates the center while the surround is dark, whereas OFF-center cells respond to darkness in the center. This organization, first described in cat retina, sharpens edges by emphasizing differences in across the . Ganglion cells are classified into major types based on their properties and projections. Parvocellular (P) ganglion cells, which constitute about 90% of the population in , have small receptive fields and are sensitive to color differences through opponent processes (e.g., red-green or blue-yellow), contributing to fine spatial detail and form . In contrast, magnocellular (M) ganglion cells have larger receptive fields, respond transiently to low-contrast stimuli, and are tuned to motion and coarse changes, supporting depth and dynamic vision. These distinctions arise from inputs from distinct bipolar and amacrine cells, preserving parallel processing streams from the . Lateral inhibition, mediated by horizontal cells in the outer and amacrine cells in the inner , enhances by suppressing activity in neighboring regions. Horizontal cells provide feedback to photoreceptors and bipolar cells, reducing responses to uniform illumination and amplifying boundaries between light and dark areas. Amacrine cells similarly inhibit bipolar and ganglion cells laterally, contributing to surround antagonism and temporal sharpening of signals. This mechanism underlies the center-surround structure and improves contrast sensitivity across the . Axons from ganglion cells project to the (LGN) of the , which acts primarily as a relay station while introducing subtle modulations. LGN neurons maintain center-surround receptive fields similar to those of ganglion cells, with parvocellular layers preserving color opponency and magnocellular layers emphasizing achromatic contrast and motion. Retinotopic organization ensures precise spatial mapping, and inputs from and cortical areas provide top-down modulation for and . Weak orientation selectivity begins to emerge in some LGN cells, particularly in koniocellular layers, though it remains rudimentary compared to cortical processing. In the primary visual cortex (V1), LGN inputs drive neurons with more specialized receptive fields. Simple cells, located mainly in layer 4, respond to oriented edges or bars at specific positions within their receptive fields, exhibiting elongated excitatory and inhibitory subregions. Complex cells, found in other layers, respond to oriented stimuli across a broader area without precise positional specificity, integrating inputs from simple cells. According to the Hubel-Wiesel model, these properties arise from convergent wiring of LGN afferents aligned along columns and orientation preferences, forming the basis of feature detection. The receptive fields of simple cells resemble Gabor functions, which are Gaussian-modulated sinusoids optimal for detecting edges in natural images. Parallel processing streams are maintained through the visual pathway, with the P pathway (via parvocellular LGN) specializing in color and high-acuity form, and the M pathway (via magnocellular LGN) handling motion, depth, and low-spatial-frequency information. These streams partially converge in V1 but remain segregated into ventral (form/color) and dorsal (motion/depth) cortical pathways. Contrast sensitivity in both streams is quantified using Michelson contrast, defined as C=LmaxLminLmax+LminC = \frac{L_{\max} - L_{\min}}{L_{\max} + L_{\min}} where LmaxL_{\max} and LminL_{\min} are the maximum and minimum luminances in the stimulus; this measure highlights how M cells detect low contrasts for global scene analysis, while P cells resolve finer details. Processed signals from V1 project to higher cortical areas for further integration.

Visual perception mechanisms

Binocular vision enables depth perception through the slight differences in the images projected onto each retina, known as . This disparity provides cues for , where the visual system computes relative depths by comparing corresponding points across the two eyes. The represents the locus of points in space that project zero disparity onto corresponding retinal locations, forming a theoretical —often approximated as the Vieth-Müller —beyond which points elicit uncrossed or crossed disparities for perceived depth. Under optimal conditions, the human threshold for detecting depth via disparity is approximately 10 arcseconds, allowing fine discrimination of distances as small as a few centimeters at arm's length. Color vision arises from the combined action of receptor and neural processing mechanisms in the visual pathway. The trichromatic theory, proposed by Thomas Young and elaborated by , posits that color perception results from the relative stimulation of three types of cone photoreceptors sensitive to short (blue), medium (green), and long (red) wavelengths. This retinal stage is complemented by the , developed by Ewald Hering and quantitatively formulated by Leo Hurvich and Dorothea Jameson, which describes post-receptoral channels encoding color differences along red-green, blue-yellow, and black-white axes to account for phenomena like afterimages and color contrast. Anomalies in these systems lead to color vision deficiencies; for instance, protanopia involves the absence of functional long-wavelength cones, resulting in confusion between reds and greens due to reliance on medium- and short-wavelength signals alone. Motion perception involves resolving ambiguities in local motion signals to form coherent global representations. The aperture problem occurs when a limited restricts observation of an object's full motion trajectory, allowing multiple possible directions perpendicular to the visible edge, as neurons in early visual areas respond only to the component normal to their orientation tuning. Optic flow patterns, first conceptualized by James J. Gibson, describe the radial expansion or contraction of visual motion during self-movement, providing cues for heading direction and environmental layout through the focus of expansion where motion vectors converge. In the middle temporal (MT) area of the , neurons exhibit robust direction selectivity for complex stimuli, integrating inputs from primary to disambiguate local motions and support of object trajectories and speeds. Visual illusions reveal how perceptual mechanisms can misinterpret sensory inputs, often due to incomplete or conflicting cues. The , where lines flanked by inward- or outward-pointing arrows appear unequal in length despite being identical, arises from the visual system's probabilistic inference of depth from angular cues, biasing length estimation as if viewing corners in a three-dimensional scene. Similarly, the , described by , induces the perception of smooth motion from sequentially flashing stationary lights, driven by low-level temporal integration in early visual areas that fills spatial gaps to create apparent continuity. These effects are modulated by feature binding, where attention links disparate attributes like position and motion into unified objects, and top-down influences from higher cortical areas, which impose expectations to resolve ambiguities in ambiguous scenes. Gestalt principles describe innate organizational rules that the visual system uses to segment and interpret complex scenes into meaningful wholes. The principle of proximity groups elements based on spatial nearness, such that dots clustered closely are perceived as forming patterns or objects separate from more distant ones, facilitating scene parsing without explicit computation. Similarity promotes grouping of elements sharing attributes like color, shape, or orientation, overriding minor positional differences to bind features into coherent entities, as seen in camouflage breakdown when uniform patterns disrupt matches. Closure completes incomplete contours into enclosed shapes, with the brain inferring missing segments to perceive a whole figure, such as recognizing a circle from a partial arc, enhancing object recognition amid clutter. These principles operate primarily in early to mid-level visual processing, aiding efficient segmentation before integration in higher cortical areas.

Development

Embryonic formation

The embryonic development of the visual system begins in the third week of , when optic grooves appear in the ventral neural folds of the developing , marking the initial site of eye formation. These grooves rapidly evaginate outward as optic vesicles, which consist of and protrude laterally from the . The optic vesicles induce the overlying surface to thicken into a lens placode, while the proximal portion of each vesicle remains connected to the brain via the optic stalk, which will later develop into the . This evagination process is regulated by genes such as Rx, which is essential for optic vesicle formation and retinal differentiation. By the fourth week, the optic vesicle invaginates to form the double-layered optic cup, where the inner layer differentiates into the and the outer layer into the . The optic stalk narrows and becomes the optic nerve precursor, facilitating axonal outgrowth from retinal ganglion cells. Concurrently, the lens placode, induced by signals from the optic vesicle including the , deepens into a lens pit and detaches to form the lens vesicle, which fills with elongating primary . Mutations in , a master regulator of , disrupt lens placode induction and can lead to congenital cataracts due to impaired lens fiber differentiation. The ventral optic cup also develops the choroid fissure, a groove through which the hyaloid artery enters to nourish the lens and . Retinal lamination proceeds in an inside-out manner starting around week 5, with progenitor cells in the inner neuroblastic layer generating neurons in a sequential order: ganglion cells differentiate first, followed by amacrine, horizontal, and photoreceptor cells, while rod photoreceptors and bipolar cells emerge later. By week 20 of gestation, the major retinal layers are established, including the ganglion cell layer, inner plexiform layer, inner nuclear layer, outer plexiform layer, outer nuclear layer containing photoreceptors, and the , though synaptic refinement continues. Vascularization of the eye involves the hyaloid artery, which supplies the avascular lens and inner during early development; the begins forming around weeks 6-7 from surrounding the optic cup, developing capillaries by week 12 and mature vessels by week 22. The hyaloid artery subsequently regresses postnatally, leaving remnants like the of the . Critical periods for visual system occur between weeks 3 and 8, during which teratogens can disrupt key processes such as optic fissure closure. The choroid fissure must fuse by week 7 to seal the ventral optic cup; failure leads to , a gap in ocular structures like the , , or iris. exposure during this window interferes with fissure closure, resulting in and other ocular defects in affected embryos, highlighting the vulnerability of these early developmental stages. Genes like Pax2 are crucial for proper fissure closure and optic stalk development.

Postnatal maturation

At birth, infants exhibit limited visual capabilities, with acuity estimated at approximately 20/400, allowing them to detect only large, high-contrast features at close range. Newborns show a for high-contrast edges and patterns, which guide their initial visual exploration and support early perceptual learning. Foveal development, including pit formation and cone specialization, progresses rapidly postnatally, reaching significant maturity by around 6 months of age, thereby improving central acuity and fixation stability. The visual system undergoes critical periods of heightened plasticity during early infancy, particularly in the primary visual cortex (V1), where ocular dominance columns segregate inputs from each eye between 3 and 8 weeks in animal models like cats, establishing binocular organization. In humans, this sensitivity window extends into the first few years; disruptions such as eye misalignment () during this period can lead to , or "lazy eye," due to competitive imbalances in cortical representation. Myelination of the optic pathways, which enhances signal conduction speed, begins prenatally but continues postnatally, with completion in the optic nerve and radiations typically by age 2 years, coinciding with refinements in visual processing efficiency. Synaptic pruning in the visual cortex refines neural circuits by eliminating excess connections, with synaptic density peaking around 8 months to 2 years before progressive elimination, stabilizing toward adult levels by puberty to optimize visual function. This process, driven by activity-dependent mechanisms, sharpens receptive fields and enhances feature selectivity in higher visual areas. During , surges in sex hormones such as and testosterone modulate visual processing, with evidence indicating influences on color discrimination abilities, where females often exhibit superior performance potentially linked to estrogen's role in cortical plasticity. These hormonal changes contribute to subtle sex differences in that emerge or consolidate in .

Clinical aspects

Common disorders

The visual system is susceptible to a range of disorders that impair various stages of visual processing, from refractive errors affecting light focus to degenerative conditions damaging neural components. Globally, at least 2.2 billion experience near or distance vision impairment, with many cases linked to preventable or treatable conditions such as uncorrected refractive errors and cataracts. Cataracts involve opacification of the eye's crystalline lens, leading to progressive blurring, glare, and reduced , and represent the leading cause of blindness worldwide, affecting over 94 million people with moderate or worse vision impairment as of 2020. Primarily age-related, they result from and , though congenital, traumatic, or secondary forms (e.g., from or steroids) also occur; surgical removal with implantation restores vision in most cases and is highly effective when accessible. Refractive errors, including , hyperopia, and , occur when the eye's shape prevents proper light focusing on the , leading to . , or nearsightedness, affects approximately 30% of the global population currently, with projections estimating nearly 50% by 2050 due to increasing driven by factors like prolonged near work and reduced outdoor time. This rise is particularly evident in urbanized regions with high educational demands, where extended reading or screen use correlates with axial eye elongation and myopia progression. Diabetic retinopathy (DR) arises in individuals with mellitus due to microvascular damage from chronic , leading to retinal hemorrhages, exudates, neovascularization, and that impair vision. Globally, DR affects about 22% of people with , contributing to over 1 million cases of blindness and 3 million of moderate-to-severe as of 2020, with prevalence rising alongside the epidemic. Proliferative DR can cause vitreous hemorrhage or , while non-proliferative forms progress variably; early screening and glycemic control are key to prevention. Glaucoma encompasses a group of disorders characterized by progressive damage to the , often resulting from elevated exceeding 21 mmHg, which compresses nerve fibers and leads to loss. While primary open-angle glaucoma is the most common form, affecting aqueous humor drainage, the condition can also arise from normal in susceptible individuals, ultimately causing irreversible defects if undetected. Retinal diseases significantly impact photoreceptor function and central vision. Age-related macular degeneration (AMD) is a leading cause of vision loss in older adults, manifesting in two forms: dry AMD, involving gradual atrophy of the and , and wet AMD, where (VEGF) promotes abnormal , leading to fluid leakage and rapid central vision deterioration. (RP) represents a heterogeneous group of inherited disorders primarily affecting rod photoreceptors, causing initial night blindness and loss due to genetic mutations that trigger progressive rod degeneration followed by secondary cone death. Cortical disorders arise from damage to higher visual processing areas in the . Homonymous hemianopia results from lesions in the optic tract, , or , often due to , causing loss of the same half of the in both eyes and impairing spatial . , or cerebral motion blindness, is a rare condition stemming from bilateral lesions in the middle temporal (MT) area, disrupting the of moving objects while preserving static vision, as documented in cases of or hypoxic injury. Color vision deficiencies impair the discrimination of hues due to cone photoreceptor anomalies. Red-green color blindness, the most prevalent form, affects about 8% of males due to X-linked recessive mutations in opsin genes on the , resulting in altered . , a rarer autosomal recessive disorder, involves complete or near-total loss of color perception from absent or dysfunctional s, often accompanied by reduced and .

Diagnostic methods

Diagnostic methods for assessing the integrity of the visual system encompass a range of techniques that evaluate acuity, visual fields, structural integrity, and electrophysiological responses, enabling clinicians to identify impairments from retinal to cortical levels. These methods are essential for detecting conditions such as , , and optic neuropathies, with selections based on suspected pathology. Visual acuity tests measure the clarity of central vision by determining the smallest letters or symbols a can resolve at a standardized . The , introduced as a clinical standard, uses rows of letters decreasing in size, where 20/20 vision indicates the ability to resolve details subtending 1 arcminute of , equivalent to normal resolution at 20 feet. For greater precision, especially in research and low-vision assessments, the employs a with evenly spaced letter sizes and consistent spacing, offering advantages in repeatability and sensitivity to small changes in acuity over traditional Snellen testing. These tests are typically conducted monocularly with refractive correction to isolate central and optic pathway function. Visual field testing, or perimetry, maps the extent and sensitivity of to detect defects like scotomas, which are blind spots arising from localized damage. Goldmann perimetry, a kinetic manual method, uses a moving stimulus to delineate field boundaries and is particularly useful for patients with low vision or unreliable fixation, providing qualitative isopters for overall field shape. In contrast, automated static perimetry, such as the Humphrey Field Analyzer, presents fixed-intensity stimuli at predefined locations to quantify threshold sensitivities, offering higher and quantitative data for monitoring progressive losses, though it may miss subtle peripheral defects compared to kinetic approaches. Imaging techniques provide detailed structural evaluation of the visual pathway. (OCT) delivers non-invasive, high-resolution cross-sectional images of retinal layers, achieving axial resolutions of 5-10 μm to quantify thicknesses of the nerve fiber layer and detect early thinning indicative of axonal loss. captures wide-field color images of the , , and vasculature, facilitating documentation of abnormalities like hemorrhages or for longitudinal comparison. For posterior pathway assessment, (MRI) excels in visualizing the , chiasm, and tracts, with enhancement highlighting or compression, though it is less sensitive for subtle retinal changes. Electrophysiological tests objectively measure neural responses along the visual pathway. Visual evoked potentials (VEP) record cortical responses to patterned stimuli via scalp electrodes, with the P100 component—a positive peak around 100 ms post-stimulus—reflecting conduction time from to ; delays beyond 115-120 ms suggest demyelination or axonal damage. (ERG) assesses retinal function by detecting electrical potentials from photoreceptors and bipolar cells in response to full-field flashes, with standardized protocols like the ISCEV guidelines distinguishing rod versus cone contributions to diagnose widespread retinal dysfunction. Emerging advancements in the 2020s incorporate (AI) for automated analysis of fundus images, enhancing early detection of age-related macular degeneration () by identifying subtle or pigment changes with sensitivities exceeding 90% in validation studies, thus supporting triage in screening programs. These AI tools, often based on convolutional neural networks, integrate with OCT and to predict progression risks, thereby aiding identification of disorders like before symptomatic vision loss.

Comparative aspects

Invertebrate visual systems

Invertebrate visual systems exhibit remarkable diversity, ranging from simple photoreceptive structures to complex organs that rival capabilities in specific functions, such as motion detection and polarization sensitivity. Unlike the centralized camera-type eyes of , many rely on distributed arrays or specialized detectors adapted to their ecological niches, enabling behaviors like rapid flight navigation or underwater predation. A prominent example is the compound eye found in insects and crustaceans, composed of numerous repeating units called ommatidia that collectively form a mosaic image. In the fruit fly Drosophila melanogaster, each compound eye contains approximately 800 ommatidia, each functioning as an independent optical unit with a corneal lens, crystalline cone, and photoreceptor cluster. These eyes operate via two primary optical mechanisms: apposition optics, where screening pigment isolates light to individual ommatidia for high spatial resolution in bright conditions, and superposition optics, which allows overlapping light paths from multiple ommatidia to enhance sensitivity in dim light. Compound eyes provide exceptional temporal resolution, with flicker fusion rates exceeding 200 Hz in some insects, facilitating precise motion detection during high-speed activities like flight. In addition to compound eyes, many insects possess ocelli—simple, non-imaging photoreceptors that detect light intensity and direction. In bees, such as bumblebees (Bombus spp.), the three dorsal ocelli serve primarily as light sensors for sky polarization navigation, particularly in low-light conditions like dusk, where they help maintain stable flight orientation by processing polarized skylight patterns. Cephalopods, including octopuses and cuttlefish, possess camera-like eyes that convergently resemble those of vertebrates but with key structural differences, such as an everted retina where photoreceptors face the incoming light, mirroring the vertebrate inverted configuration in reverse. These eyes feature a dynamic pupil; in cuttlefish (Sepia officinalis), it adopts a W-shaped form in bright light, which projects a blurred pattern onto the retina to balance vertically uneven illumination from above-water sources, enhancing contrast in shallow aquatic environments. At the cellular level, invertebrate photoreceptors typically employ rhabdomeric opsins embedded in microvillar membranes, contrasting with the ciliary opsins in ciliated photoreceptors of vertebrates; this distinction reflects ancient evolutionary divergence in phototransduction pathways. These anatomical features underpin specialized behavioral adaptations, such as (UV) vision in bees, where trichromatic photoreceptors sensitive to 300–650 nm wavelengths allow detection of UV-reflective nectar guides on flowers, guiding foraging efficiency. Similarly, mantis shrimps (Stomatopoda) exhibit advanced polarization vision through up to 16 spectral channels in their compound eyes, enabling dynamic processing of linear and circular polarized light for prey detection and intraspecific signaling in complex underwater scenes. This diversity highlights with vertebrate systems, particularly in cephalopod camera eyes, yet underscores unique invertebrate solutions to visual challenges.

Vertebrate variations

The visual systems of vertebrates exhibit diverse adaptations shaped by ecological niches, ranging from aquatic environments to aerial and terrestrial habitats. In aquatic vertebrates like fish, the cornea is notably large and contributes to enhanced light gathering, compensating for the minimal refractive power it provides underwater due to the similar refractive indices of corneal tissue and water; instead, focusing relies primarily on a spherical lens that can shift position for accommodation. Many fish species also possess tetrachromatic vision, incorporating ultraviolet-sensitive cones alongside red-, green-, and blue-sensitive ones, which enables detection of UV-reflecting patterns for communication and foraging in underwater light spectra. Avian visual systems are characterized by , with four cone types including a UV-sensitive variant, allowing birds to perceive a broader color than trichromatic mammals. These cones contain colored oil droplets that act as spectral filters, sharpening color discrimination by reducing and enhancing contrast in bright daylight environments. Additionally, birds feature a unique , a vascular structure projecting into the vitreous humor that nourishes the avascular and may stabilize during head movements. Nocturnal mammals have evolved structures to maximize sensitivity in low-light conditions, including the , a reflective layer behind the that recycles unabsorbed light to increase photon detection efficiency. Their are dominated by over cones; for example, domestic cats exhibit a of approximately 95:5, prioritizing for hunting at dusk or night while sacrificing color acuity. Primate visual evolution reflects adaptations for diurnal frugivory and arboreal life, with primates () achieving routine through duplication of the long-wavelength-sensitive (LWS) gene on the , enabling separate medium- and long-wavelength cones for red-green color discrimination centered in the fovea. In contrast, (Platyrrhini) exhibit polymorphic vision: males and homozygous females are dichromats relying on short- and medium-wavelength opsins, while heterozygous females achieve via allelic variation in a single locus, a mechanism that arose independently after the divergence from lineages around 40 million years ago.

Historical perspectives

Early discoveries

The earliest insights into the visual system emerged in , where philosophers and early anatomists began to conceptualize the eye's role in . Around the 5th century BCE, proposed that the eye served as a pathway for light and sensory impressions to reach the through channels known as poroi, marking one of the first attempts to link vision to internal anatomy rather than external emanations from the eye. This idea was advanced in the 2nd century CE by the Roman physician , who provided a detailed description of the as a conduit for visual "spirits" or , emphasizing its role in transmitting sensory information from the eye to the while integrating it into his broader theory of extramission, where visual rays emanated from the eye. During the , anatomical studies benefited from direct dissections and improved illustrations, leading to more accurate depictions of ocular structures. In 1543, published De humani corporis fabrica, featuring precise woodcut illustrations of the eye's layers, including the , , , and , which challenged Galenic errors and established a foundation for modern ocular anatomy through empirical observation. Building on this, in 1619, Jesuit Christoph Scheiner demonstrated the inversion of the retinal image in his treatise Oculus hoc est: Fundamentum opticum, using pinhole experiments on animal eyes to show that rays cross at the lens, forming an upside-down image on the retina—a key physiological insight that aligned the eye with optical principles. The 19th century saw further physiological explorations, often leveraging emerging technologies like the microscope to reveal entoptic phenomena and cellular details. In 1825, Czech physiologist Jan Evangelista Purkinje documented entoptic images, such as the shadows of retinal blood vessels visible against bright light (now known as the Purkinje tree), providing early evidence of internal ocular structures influencing perception without external aids. Hermann von Helmholtz, in his Handbuch der physiologischen Optik during the 1850s, modeled the eye as a camera obscura, with the lens focusing light onto the retina to form an image, integrating optics and physiology to explain accommodation and refraction. Theoretical advancements in color vision also took shape in the 1800s. Thomas Young proposed the trichromatic theory in 1802, suggesting three distinct retinal receptors sensitive to red, green, and blue-violet wavelengths, later refined by Helmholtz in the 1850s through quantitative analyses of . In the 1870s, Ewald Hering introduced the , positing paired color channels (red-green, blue-yellow, and black-white) that explained phenomena like afterimages, challenging and complementing the trichromatic model. The advent of the compound in the mid-19th century enabled histological breakthroughs, such as those by in the 1850s, who identified distinct retinal cell types—including , cones, and supporting elements—through detailed examinations of fixed tissues, laying groundwork for understanding photoreceptor diversity.

Modern advancements

In the mid-20th century, and Torsten N. Wiesel's groundbreaking electrophysiological studies in the 1960s revealed the functional organization of the primary (V1), demonstrating that neurons exhibit orientation selectivity, responding preferentially to lines or edges at specific angles. Their work established the hierarchical processing of visual information, where simple cells detect oriented edges and complex cells integrate motion and position, laying the foundation for understanding cortical feature detection; this research earned them the in Physiology or Medicine in 1981 alongside Roger Sperry. Advancements in during the 1990s enabled non-invasive mapping of visual areas in s, with (fMRI) allowing researchers to delineate retinotopic organization—the orderly representation of the on the cortical surface—in areas V1 through V4. B. H. Tootell and colleagues' 1995 study used fMRI to precisely identify borders of multiple visual areas by measuring responses to visual stimuli, confirming the retinotopic maps previously observed in animals and extending these findings to awake human subjects without invasive procedures. Complementary techniques like (EEG) further supported these mappings by capturing temporal dynamics of visual processing, enhancing the spatial precision of fMRI. Optogenetics emerged in the early as a transformative tool for manipulating neural activity with light, pioneered by Edward S. Boyden and colleagues in 2005, who introduced channelrhodopsin-2—a light-sensitive from —into mammalian neurons to achieve millisecond-precision control of spiking and synaptic transmission. This technique has been applied to vision restoration, enabling targeted activation of surviving retinal ganglion cells in degenerative diseases like . Building on this, retinal prostheses such as the Argus II system, approved by the U.S. in 2013 for humanitarian use but whose manufacturer ceased support in 2020 and filed for bankruptcy in 2022, delivered electrical stimulation to the via an epiretinal array, restoring basic light perception and in profoundly blind patients with severe . Subsequent optogenetic therapies, such as GS030, have advanced to clinical stages, with phase 1/2 data reported in 2023 showing restored light perception in patients with . Genetic research in the late 1990s identified in the cone-rod homeobox (CRX) gene as a cause of cone-rod , a progressive retinal disorder affecting both cone and rod photoreceptors, with Charles L. Freund and colleagues linking CRX variants to impaired photoreceptor and differentiation in 1997. More recently, -Cas9 gene editing has advanced toward clinical application for (LCA), a severe form of inherited blindness; the EDIT-101 trial (BRILLIANCE), initiated in 2020, directly injects components into the eye to disrupt a pathogenic in the CEP290 gene, marking the first human use of this technology for retinal disease and, as reported in 2024, demonstrating safety with vision and quality-of-life improvements in 79% of treated participants (11 of 14) in at least one outcome measure. In the 2010s, computational models using deep neural networks (DNNs) provided insights into visual processing by mimicking hierarchical representations in cortical areas V1 to V4, with Daniel L. K. Yamins and colleagues developing performance-optimized DNNs in 2014 that predicted neural responses in the inferior temporal cortex with accuracy rivaling biological systems during tasks. These models, trained on natural images, replicate orientation selectivity in early layers analogous to V1 and invariant object features in later layers akin to V4, bridging and to test hypotheses about visual computation.

References

Add your contribution
Related Hubs
User Avatar
No comments yet.