Hubbry Logo
Language processing in the brainLanguage processing in the brainMain
Open search
Language processing in the brain
Community hub
Language processing in the brain
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Language processing in the brain
Language processing in the brain
from Wikipedia
Dual stream connectivity between the auditory cortex and frontal lobe of monkeys and humans. Top: The auditory cortex of the monkey (left) and human (right) is schematically depicted on the supratemporal plane and observed from above (with the parieto- frontal operculi removed). Bottom: The brain of the monkey (left) and human (right) is schematically depicted and displayed from the side. Orange frames mark the region of the auditory cortex, which is displayed in the top sub-figures. Top and Bottom: Blue colors mark regions affiliated with the ADS, and red colors mark regions affiliated with the AVS (dark red and blue regions mark the primary auditory fields). Material was copied from this source, which is available under a Creative Commons Attribution 4.0 International License.

In psycholinguistics, language processing refers to the way humans use words to communicate ideas and feelings, and how such communications are processed and understood. Language processing is considered to be a uniquely human ability that is not produced with the same grammatical understanding or systematicity in even human's closest primate relatives.[1]

Throughout the 20th century the dominant model[2] for language processing in the brain was the Geschwind–Lichteim–Wernicke model, which is based primarily on the analysis of brain-damaged patients. However, due to improvements in intra-cortical electrophysiological recordings of monkey and human brains, as well non-invasive techniques such as fMRI, PET, MEG and EEG, an auditory pathway consisting of two parts[3][4] has been revealed and a two-streams model has been developed. In accordance with this model, there are two pathways that connect the auditory cortex to the frontal lobe, each pathway accounting for different linguistic roles. The auditory ventral stream pathway is responsible for sound recognition, and is accordingly known as the auditory 'what' pathway. The auditory dorsal stream in both humans and non-human primates is responsible for sound localization, and is accordingly known as the auditory 'where' pathway. In humans, this pathway (especially in the left hemisphere) is also responsible for speech production, speech repetition, lip-reading, and phonological working memory and long-term memory. In accordance with the 'from where to what' model of language evolution,[5][6] the reason the ADS is characterized with such a broad range of functions is that each indicates a different stage in language evolution.

The division of the two streams first occurs in the auditory nerve where the anterior branch enters the anterior cochlear nucleus in the brainstem which gives rise to the auditory ventral stream. The posterior branch enters the dorsal and posteroventral cochlear nucleus to give rise to the auditory dorsal stream.[7]: 8 

Language processing can also occur in relation to signed languages or written content.

Early neurolinguistics models

[edit]
Language Areas of the human brain. The angular gyrus is represented in orange, supramarginal gyrus is represented in yellow, Broca's area is represented in blue, Wernicke's area is represented in green and the primary auditory cortex is represented in pink.

Throughout the 20th century, our knowledge of language processing in the brain was dominated by the Wernicke–Lichtheim–Geschwind model.[8][2][9] The Wernicke–Lichtheim–Geschwind model is primarily based on research conducted on brain-damaged individuals who were reported to possess a variety of language related disorders. In accordance with this model, words are perceived via a specialized word reception center (Wernicke's area) that is located in the left temporoparietal junction. This region then projects to a word production center (Broca's area) that is located in the left inferior frontal gyrus. Because almost all language input was thought to funnel via Wernicke's area and all language output to funnel via Broca's area, it became extremely difficult to identify the basic properties of each region. This lack of clear definition for the contribution of Wernicke's and Broca's regions to human language rendered it extremely difficult to identify their homologues in other primates.[10] With the advent of the fMRI and its application for lesion mappings, however, it was shown that this model is based on incorrect correlations between symptoms and lesions.[11][12][13][14][15][16][17] The refutation of such an influential and dominant model opened the door to new models of language processing in the brain.

Current neurolinguistics models

[edit]

Anatomy

[edit]

In the last two decades, significant advances occurred in our understanding of the neural processing of sounds in primates. Initially by recording of neural activity in the auditory cortices of monkeys[18][19] and later elaborated via histological staining[20][21][22] and fMRI scanning studies,[23] 3 auditory fields were identified in the primary auditory cortex, and 9 associative auditory fields were shown to surround them (Figure 1 top left). Anatomical tracing and lesion studies further indicated of a separation between the anterior and posterior auditory fields, with the anterior primary auditory fields (areas R-RT) projecting to the anterior associative auditory fields (areas AL-RTL), and the posterior primary auditory field (area A1) projecting to the posterior associative auditory fields (areas CL-CM).[20][24][25][26] Recently, evidence accumulated that indicates homology between the human and monkey auditory fields. In humans, histological staining studies revealed two separate auditory fields in the primary auditory region of Heschl's gyrus,[27][28] and by mapping the tonotopic organization of the human primary auditory fields with high resolution fMRI and comparing it to the tonotopic organization of the monkey primary auditory fields, homology was established between the human anterior primary auditory field and monkey area R (denoted in humans as area hR) and the human posterior primary auditory field and the monkey area A1 (denoted in humans as area hA1).[29][30][31][32][33] Intra-cortical recordings from the human auditory cortex further demonstrated similar patterns of connectivity to the auditory cortex of the monkey. Recording from the surface of the auditory cortex (supra-temporal plane) reported that the anterior Heschl's gyrus (area hR) projects primarily to the middle-anterior superior temporal gyrus (mSTG-aSTG) and the posterior Heschl's gyrus (area hA1) projects primarily to the posterior superior temporal gyrus (pSTG) and the planum temporale (area PT; Figure 1 top right).[34][35] Consistent with connections from area hR to the aSTG and hA1 to the pSTG is an fMRI study of a patient with impaired sound recognition (auditory agnosia), who was shown with reduced bilateral activation in areas hR and aSTG but with spared activation in the mSTG-pSTG.[36] This connectivity pattern is also corroborated by a study that recorded activation from the lateral surface of the auditory cortex and reported of simultaneous non-overlapping activation clusters in the pSTG and mSTG-aSTG while listening to sounds.[37]

Downstream to the auditory cortex, anatomical tracing studies in monkeys delineated projections from the anterior associative auditory fields (areas AL-RTL) to ventral prefrontal and premotor cortices in the inferior frontal gyrus (IFG)[38][39] and amygdala.[40] Cortical recording and functional imaging studies in macaque monkeys further elaborated on this processing stream by showing that acoustic information flows from the anterior auditory cortex to the temporal pole (TP) and then to the IFG.[41][42][43][44][45][46] This pathway is commonly referred to as the auditory ventral stream (AVS; Figure 1, bottom left-red arrows). In contrast to the anterior auditory fields, tracing studies reported that the posterior auditory fields (areas CL-CM) project primarily to dorsolateral prefrontal and premotor cortices (although some projections do terminate in the IFG.[47][39] Cortical recordings and anatomical tracing studies in monkeys further provided evidence that this processing stream flows from the posterior auditory fields to the frontal lobe via a relay station in the intra-parietal sulcus (IPS).[48][49][50][51][52][53] This pathway is commonly referred to as the auditory dorsal stream (ADS; Figure 1, bottom left-blue arrows). Comparing the white matter pathways involved in communication in humans and monkeys with diffusion tensor imaging techniques indicates of similar connections of the AVS and ADS in the two species (Monkey,[52] Human[54][55][56][57][58][59]). In humans, the pSTG was shown to project to the parietal lobe (sylvian parietal-temporal junction-inferior parietal lobule; Spt-IPL), and from there to dorsolateral prefrontal and premotor cortices (Figure 1, bottom right-blue arrows), and the aSTG was shown to project to the anterior temporal lobe (middle temporal gyrus-temporal pole; MTG-TP) and from there to the IFG (Figure 1 bottom right-red arrows).

Auditory ventral stream

[edit]

The auditory ventral stream (AVS) connects the auditory cortex with the middle temporal gyrus and temporal pole, which in turn connects with the inferior frontal gyrus. This pathway is responsible for sound recognition, and is accordingly known as the auditory 'what' pathway. The functions of the AVS include the following.

Sound recognition

[edit]

Accumulative converging evidence indicates that the AVS is involved in recognizing auditory objects. At the level of the primary auditory cortex, recordings from monkeys showed higher percentage of neurons selective for learned melodic sequences in area R than area A1,[60] and a study in humans demonstrated more selectivity for heard syllables in the anterior Heschl's gyrus (area hR) than posterior Heschl's gyrus (area hA1).[61] In downstream associative auditory fields, studies from both monkeys and humans reported that the border between the anterior and posterior auditory fields (Figure 1-area PC in the monkey and mSTG in the human) processes pitch attributes that are necessary for the recognition of auditory objects.[18] The anterior auditory fields of monkeys were also demonstrated with selectivity for con-specific vocalizations with intra-cortical recordings.[41][19][62] and functional imaging[63][42][43] One fMRI monkey study further demonstrated a role of the aSTG in the recognition of individual voices.[42] The role of the human mSTG-aSTG in sound recognition was demonstrated via functional imaging studies that correlated activity in this region with isolation of auditory objects from background noise,[64][65] and with the recognition of spoken words,[66][67][68][69][70][71][72] voices,[73] melodies,[74][75] environmental sounds,[76][77][78] and non-speech communicative sounds.[79] A meta-analysis of fMRI studies[80] further demonstrated functional dissociation between the left mSTG and aSTG, with the former processing short speech units (phonemes) and the latter processing longer units (e.g., words, environmental sounds). A study that recorded neural activity directly from the left pSTG and aSTG reported that the aSTG, but not pSTG, was more active when the patient listened to speech in her native language than unfamiliar foreign language.[81] Consistently, electro stimulation to the aSTG of this patient resulted in impaired speech perception[81] (see also[82][83] for similar results). Intra-cortical recordings from the right and left aSTG further demonstrated that speech is processed laterally to music.[81] An fMRI study of a patient with impaired sound recognition (auditory agnosia) due to brainstem damage was also shown with reduced activation in areas hR and aSTG of both hemispheres when hearing spoken words and environmental sounds.[36] Recordings from the anterior auditory cortex of monkeys while maintaining learned sounds in working memory,[46] and the debilitating effect of induced lesions to this region on working memory recall,[84][85][86] further implicate the AVS in maintaining the perceived auditory objects in working memory. In humans, area mSTG-aSTG was also reported active during rehearsal of heard syllables with MEG.[87] and fMRI[88] The latter study further demonstrated that working memory in the AVS is for the acoustic properties of spoken words and that it is independent to working memory in the ADS, which mediates inner speech. Working memory studies in monkeys also suggest that in monkeys, in contrast to humans, the AVS is the dominant working memory store.[89]

In humans, downstream to the aSTG, the MTG and TP are thought to constitute the semantic lexicon, which is a long-term memory repository of audio-visual representations that are interconnected on the basis of semantic relationships. (See also the reviews by[3][4] discussing this topic). The primary evidence for this role of the MTG-TP is that patients with damage to this region (e.g., patients with semantic dementia or herpes simplex virus encephalitis) are reported[90][91] with an impaired ability to describe visual and auditory objects and a tendency to commit semantic errors when naming objects (i.e., semantic paraphasia). Semantic paraphasias were also expressed by aphasic patients with left MTG-TP damage[14][92] and were shown to occur in non-aphasic patients after electro-stimulation to this region.[93][83] or the underlying white matter pathway[94] Two meta-analyses of the fMRI literature also reported that the anterior MTG and TP were consistently active during semantic analysis of speech and text;[66][95] and an intra-cortical recording study correlated neural discharge in the MTG with the comprehension of intelligible sentences.[96]

Sentence comprehension

[edit]

In addition to extracting meaning from sounds, the MTG-TP region of the AVS appears to have a role in sentence comprehension, possibly by merging concepts together (e.g., merging the concept 'blue' and 'shirt' to create the concept of a 'blue shirt'). The role of the MTG in extracting meaning from sentences has been demonstrated in functional imaging studies reporting stronger activation in the anterior MTG when proper sentences are contrasted with lists of words, sentences in a foreign or nonsense language, scrambled sentences, sentences with semantic or syntactic violations and sentence-like sequences of environmental sounds.[97][98][99][100][101][102][103][104] One fMRI study[105] in which participants were instructed to read a story further correlated activity in the anterior MTG with the amount of semantic and syntactic content each sentence contained. An EEG study[106] that contrasted cortical activity while reading sentences with and without syntactic violations in healthy participants and patients with MTG-TP damage, concluded that the MTG-TP in both hemispheres participate in the automatic (rule based) stage of syntactic analysis (ELAN component), and that the left MTG-TP is also involved in a later controlled stage of syntax analysis (P600 component). Patients with damage to the MTG-TP region have also been reported with impaired sentence comprehension.[14][107][108] See review[109] for more information on this topic.

Bilaterality

[edit]

In contradiction to the Wernicke–Lichtheim–Geschwind model that implicates sound recognition to occur solely in the left hemisphere, studies that examined the properties of the right or left hemisphere in isolation via unilateral hemispheric anesthesia (i.e., the WADA procedure[110]) or intra-cortical recordings from each hemisphere[96] provided evidence that sound recognition is processed bilaterally. Moreover, a study that instructed patients with disconnected hemispheres (i.e., split-brain patients) to match spoken words to written words presented to the right or left hemifields, reported vocabulary in the right hemisphere that almost matches in size with the left hemisphere[111] (The right hemisphere vocabulary was equivalent to the vocabulary of a healthy 11-years old child). This bilateral recognition of sounds is also consistent with the finding that unilateral lesion to the auditory cortex rarely results in deficit to auditory comprehension (i.e., auditory agnosia), whereas a second lesion to the remaining hemisphere (which could occur years later) does.[112][113] Finally, as mentioned earlier, an fMRI scan of an auditory agnosia patient demonstrated bilateral reduced activation in the anterior auditory cortices,[36] and bilateral electro-stimulation to these regions in both hemispheres resulted with impaired speech recognition.[81]

Auditory dorsal stream

[edit]

The auditory dorsal stream connects the auditory cortex with the parietal lobe, which in turn connects with inferior frontal gyrus. In both humans and non-human primates, the auditory dorsal stream is responsible for sound localization, and is accordingly known as the auditory 'where' pathway. In humans, this pathway (especially in the left hemisphere) is also responsible for speech production, speech repetition, lip-reading, and phonological working memory and long-term memory.

Speech production

[edit]

Studies of present-day humans have demonstrated a role for the ADS in speech production, particularly in the vocal expression of the names of objects. For instance, in a series of studies in which sub-cortical fibers were directly stimulated[94] interference in the left pSTG and IPL resulted in errors during object-naming tasks, and interference in the left IFG resulted in speech arrest. Magnetic interference in the pSTG and IFG of healthy participants also produced speech errors and speech arrest, respectively[114][115] One study has also reported that electrical stimulation of the left IPL caused patients to believe that they had spoken when they had not and that IFG stimulation caused patients to unconsciously move their lips.[116] The contribution of the ADS to the process of articulating the names of objects could be dependent on the reception of afferents from the semantic lexicon of the AVS, as an intra-cortical recording study reported of activation in the posterior MTG prior to activation in the Spt-IPL region when patients named objects in pictures[117] Intra-cortical electrical stimulation studies also reported that electrical interference to the posterior MTG was correlated with impaired object naming[118][82]

Additionally, lesion studies of stroke patients have provided evidence supporting the dual stream model's role in speech production. Recent research using multivariate lesion/disconnectome symptom mapping has shown that lower scores in speech production tasks are associated with lesions and abnormalities in the left inferior parietal lobe and frontal lobe. These findings from stroke patients further support the involvement of the dorsal stream pathway in speech production, complementing the stimulation and interference studies in healthy participants.[119]

Vocal mimicry

[edit]

Although sound perception is primarily ascribed with the AVS, the ADS appears associated with several aspects of speech perception. For instance, in a meta-analysis of fMRI studies[120] (Turkeltaub and Coslett, 2010), in which the auditory perception of phonemes was contrasted with closely matching sounds, and the studies were rated for the required level of attention, the authors concluded that attention to phonemes correlates with strong activation in the pSTG-pSTS region. An intra-cortical recording study in which participants were instructed to identify syllables also correlated the hearing of each syllable with its own activation pattern in the pSTG.[121] The involvement of the ADS in both speech perception and production has been further illuminated in several pioneering functional imaging studies that contrasted speech perception with overt or covert speech production.[122][123][124] These studies demonstrated that the pSTS is active only during the perception of speech, whereas area Spt is active during both the perception and production of speech. The authors concluded that the pSTS projects to area Spt, which converts the auditory input into articulatory movements.[125][126] Similar results have been obtained in a study in which participants' temporal and parietal lobes were electrically stimulated. This study reported that electrically stimulating the pSTG region interferes with sentence comprehension and that stimulation of the IPL interferes with the ability to vocalize the names of objects.[83] The authors also reported that stimulation in area Spt and the inferior IPL induced interference during both object-naming and speech-comprehension tasks. The role of the ADS in speech repetition is also congruent with the results of the other functional imaging studies that have localized activation during speech repetition tasks to ADS regions.[127][128][129] An intra-cortical recording study that recorded activity throughout most of the temporal, parietal and frontal lobes also reported activation in the pSTG, Spt, IPL and IFG when speech repetition is contrasted with speech perception.[130] Neuropsychological studies have also found that individuals with speech repetition deficits but preserved auditory comprehension (i.e., conduction aphasia) suffer from circumscribed damage to the Spt-IPL area[131][132][133][134][135][136][137] or damage to the projections that emanate from this area and target the frontal lobe[138][139][140][141] Studies have also reported a transient speech repetition deficit in patients after direct intra-cortical electrical stimulation to this same region.[11][142][143] Insight into the purpose of speech repetition in the ADS is provided by longitudinal studies of children that correlated the learning of foreign vocabulary with the ability to repeat nonsense words.[144][145]

Speech monitoring

[edit]

In addition to repeating and producing speech, the ADS appears to have a role in monitoring the quality of the speech output. Neuroanatomical evidence suggests that the ADS is equipped with descending connections from the IFG to the pSTG that relay information about motor activity (i.e., corollary discharges) in the vocal apparatus (mouth, tongue, vocal folds). This feedback marks the sound perceived during speech production as self-produced and can be used to adjust the vocal apparatus to increase the similarity between the perceived and emitted calls. Evidence for descending connections from the IFG to the pSTG has been offered by a study that electrically stimulated the IFG during surgical operations and reported the spread of activation to the pSTG-pSTS-Spt region[146] A study[147] that compared the ability of aphasic patients with frontal, parietal or temporal lobe damage to quickly and repeatedly articulate a string of syllables reported that damage to the frontal lobe interfered with the articulation of both identical syllabic strings ("Bababa") and non-identical syllabic strings ("Badaga"), whereas patients with temporal or parietal lobe damage only exhibited impairment when articulating non-identical syllabic strings. Because the patients with temporal and parietal lobe damage were capable of repeating the syllabic string in the first task, their speech perception and production appears to be relatively preserved, and their deficit in the second task is therefore due to impaired monitoring. Demonstrating the role of the descending ADS connections in monitoring emitted calls, an fMRI study instructed participants to speak under normal conditions or when hearing a modified version of their own voice (delayed first formant) and reported that hearing a distorted version of one's own voice results in increased activation in the pSTG.[148] Further demonstrating that the ADS facilitates motor feedback during mimicry is an intra-cortical recording study that contrasted speech perception and repetition.[130] The authors reported that, in addition to activation in the IPL and IFG, speech repetition is characterized by stronger activation in the pSTG than during speech perception.

Integration of phonemes with lip-movements

[edit]

Although sound perception is primarily ascribed with the AVS, the ADS appears associated with several aspects of speech perception. For instance, in a meta-analysis of fMRI studies[120] in which the auditory perception of phonemes was contrasted with closely matching sounds, and the studies were rated for the required level of attention, the authors concluded that attention to phonemes correlates with strong activation in the pSTG-pSTS region. An intra-cortical recording study in which participants were instructed to identify syllables also correlated the hearing of each syllable with its own activation pattern in the pSTG.[149] Consistent with the role of the ADS in discriminating phonemes,[120] studies have ascribed the integration of phonemes and their corresponding lip movements (i.e., visemes) to the pSTS of the ADS. For example, an fMRI study[150] has correlated activation in the pSTS with the McGurk illusion (in which hearing the syllable "ba" while seeing the viseme "ga" results in the perception of the syllable "da"). Another study has found that using magnetic stimulation to interfere with processing in this area further disrupts the McGurk illusion.[151] The association of the pSTS with the audio-visual integration of speech has also been demonstrated in a study that presented participants with pictures of faces and spoken words of varying quality. The study reported that the pSTS selects for the combined increase of the clarity of faces and spoken words.[152] Corroborating evidence has been provided by an fMRI study[153] that contrasted the perception of audio-visual speech with audio-visual non-speech (pictures and sounds of tools). This study reported the detection of speech-selective compartments in the pSTS. In addition, an fMRI study[154] that contrasted congruent audio-visual speech with incongruent speech (pictures of still faces) reported pSTS activation. For a review presenting additional converging evidence regarding the role of the pSTS and ADS in phoneme-viseme integration see.[155]

Empirical research has demonstrated that visual lip movements enhance speech processing along the auditory dorsal stream, particularly in noisy conditions. Recent studies [156] discovered that the dorsal stream regions, including frontal speech motor areas and supramarginal gyrus, show improved neural representations of speech sounds when visual lip movements are available.

Phonological long-term memory

[edit]

A growing body of evidence indicates that humans, in addition to having a long-term store for word meanings located in the MTG-TP of the AVS (i.e., the semantic lexicon), also have a long-term store for the names of objects located in the Spt-IPL region of the ADS (i.e., the phonological lexicon). For example, a study[157][158] examining patients with damage to the AVS (MTG damage) or damage to the ADS (IPL damage) reported that MTG damage results in individuals incorrectly identifying objects (e.g., calling a "goat" a "sheep," an example of semantic paraphasia). Conversely, IPL damage results in individuals correctly identifying the object but incorrectly pronouncing its name (e.g., saying "gof" instead of "goat," an example of phonemic paraphasia). Semantic paraphasia errors have also been reported in patients receiving intra-cortical electrical stimulation of the AVS (MTG), and phonemic paraphasia errors have been reported in patients whose ADS (pSTG, Spt, and IPL) received intra-cortical electrical stimulation.[83][159][94] Further supporting the role of the ADS in object naming is an MEG study that localized activity in the IPL during the learning and during the recall of object names.[160] A study that induced magnetic interference in participants' IPL while they answered questions about an object reported that the participants were capable of answering questions regarding the object's characteristics or perceptual attributes but were impaired when asked whether the word contained two or three syllables.[161] An MEG study has also correlated recovery from anomia (a disorder characterized by an impaired ability to name objects) with changes in IPL activation.[162] Further supporting the role of the IPL in encoding the sounds of words are studies reporting that, compared to monolinguals, bilinguals have greater cortical density in the IPL but not the MTG.[163][164] Because evidence shows that, in bilinguals, different phonological representations of the same word share the same semantic representation,[165] this increase in density in the IPL verifies the existence of the phonological lexicon: the semantic lexicon of bilinguals is expected to be similar in size to the semantic lexicon of monolinguals, whereas their phonological lexicon should be twice the size. Consistent with this finding, cortical density in the IPL of monolinguals also correlates with vocabulary size.[166][167] Notably, the functional dissociation of the AVS and ADS in object-naming tasks is supported by cumulative evidence from reading research showing that semantic errors are correlated with MTG impairment and phonemic errors with IPL impairment. Based on these associations, the semantic analysis of text has been linked to the inferior-temporal gyrus and MTG, and the phonological analysis of text has been linked to the pSTG-Spt- IPL[168][169][170]

Phonological working memory

[edit]

Working memory is often treated as the temporary activation of the representations stored in long-term memory that are used for speech (phonological representations). This sharing of resources between working memory and speech is evident by the finding[171][172] that speaking during rehearsal results in a significant reduction in the number of items that can be recalled from working memory (articulatory suppression). The involvement of the phonological lexicon in working memory is also evidenced by the tendency of individuals to make more errors when recalling words from a recently learned list of phonologically similar words than from a list of phonologically dissimilar words (the phonological similarity effect).[171] Studies have also found that speech errors committed during reading are remarkably similar to speech errors made during the recall of recently learned, phonologically similar words from working memory.[173] Patients with IPL damage have also been observed to exhibit both speech production errors and impaired working memory[174][175][176][177] Finally, the view that verbal working memory is the result of temporarily activating phonological representations in the ADS is compatible with recent models describing working memory as the combination of maintaining representations in the mechanism of attention in parallel to temporarily activating representations in long-term memory.[172][178][179][180] It has been argued that the role of the ADS in the rehearsal of lists of words is the reason this pathway is active during sentence comprehension[181] For a review of the role of the ADS in working memory, see.[182]

Studies have shown that performance on phonological working memory tasks correlates with properties of the left dorsal branch of the arcuate fasciculus (AF), which connects posterior temporal language regions with attention-regulating areas in the middle frontal gyrus. The arcuate fasciculus is a white matter pathway in the brain which contains two branches: a ventral branch connecting Wernicke's area with Broca's area and a dorsal branch connecting the posterior temporal region with the middle frontal gyrus. This dorsal branch appears to be particularly important for phonological working memory processes.[183]

The 'from where to what' model of language evolution hypotheses 7 stages of language evolution.

Linguistic theories

[edit]

Language-processing research informs theories of language. The primary theoretical question is whether linguistic structures follow from the brain structures or vice versa. Externalist models, such as Ferdinand de Saussure's structuralism, argue that language as a social phenomenon is external to the brain. The individual receives the linguistic system from the outside, and the given language shapes the individual's brain.[184]

This idea is opposed by internalist models including Noam Chomsky's transformational generative grammar, George Lakoff's Cognitive Linguistics, and John A. Hawkins's efficiency hypothesis. According to Chomsky, language is acquired from an innate brain structure independently of meaning.[185] Lakoff argues that language emerges from the sensory systems.[186] Hawkins hypothesizes that cross-linguistically prevalent patterns are based on the brain's natural processing preferences.[187]

Additionally, models inspired by Richard Dawkins's memetics, including Construction Grammar and Usage-Based Linguistics, advocate a two-way model arguing that the brain shapes language, and language shapes the brain.[188][189]

Evidence from neuroimaging studies points towards the externalist position. ERP studies suggest that language processing is based on the interaction of syntax and semantics, and the research does not support innate grammatical structures.[190][191] MRI studies suggest that the structural characteristics of the child's first language shapes the processing connectome of the brain.[192] Processing research has failed to find support for the inverse idea that syntactic structures reflect the brain's natural processing preferences cross-linguistically.[193]

The evolution of language

[edit]

The auditory dorsal stream also has non-language related functions, such as sound localization[194][195][196][197][198] and guidance of eye movements.[199][200] Recent studies also indicate a role of the ADS in localization of family/tribe members, as a study[201] that recorded from the cortex of an epileptic patient reported that the pSTG, but not aSTG, is selective for the presence of new speakers. An fMRI[202] study of fetuses at their third trimester also demonstrated that area Spt is more selective to female speech than pure tones, and a sub-section of Spt is selective to the speech of their mother in contrast to unfamiliar female voices.

It is presently unknown why so many functions are ascribed to the human ADS. An attempt to unify these functions under a single framework was conducted in the 'From where to what' model of language evolution[203][204] In accordance with this model, each function of the ADS indicates of a different intermediate phase in the evolution of language. The roles of sound localization and integration of sound location with voices and auditory objects is interpreted as evidence that the origin of speech is the exchange of contact calls (calls used to report location in cases of separation) between mothers and offspring. The role of the ADS in the perception and production of intonations is interpreted as evidence that speech began by modifying the contact calls with intonations, possibly for distinguishing alarm contact calls from safe contact calls. The role of the ADS in encoding the names of objects (phonological long-term memory) is interpreted as evidence of gradual transition from modifying calls with intonations to complete vocal control. The role of the ADS in the integration of lip movements with phonemes and in speech repetition is interpreted as evidence that spoken words were learned by infants mimicking their parents' vocalizations, initially by imitating their lip movements. The role of the ADS in phonological working memory is interpreted as evidence that the words learned through mimicry remained active in the ADS even when not spoken. This resulted with individuals capable of rehearsing a list of vocalizations, which enabled the production of words with several syllables. Further developments in the ADS enabled the rehearsal of lists of words, which provided the infra-structure for communicating with sentences.

Sign language in the brain

[edit]

Neuroscientific research has provided a scientific understanding of how sign language is processed in the brain. There are over 135 discrete sign languages around the world- making use of different accents formed by separate areas of a country.[205]

By utilizing lesion analyses and neuroimaging, neuroscientists have discovered that whether it be spoken or sign language, human brains process language in general, in a similar manner regarding which area of the brain is being used.[205]Lesion analyses are used to examine the consequences of damage to specific brain regions involved in language while neuroimaging explore regions that are engaged in the processing of language.[205]

Previous hypotheses have been made that damage to Broca's area or Wernicke's area does not affect sign language being perceived; however, it is not the case. Studies have shown that damage to these areas are similar in results in spoken language where sign errors are present and/or repeated.[205]In both types of languages, they are affected by damage to the left hemisphere of the brain rather than the right -usually dealing with the arts.

There are obvious patterns for utilizing and processing language. Both speaking and signing activate Broca's area, while processing sign language employs Wernicke's area similar to that of spoken language.[205]

There have been other hypotheses about the lateralization of the two hemispheres. Specifically, the right hemisphere was thought to contribute to the overall communication of a language globally whereas the left hemisphere would be dominant in generating the language locally.[206] Through research in aphasias, RHD signers were found to have a problem maintaining the spatial portion of their signs, confusing similar signs at different locations necessary to communicate with another properly.[206] LHD signers, on the other hand, had similar results to those of hearing patients. Furthermore, other studies have emphasized that sign language is present bilaterally but will need to continue researching to reach a conclusion.[206]

Writing in the brain

[edit]

There is a comparatively small body of research on the neurology of reading and writing.[207] Most of the studies performed deal with reading rather than writing or spelling, and the majority of both kinds focus solely on the English language.[208] English orthography is less transparent than that of other languages using a Latin script.[207] Another difficulty is that some studies focus on spelling words of English and omit the few logographic characters found in the script.[207]

In terms of spelling, English words can be divided into three categories – regular, irregular, and "novel words" or "nonwords." Regular words are those in which there is a regular, one-to-one correspondence between grapheme and phoneme in spelling. Irregular words are those in which no such correspondence exists. Nonwords are those that exhibit the expected orthography of regular words but do not carry meaning, such as nonce words and onomatopoeia.[207]

An issue in the cognitive and neurological study of reading and spelling in English is whether a single-route or dual-route model best describes how literate speakers are able to read and write all three categories of English words according to accepted standards of orthographic correctness. Single-route models posit that lexical memory is used to store all spellings of words for retrieval in a single process. Dual-route models posit that lexical memory is employed to process irregular and high-frequency regular words, while low-frequency regular words and nonwords are processed using a sub-lexical set of phonological rules.[207]

The single-route model for reading has found support in computer modelling studies, which suggest that readers identify words by their orthographic similarities to phonologically alike words.[207] However, cognitive and lesion studies lean towards the dual-route model. Cognitive spelling studies on children and adults suggest that spellers employ phonological rules in spelling regular words and nonwords, while lexical memory is accessed to spell irregular words and high-frequency words of all types.[207] Similarly, lesion studies indicate that lexical memory is used to store irregular words and certain regular words, while phonological rules are used to spell nonwords.[207]

More recently, neuroimaging studies using positron emission tomography and fMRI have suggested a balanced model in which the reading of all word types begins in the visual word form area, but subsequently branches off into different routes depending upon whether or not access to lexical memory or semantic information is needed (which would be expected with irregular words under a dual-route model).[207] A 2007 fMRI study found that subjects asked to produce regular words in a spelling task exhibited greater activation in the left posterior STG, an area used for phonological processing, while the spelling of irregular words produced greater activation of areas used for lexical memory and semantic processing, such as the left IFG and left SMG and both hemispheres of the MTG.[207] Spelling nonwords was found to access members of both pathways, such as the left STG and bilateral MTG and ITG.[207] Significantly, it was found that spelling induces activation in areas such as the left fusiform gyrus and left SMG that are also important in reading, suggesting that a similar pathway is used for both reading and writing.[207]

Far less information exists on the cognition and neurology of non-alphabetic and non-English scripts. Every language has a morphological and a phonological component, either of which can be recorded by a writing system. Scripts recording words and morphemes are considered logographic, while those recording phonological segments, such as syllabaries and alphabets, are phonographic.[208] Most systems combine the two and have both logographic and phonographic characters.[208]

In terms of complexity, writing systems can be characterized as "transparent" or "opaque" and as "shallow" or "deep". A "transparent" system exhibits an obvious correspondence between grapheme and sound, while in an "opaque" system this relationship is less obvious. The terms "shallow" and "deep" refer to the extent that a system's orthography represents morphemes as opposed to phonological segments.[208] Systems that record larger morphosyntactic or phonological segments, such as logographic systems and syllabaries put greater demand on the memory of users.[208] It would thus be expected that an opaque or deep writing system would put greater demand on areas of the brain used for lexical memory than would a system with transparent or shallow orthography.

See also

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Language processing in the brain encompasses the neural mechanisms that enable humans to comprehend, produce, and manipulate linguistic information, primarily involving a distributed network in the left cerebral hemisphere. This process relies on specialized regions such as Broca's area in the inferior frontal gyrus (Brodmann areas 44 and 45), which supports speech production and grammatical processing, and Wernicke's area in the posterior superior temporal gyrus (Brodmann area 22), which is essential for language comprehension and semantic interpretation. These areas are interconnected through white matter tracts forming dorsal and ventral streams: the dorsal pathway, via the arcuate fasciculus, handles phonological mapping and sound-to-articulation transformations, while the ventral pathway, involving the inferior fronto-occipital fasciculus, facilitates semantic processing and meaning retrieval. The foundational understanding of these mechanisms emerged in the 19th century through clinical observations of aphasia patients. In 1861, French neurologist Paul Broca identified a lesion in the left inferior frontal gyrus of a patient with expressive aphasia, establishing Broca's area as critical for articulate speech production. Over a decade later, in 1874, German neurologist Carl Wernicke described fluent but incomprehensible speech resulting from damage to the posterior superior temporal region, linking Wernicke's area to receptive language functions and proposing an early connectionist model between these regions. Subsequent neuroimaging studies, including functional MRI, have confirmed and expanded this classical model, revealing bilateral involvement in some aspects of processing and a broader perisylvian network that integrates auditory, motor, and cognitive systems. Modern neurobiological models emphasize causal mechanisms grounded in neural dynamics, such as spiking neurons, , and states within cortical microcircuits. The dual-stream framework, refined through and electrophysiological studies, posits that the dorsal stream supports interface between phonological and articulatory representations, while the ventral stream enables comprehension via lexical-semantic access. Language processing also involves unification operations for combining words into meaningful sentences, occurring rapidly in prefrontal and temporal regions on timescales of hundreds of milliseconds. Disruptions, as seen in aphasias, highlight the system's vulnerability, yet plasticity allows recovery through compensatory recruitment of homologous right-hemisphere areas or adjacent networks. Ongoing integrates computational simulations with invasive recordings to elucidate how these mechanisms adapt during learning and bilingualism.

Neuroanatomy of Language

Core Language Areas

The core language areas are primarily located in the left and form the foundational neural substrate for functions in most individuals. These regions, identified through historical lesion studies and modern , include , , and the , each contributing distinct aspects of processing. Subcortical structures such as the and provide essential support for coordinating these cortical activities. processing exhibits strong lateralization, with approximately 90-95% of right-handed individuals showing left-hemisphere dominance for core functions, though left-handers display greater variability, with right-hemisphere involvement in up to 30% of cases depending on the degree of . Broca's area, situated in the left (Brodmann areas 44 and 45), plays a central role in , syntactic processing, and the articulation of complex grammatical structures. It was first identified in 1861 by French physician through postmortem examinations of patients with non-fluent , such as the famous case of "Tan" (Louis Leborgne), whose in this region correlated with severe expressive language deficits despite preserved comprehension. studies and confirm that damage to Broca's area disrupts motor planning for speech and hierarchical syntactic operations, often resulting in characterized by short, agrammatic phrases. Wernicke's area, located in the posterior () of the left hemisphere, is essential for language comprehension, semantic processing, and the interpretation of auditory verbal input. Discovered in 1874 by German neurologist , it was characterized through observations of patients exhibiting fluent but nonsensical speech () due to lesions in this region, highlighting its role in mapping sounds to meaning. evidence supports its involvement in phonological and lexical-semantic analysis, where it integrates auditory signals to form coherent linguistic representations. The , found in the left inferior (), facilitates the integration of visual and auditory information crucial for reading, writing, and multimodal tasks. It serves as a for associating orthographic, phonological, and semantic elements, enabling processes like and cross-modal mapping in . Lesion studies demonstrate that angular gyrus damage impairs and mathematical-spatial integration, underscoring its role in higher-order associative functions beyond basic speech or audition. Subcortically, the and modulate language by supporting motor sequencing, procedural learning, and attentional gating of linguistic elements. The , including structures like the and , contribute to the rhythmic timing and sequential organization of , as evidenced by their involvement in disorders like , which disrupt fluent articulation. The acts as a relay hub, filtering sensory inputs to cortical language areas and enhancing selective during comprehension tasks, with its anterior and showing activation in verbal fluency and semantic retrieval paradigms.

Supporting Networks and Connectivity

The arcuate fasciculus is a major tract that connects in the to in the , facilitating the mapping of phonological representations during processing. This pathway supports the repetition and production of speech sounds, with disruptions leading to , characterized by impaired phonological repetition despite preserved comprehension and fluent speech. Diffusion tensor imaging (DTI) studies since the 2000s have revealed bidirectional connectivity along the arcuate fasciculus, enabling efficient information flow between auditory and motor regions for phonological integration. The superior longitudinal fasciculus (SLF) forms a key component of the dorsal language pathway, linking parietal and frontal regions to support articulation and sensory-motor mapping in . DTI evidence indicates that the SLF provides bidirectional connections that coordinate articulatory planning, with integrity correlating to repetition and naming abilities post-stroke. Complementing this, the inferior fronto-occipital fasciculus (IFOF) constitutes the ventral pathway, connecting frontal and occipital-temporal areas to underpin semantic processing and meaning retrieval. and studies highlight the IFOF's role in integrating , with left-hemisphere damage impairing object naming and conceptual associations. The enables interhemispheric transfer of linguistic information, particularly relevant for bilingual individuals who exhibit enhanced callosal microstructure to manage dual- systems. In recovery from brain injury, DTI reveals that preserved callosal integrity facilitates reorganization by promoting compensatory activation across hemispheres. Additionally, the (DMN) overlaps with networks during processing, supporting internal and story comprehension through connectivity in medial prefrontal and posterior cingulate regions. This integration allows for the of coherent mental , as evidenced by functional connectivity analyses during semantic and episodic tasks.

Historical and Current Models

Early Neurolinguistic Models

Early neurolinguistic models emerged in the mid-19th century through clinico-pathological studies of , driven by the localizationist principle that specific brain regions govern distinct cognitive functions. In 1861, identified a lesion in the left of a patient with severe speech production deficits but preserved comprehension, establishing this area—now known as —as critical for articulate speech. This finding, based on correlations, marked a pivotal shift toward cerebral localization of , influencing subsequent aphasiology by emphasizing hemispheric asymmetry, particularly left-hemisphere dominance. Building on Broca's work, in 1874 described a sensory form of linked to lesions in the posterior , termed , characterized by fluent but semantically empty speech and impaired auditory comprehension. Patients with produce paraphasic errors, such as neologisms or , while maintaining normal prosody and , yet struggle to understand spoken or written due to disrupted sound-to-meaning mapping. In the , Norman Geschwind synthesized these observations into the Wernicke-Geschwind model, positing a hierarchical pathway for processing: auditory input from the primary projects to for comprehension, then via the arcuate fasciculus—a tract—to Broca's area for . Damage to the arcuate fasciculus was theorized to cause , featuring intact comprehension and fluent speech but severe repetition deficits and phonemic errors, as the disconnection prevents phonological relay between comprehension and production centers. This model embodied a modular view of language, treating comprehension and production as discrete, localized modules connected by dedicated conduits, which simplified explanations of aphasia syndromes but overlooked broader cognitive integrations. Critiques highlight its oversimplification of semantics, as it primarily addressed phonological and syntactic processing while neglecting how meaning is constructed across distributed networks. Additionally, the model's strict left-lateralized focus ignored evidence of bilateral contributions to language, particularly in prosody and semantics, and failed to account for variability in lesion outcomes, where arcuate damage does not uniformly produce conduction aphasia. These limitations, revealed through later anatomical and functional studies, underscored the need for more dynamic frameworks beyond lesion-based localizationism.

Modern Neurolinguistic Models

Modern neurolinguistic models have evolved to incorporate interactive and dynamic processes, addressing limitations of earlier lesion-based approaches by integrating data and computational simulations. These frameworks emphasize parallelism and bidirectional interactions in language processing, revealing how the handles comprehension and production through overlapping neural networks rather than isolated modules. The dual-stream model, proposed by Hickok and Poeppel in the 2000s, posits two partially segregated pathways for speech and language: a ventral stream primarily supporting comprehension by mapping sound to meaning, and a dorsal stream facilitating production and phonological mapping for articulation. This model, informed by studies, highlights bilateral processing in superior temporal regions for early auditory analysis, with asymmetries emerging in higher-level functions. Updates from () and () further demonstrate parallel activation across streams during real-time language tasks, enabling flexible integration of sensory input and motor output. Predictive coding theories, advanced by Friston in the 2010s, frame processing as where the brain generates top-down about incoming linguistic input to minimize errors. In this hierarchical scheme, cortical layers anticipate syntactic and semantic structures, with evidence from fMRI showing reduced neural activity for expected words and heightened responses to surprises, supporting efficient communication. Applied to , these models explain phenomena like syntactic priming, where prior exposure facilitates of grammatical continuations. The discovery of mirror neurons in the 1990s by Rizzolatti and colleagues in premotor areas like the has informed modern views on action- interfaces, suggesting these cells activate both during action observation and execution, potentially linking motor simulation to linguistic comprehension. fMRI and MEG studies in the 2000s and beyond have extended this to , showing mirror neuron involvement in action verbs and metaphors, though their precise role remains debated as facilitative rather than essential. Connectionist models, utilizing artificial neural networks to simulate , challenge strict by demonstrating emergent linguistic abilities through distributed learning rather than innate rules. These networks, trained on probabilistic input, replicate patterns like past-tense formation without explicit programming, aligning with brain-like parallel distributed processing. Critiques of , bolstered by evidence from recovery studies, argue that language networks reorganize dynamically across development and injury, integrating with broader cognitive systems rather than operating in isolation. In the 2020s, AI-inspired models have advanced syntax prediction in , with large language models (LLMs) like exhibiting brain-like hierarchical representations that correlate with fMRI activations during sentence processing. These approaches reveal how predictive mechanisms in artificial networks mirror cortical hierarchies for syntactic disambiguation, offering insights into plasticity and offering testable hypotheses for .

Auditory Processing Streams

Ventral Stream Functions

The ventral stream in the , often termed the "what" pathway, primarily facilitates the mapping of auditory input to meaning during comprehension, involving a network of regions that process sounds from basic acoustic features to complex semantic representations. This pathway originates in early auditory areas and extends anteriorly through the (STG) and sulcus (STS), progressing to the middle (MTG) and anterior for higher-level integration. In contrast to the dorsal stream's role in sound-to-articulation mapping for production, the ventral stream focuses on recognition and interpretation without direct involvement in motor output. Sound-to-lexicon mapping occurs progressively along the ventral stream, beginning in the anterior (aSTS), where phonetic representations are formed from incoming speech signals, and extending to the MTG, which supports lexical access by linking these representations to word meanings. This process involves hierarchical , with posterior STS regions encoding spectrotemporal features of sounds and more anterior MTG areas integrating them into recognizable lexical items. For instance, studies show that aSTS activation correlates with the resolution of phonetic ambiguities in spoken words, facilitating the transition from auditory input to stored lexical knowledge. Sound recognition in the ventral stream unfolds in distinct stages, starting with the extraction of acoustic features such as frequency and amplitude modulations in primary , followed by phonetic feature assembly in the posterior STG and STS. These stages enable the transformation of raw sound waves into invariant phonetic units, with increasing tolerance to acoustic variability (e.g., speaker differences) as signals propagate anteriorly along the pathway. This hierarchical processing ensures robust recognition of speech elements before lexical integration. Neural adaptations of lexical access models, such as the cohort theory, illustrate how the ventral stream handles by activating multiple candidate words based on initial phonetic input and narrowing them via contextual cues. In this framework, posterior temporal regions initiate a "cohort" of phonetically similar activations, while anterior MTG regions resolve competition through semantic and syntactic constraints, mirroring behavioral predictions of the theory at the neural level. Functional MRI evidence supports this, showing graded activation patterns along the ventral stream during tasks requiring unique word identification from partial acoustic information. At the sentence level, the ventral stream contributes to comprehension by integrating syntactic structure and semantic content through connections between the and . This pathway supports the unification of lexical items into coherent meanings, with MTG and facilitating both local phrase building and global thematic role assignment. For example, during processing of ambiguous sentences, ventral stream activity resolves semantic dependencies, ensuring interpretive coherence. The ventral stream exhibits bilaterality, with the left hemisphere dominating lexical and syntactic processing, while the right hemisphere contributes to prosody and broader contextual integration. Right ventral temporal regions, including the STS, enhance comprehension of intonational cues and emotional tone in speech, aiding disambiguation in naturalistic contexts. Evidence from studies demonstrates that isolated right hemispheres can process prosodic elements and simple semantic relations, underscoring the pathway's distributed nature. Disruptions to ventral stream regions, as seen in , lead to profound deficits in conceptual knowledge and word meaning retrieval while sparing phonological and syntactic abilities. Atrophy in the anterior temporal lobes, particularly the ATL, impairs multimodal semantic associations, resulting in anomia and category-specific comprehension failures. studies confirm that ventral pathway damage selectively hinders lexico-semantic access, with preserved dorsal functions allowing fluent but empty speech output. Recent research highlights the ventral stream's role in predictive semantics, where anterior temporal regions generate expectations about upcoming words based on prior context, facilitating efficient comprehension. For instance, fMRI studies show enhanced MTG activity during predictive integration in narratives, with top-down signals from prefrontal areas modulating ventral processing to anticipate meanings. This predictive mechanism, supported by hierarchical models, reduces processing load in real-time language use.

Dorsal Stream Functions

The dorsal stream in the auditory processing of serves as the primary pathway for mapping acoustic speech signals to articulatory motor representations, facilitating the transformation of sound into action. This "how" pathway connects the posterior (pSTG) to the through the dorsal branch of the arcuate fasciculus, enabling planning by integrating auditory input with motor output for phonological encoding and articulation sequencing. In this process, the stream supports the online adjustment of speech motor commands based on auditory feedback, ensuring precise coordination during verbal output. A key function of the dorsal stream involves phonological , particularly through syllable mechanisms in the left (IPL), which acts as a temporary buffer for maintaining and manipulating speech sounds. This region, part of the phonological loop, relies on subvocal articulation to refresh decaying phonological traces, allowing for the of verbal sequences essential for tasks like repetition and sentence construction. The IPL's role in this process highlights the dorsal stream's contribution to short-term storage of sublexical units, distinct from semantic processing. The dorsal stream also underpins vocal mimicry and during , utilizing efference copies—internal predictions of sensory consequences from motor commands—to distinguish self-generated sounds from external auditory input. These efference signals, generated in frontal regions and propagated through the dorsal pathway, suppress responses to one's own voice, enabling real-time . Furthermore, the stream integrates auditory phonemes with visual cues from lip movements, providing a neural basis for phenomena like the , where conflicting audiovisual inputs fuse into a unified percept that influences motor planning via posterior projections to premotor areas. In addition to transient processing, the dorsal stream contributes to long-term phonological storage by supporting retrieval for production, with the serving as a dorsal phonological that maps conceptual representations to articulatory forms. Disruptions to this pathway, such as damage to the arcuate fasciculus, manifest in , characterized by impaired planning and execution of speech movements despite intact muscle control, leading to effortful, groping articulations. Recent advances in ultrasound tongue imaging have revealed the dorsal stream's adaptability in accent learning, demonstrating neural plasticity in motor as learners adjust articulatory gestures to novel phonological patterns through visual .

Integration with Linguistics and Cognition

Neurolinguistic Theories

Neurolinguistic theories seek to integrate insights from with linguistic principles to explain how the brain processes syntax, semantics, and universal linguistic structures. These theories draw on brain imaging, lesion studies, and genetic evidence to test hypotheses about innate versus learned language mechanisms, often challenging or refining classical linguistic models. For instance, Noam Chomsky's concept of (UG), which posits an innate biological endowment for including recursive syntax, has been examined through neurogenetic links, particularly the gene discovered in the as a key regulator of speech and grammatical processing. Mutations in FOXP2, identified in families with , disrupt orofacial motor control and syntactic abilities, providing genetic evidence for a biological basis of UG-like structures, though the gene's role extends beyond syntax to broader vocal learning circuits. Neural evidence supports specific syntactic operations central to UG, such as the "merge" mechanism, which builds hierarchical phrase structures. Functional magnetic resonance imaging (fMRI) studies show that processing hierarchical dependencies in sentences activates subregions of the left inferior frontal gyrus (IFG), particularly Brodmann areas 44 and 45 (Broca's area), with greater activation for deeper embeddings compared to flat structures, indicating a dedicated neural computation for recursion. Lesions or disruptions in the left IFG impair hierarchical syntax comprehension while sparing simpler associations. In semantics, the temporal lobes, especially the anterior temporal lobe (ATL) and superior temporal gyrus, encode word meanings in distributed patterns that reflect the distributional hypothesis—that semantic similarity arises from contextual co-occurrences in language use. Voxel-wise modeling of natural language narratives via fMRI reveals semantic maps in the ATL mirroring vector-space representations from distributional models, where words like "carnivore" cluster near related concepts based on usage patterns, supporting a usage-driven rather than purely innate semantic foundation. Embodied cognition theories further bridge and by proposing that comprehension is grounded in sensorimotor experiences, rather than abstract symbols alone. For example, action verbs like "" activates motor cortex regions corresponding to leg movements, as shown in (TMS) and fMRI experiments, suggesting meanings are simulated through bodily states to facilitate understanding. This grounding challenges disembodied views, emphasizing how sensorimotor simulations in premotor and somatosensory areas integrate with linguistic for contextual interpretation. Usage-based models critique Chomskyan by arguing that linguistic structures emerge from statistical patterns in input, without requiring a domain-specific UG module; neurolinguistic evidence from electroencephalography (EEG) shows that frequency of exposure modulates neural responses to grammatical constructions, supporting emergentist accounts over hardwired rules. These models highlight in perisylvian networks, where repeated usage strengthens connections for syntax and semantics via Hebbian learning. Recent 2020s debates on compositionality— the principle that complex meanings arise systematically from simpler parts—have been informed by BERT-like models, which approximate human-like semantic processing but reveal limitations in true hierarchical generalization. alignments between activations and brain data show that models like BERT capture in temporal regions but struggle with novel recombinations, prompting questions about whether human compositionality relies on innate biases or learned approximations. Studies decoding fMRI responses during compositional tasks indicate that the left IFG and ATL support flexible meaning integration beyond static embeddings, suggesting neural mechanisms exceed current AI capabilities in handling ambiguity and context. These findings refine neurolinguistic theories by integrating computational modeling with brain data, highlighting ongoing tensions between innatist and experiential accounts, with 2025 studies further exploring AI-brain alignments for predictive processing.

Language Evolution and Development

The evolution of language processing in the human brain involved key genetic and neuroanatomical milestones that distinguished Homo sapiens from other primates. Mutations in the FOXP2 gene, essential for vocal motor control and sequencing of orofacial movements, emerged after the human-chimpanzee divergence approximately 6–7 million years ago, with two amino acid substitutions shared with Neandertals becoming fixed prior to their divergence around 500,000–800,000 years ago, aligning with early Homo species and potentially enabling complex speech production. Concurrently, the human brain underwent significant restructuring, including a taller frontal lobe and anterior-medial reorientation of the temporal poles between 100,000 and 35,000 years ago, which enlarged and refined frontotemporal regions critical for integrating auditory and articulatory aspects of language. These changes supported the phylogenetic origins of language by enhancing neural circuits for symbolic communication, as evidenced by fossil endocasts showing globularization of the brain shape during this period. Cultural and genetic interactions further shaped language evolution through mechanisms like the , where learned behaviors transmitted culturally become genetically assimilated over generations, accelerating adaptation. In language contexts, this process may have genetically encoded predispositions for acquiring universal grammatical features after initial cultural transmission refined them across populations. Recent 2020s comparative neuroimaging underscores proto-language precursors in great apes, revealing experience-dependent plasticity in language-trained individuals, such as human-like asymmetries in the superior longitudinal fasciculus and expansions in 44 (a Broca's homolog), suggesting foundational networks for that prefigure human capabilities. For instance, studies using diffusion tensor on chimpanzees and bonobos trained in lexigrams show enhanced connectivity in auditory-motor pathways, indicating neural adaptations akin to early human language hubs. Ontogenetically, language processing develops through sequential stages in children, leveraging high neural plasticity during early life. Preverbal communication begins with gestures and cooing from birth to 6 months, transitioning to canonical babbling around 6–12 months, where reveals activation of dorsal stream components, including motor regions in and the , as infants link auditory to vocal production. This progresses to holophrastic single-word utterances (12–18 months), two-word combinations (18–24 months), and full acquisition by 3–4 years, with expanding and grammatical reflecting maturation of frontotemporal connectivity. The asserts that this plasticity peaks in childhood, declining sharply after age 7–8, as demonstrated by studies on late second-language learners, such as Johnson and Newport (1989), who found that ultimate attainment in L2 grammar declines after age ~7, with immigrants exposed after this age showing non-native-like proficiency despite immersion, highlighting enduring neural constraints on .

Multimodal Language Processing

Sign Language Processing

Sign language processing in the brain relies on visual-manual modalities rather than auditory-oral ones, yet it recruits homologous neural regions to those involved in , demonstrating significant overlap in core linguistic functions. Deaf signers activate left perisylvian areas, including homologs of Broca's and Wernicke's areas, for processing sign syntax and semantics, such as grammatical structure and meaning integration. studies confirm that these regions support combinatorial processing in sign languages like (ASL), mirroring their role in spoken syntax. The visual ventral stream, spanning occipital-temporal regions, plays a crucial role in sign recognition by discriminating linguistic handshapes and movements from non-linguistic gestures, with early neural tuning observed around 80-120 ms post-stimulus. Additionally, the bilateral integrates manual signs with non-manual features, such as expressions and head movements, essential for conveying prosody and in sign languages. fMRI evidence from deaf signers viewing signs shows activation patterns in bilateral fronto-temporal networks (left-dominant) that closely resemble those during comprehension in hearing individuals, underscoring shared neural substrates for linguistic decoding across modalities. A modality-independent core language network, involving left anterior temporal lobe and , supports universal grammatical computations in both sign and speech, suggesting that abstract linguistic principles operate beyond sensory input. Disruptions from left hemisphere lesions produce sign aphasias that parallel spoken types, including phonological errors (e.g., handshape substitutions) and (telegraphic signing), as documented in case studies of deaf signers with focal brain damage. Recent studies highlight neural plasticity in late learners of , even into adulthood. For instance, hearing adults undergoing intensive ASL training over eight months exhibit dynamic reorganization, with increased activation in left perisylvian and visual processing areas, alongside shifts in functional connectivity that adapt to linguistic demands. Short-term training in late learners also recruits frontal and parietal regions similarly to native signers, indicating compensatory plasticity that enhances comprehension despite delayed exposure. These findings emphasize the brain's adaptability, though late acquisition may involve greater reliance on visual and right-hemisphere resources compared to early bilinguals.

Written Language Processing

Written language processing involves the neural mechanisms that enable the recognition, comprehension, and production of orthographic symbols, distinct from pathways. Reading primarily engages the ventral occipito-temporal cortex, where visual input is transformed into meaningful linguistic representations, while writing recruits motor planning areas to generate graphemic output. These processes rely on specialized regions that develop through acquisition and exhibit variations across writing systems. A key structure in reading is the (VWFA), located in the left , which supports invariant recognition of words regardless of font, case, or size. This region becomes tuned to orthographic forms during reading development, facilitating rapid word identification by enhancing sensitivity to recurring visual patterns in print. studies confirm the VWFA's role in abstract orthographic processing, with activation peaking around 150-200 milliseconds post-stimulus onset. The dual-route model describes two primary pathways for reading: the lexical route, which accesses stored word representations directly via the VWFA for familiar words, and the sublexical route, which assembles pronunciations through grapheme-to-phoneme conversion for unfamiliar or pseudowords. The lexical path involves occipito-temporal regions linking to semantics, while the sublexical path engages superior temporal and parietal areas for phonological mapping. evidence supports this framework, showing distinct activation patterns for exception words (lexical) versus regular nonwords (sublexical) during reading aloud. Grapheme-to-phoneme conversion, central to the sublexical route, implicates the in integrating visual letter forms with phonological codes, often within 100-200 milliseconds of visual presentation. This process is modulated by , with shallower systems (e.g., consistent letter-sound mappings) relying more on posterior superior temporal activation. Disruptions in these pathways, such as reduced integrity of the left arcuate fasciculus—a tract connecting temporal and frontal regions—are linked to , impairing phonological decoding and reading fluency across languages. Writing engages Exner's area in the left (posterior , ), which stores motor programs for production and coordinates hand movements during script generation. Activation in this region increases with visual letter presentation, suggesting it integrates orthographic planning with motor execution. In , arcuate fasciculus anomalies further compromise writing by disrupting connections between phonological and motor areas, leading to graphemic errors. Cross-linguistic variations highlight script-specific neural adaptations: alphabetic systems emphasize grapheme-to-phoneme mapping in superior temporal regions, whereas logographic scripts like Chinese recruit additional left middle frontal and inferior parietal areas for visual-semantic processing of characters. Bilingual studies show overlapping ventral occipito-temporal involvement but greater selectivity for logographs due to their morphological complexity. Recent fMRI and eye-tracking research from the 2020s reveals in ventral occipito-temporal areas during reading, where prior linguistic anticipates upcoming words, modulating fixation durations and VWFA responses to enhance efficiency.

References

Add your contribution
Related Hubs
User Avatar
No comments yet.