Fact-checked by Grok 2 weeks ago

Sentence processing

Sentence processing is a subfield of that investigates how humans comprehend and interpret sentences in by incrementally integrating lexical, syntactic, semantic, prosodic, and contextual to derive meaning. This process unfolds rapidly as words are encountered, often involving the resolution of ambiguities at multiple levels, such as structural (e.g., garden-path sentences) or lexical (e.g., homophones), to construct a coherent representation of the intended message. Central to the field are competing theoretical models that explain how syntactic dependencies—links between elements like verbs and arguments or pronouns and antecedents—are formed during comprehension. The influential garden-path model, developed by Frazier and Fodor, proposes a serial, two-stage process where an initial parse relies on heuristics such as minimal attachment (preferring simpler structures) and late closure (attaching new material to the most recent ), leading to reanalysis upon encountering disambiguating evidence. In contrast, constraint-based models advocate for parallel evaluation of multiple interpretations, where probabilistic constraints from semantics, discourse, and world knowledge modulate activation levels to favor the most plausible parse. More recent frameworks, including good-enough processing and noisy channel accounts, suggest that comprehenders may settle for approximate or underspecified representations rather than exhaustive analysis, especially under or in noisy environments. Experimental methods have advanced understanding of these mechanisms, revealing the incremental and predictive nature of processing. Eye-tracking s, such as the visual world , capture fixations during spoken or to index attention shifts and prediction effects, as in anticipatory looks to objects matching expected sentence continuations. Event-related potentials (ERPs) provide neural evidence, with components like the N400 signaling semantic integration difficulties and the P600 indicating syntactic reanalysis. Self-paced reading tasks measure word-by-word latencies, highlighting effects in , such as attraction across languages. Contemporary research emphasizes individual differences, developmental trajectories, and cross-linguistic variations, showing that processing is modulated by factors like capacity and typology. For instance, cue-based retrieval theories explain how interference affects dependency formation similarly in diverse languages, from English subject-verb to Mandarin reflexives, though modulated by cues like case marking. Recent work as of 2025 has explored how large models mimic human sentence processing patterns, providing new insights into predictive mechanisms. Ongoing debates center on the immediacy of contextual integration and the role of in facilitating efficient , with implications for and computational modeling.

Introduction

Definition and scope

Sentence processing refers to the cognitive mechanisms by which humans incrementally linguistic input—whether spoken or written—to construct mental representations of sentence meaning in . This process involves the dynamic interplay of syntactic structure-building, semantic , and pragmatic , drawing on lexical , contextual cues, and predictive expectations to resolve the relations among words. Unlike isolated , sentence processing emphasizes the integration of multiple words into coherent syntactic and semantic structures, going beyond mere lexical retrieval to form propositional representations. The scope of sentence processing research primarily encompasses in both auditory and visual modalities, excluding , and centers on human cognitive processes while noting conceptual parallels in and systems. These parallels arise from shared principles, such as and probabilistic inference, observed in both human brains and autoregressive models during naturalistic language tasks. Foundational influences, including Noam Chomsky's , have shaped early psycholinguistic models by providing formal frameworks for syntactic analysis in . Understanding sentence processing is crucial for insights into , where children rapidly develop abilities to parse complex structures online; neurological disorders like , which impair syntactic integration and lead to comprehension deficits; and applications in , where mimicking human-like incremental parsing enhances in AI systems.

Historical overview

The study of sentence processing originated in the mid-20th century, drawing heavily from linguistic theory. In the 1950s and 1960s, Noam Chomsky's (1957) introduced transformational-generative grammar, emphasizing hierarchical syntactic structures and innate language mechanisms, which profoundly shaped early models of how humans parse sentences by proposing rules for generating and interpreting syntactic forms. This framework shifted focus from behaviorist views to mentalistic accounts, influencing to explore as a rule-based computational process. The 1970s marked the rise of as a distinct field, with experimental investigations into real-time comprehension. A pivotal development was the garden-path theory proposed by Lyn Frazier and Janet Dean Fodor in their 1978 "sausage machine" model, which posited serial syntactic processing where initial parses based on minimal attachments lead to temporary misanalyses that require reanalysis, highlighting the incremental nature of sentence interpretation. This theory emerged amid broader psycholinguistic efforts to test linguistic predictions empirically, using as a key testing ground for competing theories since the decade's start. In the and , the field expanded beyond strictly serial models to incorporate interactive and connectionist approaches. Interactive models, such as those advanced by James L. McClelland, Mark F. St. John, and Roman Taraban in 1989, integrated multiple sources of information (syntactic, semantic, and contextual) in parallel, challenging modularity by demonstrating bidirectional influences during comprehension through parallel distributed processing simulations. Concurrently, connectionist networks gained traction for modeling emergent processing behaviors without explicit rules. From the 2000s onward, sentence processing research underwent a neurocognitive transformation with the integration of and computational tools. Key milestones included the first event-related potential (ERP) studies on syntactic violations by Helen J. Neville and colleagues in 1991, revealing distinct brain responses (e.g., early left anterior negativity) to syntactic anomalies, thus linking electrophysiological measures to processing stages. This was followed by a surge in eye-tracking applications post-2000, enabled by advances in technology, which provided high-resolution data on reading times and regressions to probe incremental in natural contexts.

Fundamental Processes

Lexical access

Lexical access refers to the initial retrieval of stored word representations from the during , triggered by bottom-up activation from sensory input. In , phonological input activates a set of candidate words sharing initial sounds, while in written language, orthographic input engages letter-based features to retrieve corresponding lexical entries. This process rapidly accesses semantic meanings, syntactic categories, and other properties essential for further integration, occurring within hundreds of milliseconds of word onset. Key models describe these dynamics differently across modalities. The cohort model posits that spoken word recognition begins with bottom-up activation of all words matching the initial phonological cohort, followed by progressive narrowing as more input eliminates competitors through feature matching. This serial activation emphasizes early, exhaustive access before selection. For visual word recognition, the interactive activation model proposes across orthographic, phonological, and semantic levels, where bottom-up input from letters spreads activation bidirectionally, allowing contextual influences to modulate early stages. Both frameworks highlight how sensory-driven activation retrieves core lexical features, setting the stage for syntactic . Word frequency profoundly influences access speed, with high-frequency words eliciting faster recognition than low-frequency ones, as evidenced by shorter response times in lexical decision tasks. This effect arises from stronger resting activation levels for frequent items in the , reducing retrieval time and facilitating smoother incorporation into ongoing sentence processing. Sentence context exerts immediate effects through lexical priming, where prior words activate related candidates, speeding access to predictable targets. For instance, semantically related primes facilitate recognition of upcoming words, demonstrating rapid, automatic within the that aids predictive processing. Challenges arise with lexical ambiguities like homophones (e.g., "" as river edge or ) and polysemous words (e.g., "" as material or document), where multiple representations activate simultaneously at the lexical stage. For homophones, initial phonological input triggers exhaustive access to all matches before contextual selection, while polysemy involves competition among related senses, often resolved through semantic overlap rather than complete inhibition. These ambiguities underscore the lexicon's organization, where frequency and relatedness modulate resolution efficiency.

Syntactic parsing

Syntactic parsing involves the construction of hierarchical phrase structures from sequences of words, utilizing syntactic categories and frames provided by lexical access to organize phrases and clauses in . This process operates incrementally, building partial syntactic representations as each word is encountered, rather than waiting for the entire to be available. Key parsing strategies guide the attachment of incoming words to existing structures. The late closure principle favors attaching new constituents to the most recently processed phrase, promoting right-branching analyses, while the minimal attachment principle prefers the simplest syntactic tree with the fewest nodes and attachments. These strategies, proposed in the garden-path model, minimize computational load during initial but can lead to errors requiring later revision. For instance, in the ambiguous sentence "The dean admired the student of the teacher with the ," both minimal attachment and late closure initially favor linking "with the " to "the teacher," interpreting the teacher as using or possessing the , but this may require reanalysis if context indicates it modifies the student or the verb. Incremental processing is often head-driven, where heads such as project their structures early, constraining the possible attachments for subsequent dependents and facilitating rapid building. In this approach, the parser uses the verb's information to anticipate and integrate arguments as they appear, ensuring connectedness in partial parses. Garden-path phenomena illustrate the consequences of these strategies, where an initial parse based on late closure or minimal attachment proves incompatible with later input, necessitating reanalysis and increased processing effort. A classic example is "The horse raced past the barn fell," where "raced" is initially parsed as the main (), but "fell" forces reanalysis to a reduced ("raced" as past participle), triggering a temporary misparse. Such effects have been observed in reading times and eye movements, confirming the prevalence of these attachment preferences. Cross-linguistic variations arise from head-directionality parameters, influencing parsing efficiency and strategies. In head-initial languages like English, where heads precede complements, parsing proceeds left-to-right with heads driving attachments forward. In contrast, head-final languages like require pre-head attachments, where arguments are incrementally integrated before the arrives, as evidenced by faster reading times for compatible structures in self-paced experiments. This adaptation maintains incrementality but modulates reliance on predictive mechanisms. Syntactic anomalies, such as phrase structure violations or unexpected attachments, elicit error signals detectable via event-related potentials. The P600 component, a positive wave peaking around 600 ms post-stimulus, indexes syntactic reanalysis or integration difficulties, distinguishing it from semantic processing markers. For example, ungrammatical sentences like "The man that the nurse kissed * the doctor" evoke a robust P600, reflecting during continuous speech or reading.

Semantic and pragmatic integration

Semantic processes in sentence processing involve the compositional construction of meaning from individual lexical items, guided by principles of function-argument application often formalized using . In this framework, verbs are treated as functions that take arguments to yield propositions; for instance, the verb "chase" can be represented as λx.λy.chase(y,x), where applying it to an object like "the cat" yields λy.chase(y, the cat), and further application to a subject like "the dog" results in chase(the dog, the cat). This bottom-up merging ensures that the meaning of a complex expression is derived systematically from the meanings of its parts, as outlined in formal semantic theories adapted to processing models. Pragmatic layers enrich this semantic composition by incorporating contextual inferences and speaker intentions, particularly through implicature resolution based on Gricean maxims of , which include quantity (provide as much information as required), quality (be truthful), relation (be relevant), and manner (be clear). For example, the sentence "Some students passed" implicates that not all did, via the maxim of quantity, assuming cooperative communication. Additionally, reference assignment handles anaphora, such as linking "he" to a prior antecedent like "," drawing on discourse context and recency to resolve during incremental processing. Integration of semantics and occurs in with syntactic in interactive models, where multiple sources of constrain simultaneously, rather than strictly serially. Semantic mismatches, such as unexpected words in (e.g., "He spread the warm bread with socks"), elicit the N400 , a negative wave peaking around 400 ms post-stimulus, indexing the effort to integrate incongruent meanings. This component arises during both and , reflecting a unified mechanism for semantic . At the discourse level, is built by managing given-new , where given elements (previously mentioned or assumed known) are placed early in sentences for activation, and new follows to advance the , facilitating efficient . Syntactic frames from prior serve as scaffolds, providing slots that guide semantic filling, such as theta roles for arguments. Predictive elements further shape , with top-down expectations based on modulating ease, as measured in cloze probability tasks where participants complete fragments, revealing how likely a word is given preceding material. High cloze words (e.g., "" in "He spread the warm _____") reduce integration costs, while low-probability ones increase them, supporting accounts where the brain anticipates upcoming semantics to optimize .

Ambiguity and Resolution

Types of ambiguity

In sentence processing, ambiguity arises when linguistic input allows for multiple interpretations, challenging the comprehension system to select the appropriate meaning or structure. These ambiguities can occur at different levels of representation, including lexical, syntactic, semantic, and anaphoric, each presenting unique processing demands. Lexical ambiguities stem from words with multiple meanings, while syntactic ones involve alternative phrase structures. Semantic ambiguities concern interpretive scope, and anaphoric ambiguities involve reference for pronouns or other referring expressions. Such ambiguities are prevalent in , with temporary ambiguities (resolved later in the sentence) often leading to processing delays compared to global ambiguities (persistent throughout). Lexical ambiguity occurs when a single word form maps to multiple meanings, such as homonyms (unrelated meanings, e.g., "" referring to a or a river edge) or polysemous words (related senses, e.g., "" as a or its ). Upon encountering an ambiguous word, all possible meanings are initially activated regardless of prior context, as demonstrated in cross-modal priming studies where targets related to both dominant and subordinate meanings elicit facilitation immediately after the ambiguous word but not 200-500 milliseconds later. This rapid, exhaustive activation supports modular models of lexical access, where selection occurs post-access based on contextual fit. For instance, in the sentence "The had to the funds," "" could evoke striking a light or pairing resources, with the dominant meaning (pairing) dominating due to higher frequency. Distinctions between homonymy and influence processing, as polysemous senses share more semantic overlap and decay slower than unrelated homonym meanings. Syntactic ambiguity, also termed structural ambiguity, arises from multiple possible grammatical parses of a phrase or sentence, often exposed during parsing when initial commitments lead to revisions. A classic subtype is prepositional phrase (PP) attachment ambiguity, where a PP can modify the verb or the preceding noun phrase (NP), as in "I saw the man with the telescope," interpretable as using a telescope to see or seeing a man holding one. Another form is late closure ambiguity, where a new phrase can attach to the most recent constituent, contributing to garden path effects exemplified by "The horse raced past the barn fell," initially parsed with "raced" as the main verb but requiring reanalysis to a reduced relative clause upon encountering "fell." Phrase structure ambiguities, such as between NP and PP interpretations (e.g., "old men and women" as old (men and women) or (old men) and women), further illustrate how minimal attachment principles—favoring simpler structures—guide initial parsing. These ambiguities highlight parsing's role in incrementally building representations, with temporary syntactic ambiguities frequently causing "garden path" effects. Semantic ambiguity involves multiple plausible meanings at the level of overall sentence interpretation, often due to scope interactions among operators like quantifiers. Quantifier ambiguity occurs when the relative order of quantification is unclear, as in "Every kid climbed a tree," which can mean each kid climbed some tree (wide scope for "every") or there exists one tree that all kids climbed (wide scope for "a"). Processing favors surface order (e.g., universal quantifiers taking wide scope when preceding existentials), influenced by structural positions rather than purely semantic preferences. Another example is "Every farmer who owns a donkey beats it," where "it" introduces anaphoric ties but also scope issues between the universal "every" and the donkey's reference. These ambiguities require integrating with , differing from syntactic types by persisting beyond structural resolution. Anaphoric ambiguity emerges when a referring expression, such as a , has multiple potential antecedents in the , complicating assignment. For example, in "John hit Bill, and then he apologized," "he" could refer to , , or an unmentioned , with guided by syntactic parallelism, recency, and prominence. Sentences with multiple clause structures exacerbate this, as in "The athlete admired the coach and he was talented," where "he" ambiguously links to either the or coach. Processing involves computing prominence profiles across candidates, with ambiguities more pronounced in complex sentences featuring crossed dependencies. Unlike lexical types, anaphoric ambiguities rely on -level , often leading to delayed in reading times. Temporary ambiguities, resolvable upon further input, dominate experimental findings (e.g., PP attachments), while global ones (e.g., quantifier scopes) affect entire sentences and are rarer in corpora but critical for logical inference. Both types underscore the incremental nature of .

Resolution strategies

Resolution strategies in sentence processing rely on cognitive that prioritize efficiency during real-time disambiguation, particularly for syntactic ambiguities such as attachment preferences in relative clauses. One key heuristic is frequency-based preference, where more common structures, like subject-relative clauses (e.g., "The visited the patient was kind") over object-relative clauses (e.g., "The patient who the doctor visited was kind"), are initially favored due to their higher occurrence in language use. Another foundational heuristic is recency, embodied in the late closure principle, which attaches incoming material to the most recent open or to minimize structural complexity (e.g., interpreting "I saw with the " as using the telescope to see the man). Resolution often integrates multiple cues for robust disambiguation. Lexical cues, such as verb biases toward specific syntactic frames (e.g., "used" biasing toward an instrument phrase like "She used the spoon to eat the soup" rather than a modifier), guide initial parse decisions rapidly. Prosodic cues, including intonation contours and pauses, further aid resolution by signaling phrase boundaries (e.g., a pause after "the man" in "I saw the man, with the telescope" promoting low attachment). World knowledge integration supplements these by evaluating plausibility against general expectations, such as thematic roles or event schemas, to select viable interpretations. When initial parses fail, reanalysis incurs processing costs that vary by theoretical approach. In serial models, fully revises the committed structure, leading to higher disruption in garden-path sentences (e.g., "The horse raced past the barn fell"). Parallel models, conversely, employ soft commitments to multiple analyses, distributing costs through probabilistic weighting rather than wholesale revision. Individual differences modulate resolution efficacy, with working memory capacity determining the ability to maintain and compare alternative parses. Higher-capacity individuals better accommodate low-frequency or conflicting structures, reducing error rates in complex ambiguities. Cross-linguistically, strategy effectiveness depends on language typology and cue reliability. In , overt case marking on nouns and determiners reliably disambiguates subject-object order (e.g., accusative marking signals an object-first structure), overriding recency biases more effectively than in English.

Theoretical Frameworks

Architectural issues

Architectural issues in sentence processing center on the high-level organization of cognitive mechanisms that enable comprehension, particularly the extent to which processing is modular or interactive and serial or parallel. Modular architectures, inspired by framework, propose that sentence processing involves domain-specific modules that operate autonomously and are informationally encapsulated, preventing influences from non-linguistic knowledge during core computations such as syntactic . In this view, an independent syntax module computes structural representations without input from semantics or , ensuring efficient, rule-based processing insulated from broader contextual factors. This Fodorian modularity emphasizes vertical organization, where lower-level perceptual modules feed into higher ones without horizontal interactions across domains. In opposition, interactive architectures advocate for bidirectional flow of information between processing levels, allowing semantic and contextual cues to influence syntactic decisions from the outset. Pioneered in models of reading and , this approach posits that spreads across levels simultaneously, enabling top-down expectations to constrain bottom-up analyses and vice versa. A related dimension involves serial versus : serial designs, often tied to , sequence operations such that syntactic structure is fully built before semantic integration occurs, minimizing computational load but potentially delaying context effects. , characteristic of interactive systems, activates multiple structural hypotheses concurrently, facilitating competition and integration across levels in real time. Debates over have long pitted the hypothesis, which upholds syntactic and predicts insulation from non-syntactic influences, against interactive theories that highlight rapid contextual modulation of early processing stages. These positions carry distinct implications: anticipates delayed semantic contributions, leading to initial parses that may require later revision, whereas supports immediate exploitation of context to streamline comprehension and reduce . Recent critiques have challenged strict dichotomies, favoring architectures that integrate modular for core computations with predictive mechanisms allowing limited , thus accommodating of both encapsulation and contextual sensitivity across processing stages.

Processing models

Sentence processing models aim to explain how humans incrementally interpret linguistic input, integrating syntactic, semantic, and contextual information to construct meaning. These models vary in their assumptions about the architecture of , such as whether processing is or , and how ambiguities are resolved. Influential frameworks include models emphasizing syntactic primacy and reanalysis, models incorporating multiple probabilistic constraints, and more recent approaches highlighting strategies and predictive mechanisms. Each model makes distinct predictions about processing difficulty, often tested through reading times and error rates in ambiguous sentences. The garden-path model, proposed by Frazier and Fodor (1978), posits a serial, syntax-first approach to parsing, where comprehenders initially adopt the simplest syntactic structure consistent with minimal attachment principles, such as attaching new phrases to the lowest possible node in the existing parse tree. If this leads to an incompatible interpretation upon encountering disambiguating information, reanalysis occurs, incurring processing costs manifested as longer reading times or regressions. Frazier and Rayner (1982) provided eye-tracking evidence from structurally ambiguous sentences, like "The horse raced past the barn fell," where the initial misparse of "raced" as the main verb creates a garden-path effect, predicting difficulty from the need for syntactic revision rather than semantic influences during initial parsing. The model highlights disruptions from temporary misparses but assumes rapid recovery through reanalysis. In contrast, the constraint-based model advocates parallel evaluation of multiple sources of information, including syntactic biases, , and , which compete probabilistically to determine the most likely . MacDonald, Pearlmutter, and Seidenberg (1994) argued that ambiguity resolution emerges from the interaction of these constraints, with lexical frequencies shaping preferences; for instance, verbs with higher frequency in reduced constructions, such as "headed," bias toward that interpretation in sentences like "The headed by the professor was informative," due to higher probabilistic of subordinate structures. This framework uses connectionist-inspired mechanisms for ranking alternatives, predicting smoother processing when constraints align and difficulty only when competing interpretations are equally probable, thus avoiding the discrete reanalysis of models. The good-enough processing account, proposed by Ferreira et al. (2002), extends constraint-based ideas by emphasizing shallow, interpretations that prioritize rapid comprehension over exhaustive syntactic detail, allowing errors in complex structures if the overall meaning suffices. and Christiansen (2008) linked this to experience-based statistical learning, where frequent exposure leads individuals to adopt strategies, such as inferring agent-patient roles without full reanalysis in passive sentences like "The trophy was won by the yelling fan," resulting in error-prone but efficient understanding. This model explains persistent misinterpretations in garden-path sentences and variability across listeners, attributing it to individual differences in linguistic experience rather than parsing principles. Recent developments incorporate predictive processing, where comprehenders actively anticipate upcoming input using to update beliefs based on prior context, minimizing uncertainty. Kuperberg and Jaeger (2016) described this as probabilistic prediction across multiple representational levels, from phonemes to semantics, enabling preemptive resolution of ambiguities; for example, a verb's frame might bias expectations for object types, facilitating integration. Complementing this, the surprisal metric quantifies processing difficulty as the negative log probability of an element given its context, formally defined as \text{surprisal} = -\log P(\text{parse} \mid \text{previous}) Hale (2001) introduced surprisal in a probabilistic framework, predicting that low-probability continuations, like rare syntactic attachments, increase , as evidenced by slower reading times in surprisal-correlated experiments. This metric integrates well with predictive models by linking anticipation to information-theoretic efficiency. Comparisons among these models reveal trade-offs: the garden-path approach offers a minimalist mechanism for syntactic primacy but underemphasizes integrative constraints, while constraint-based and good-enough models provide richer, probabilistic accounts yet may overlook rapid reanalysis in clear cases. Predictive extensions address these by incorporating , though all frameworks share limitations in accounting for individual differences, such as varying capacities or language exposure, which modulate model predictions. Frazier and Fodor (1978); MacDonald et al. (1994); Ferreira et al. (2002); MacDonald and Christiansen (2008).

Research Methods

Behavioral techniques

Behavioral techniques in sentence processing involve non-invasive methods that assess through participants' explicit responses, such as reaction times or judgments, in controlled experimental settings. These approaches provide insights into the speed and accuracy of processing by measuring how individuals handle linguistic structures under varying conditions. Self-paced reading presents sentences word-by-word or in segments on a computer screen, with participants advancing at their own by pressing a key, allowing researchers to record reading times for each segment as an index of processing difficulty. Longer reading times at specific words or phrases indicate higher , such as during syntactic integration or . This technique was pioneered in studies examining processes, where it revealed patterns like increased times for object-relative clauses compared to subject-relative clauses. Acceptability judgments require participants to rate sentences on scales of grammaticality, naturalness, or plausibility, probing implicit linguistic intuitions without time pressure. These ratings help identify subtle effects of structure on perceived well-formedness, such as gradient acceptability in complex embeddings. Methodological advancements have emphasized objective scaling to minimize bias, ensuring reliable measurement of processing preferences. Maze tasks force incremental processing by displaying the sentence with the next word masked and offering multiple grammatical continuations for rapid selection via keypress, capturing decision times at each step. This method highlights momentary processing commitments, such as faster selections for predictable continuations, and serves as an alternative to traditional reading paradigms by reducing visual masking artifacts. Lexical decision in embeds a target word within a sentence, asking participants to quickly judge if it is a real word while prior influences response speed. Facilitation from semantic or syntactic fit demonstrates effects, with faster decisions for contextually congruent words. This task has been instrumental in showing how sentence-level constraints modulate lexical . These techniques offer high for tracking processing increments and are cost-effective for large samples, though they may introduce metalinguistic awareness that alters natural comprehension. For instance, self-paced reading and tasks have been applied to study ambiguity resolution, revealing reanalysis costs in garden-path sentences.

Eye-tracking studies

Eye-tracking studies offer a window into the dynamics of sentence processing by monitoring oculomotor responses, such as fixations and saccades, during reading or auditory comprehension tasks. These measures provide online indices of cognitive effort, revealing how readers or listeners incrementally build interpretations without relying on reports. Unlike offline behavioral techniques, which capture end-of-trial responses, eye-tracking captures implicit processing fluctuations on a scale. Central metrics in these studies include first-pass fixation duration, which quantifies the time spent initially processing a word or region before moving forward, indexing early syntactic and semantic integration; regressions, backward eye movements signaling reanalysis or comprehension breakdowns; and go-past times, encompassing the total duration from first fixation to leaving a region forward, reflecting overall processing difficulty. For instance, increased first-pass durations often mark encounters with unexpected , while regressions highlight recovery from misanalyses. In reading paradigms, eye movements are recorded as participants silently read sentences displayed on a screen, allowing precise localization of processing costs to specific words. A key technique is the , where an invisible boundary precedes a target word; upon the eyes crossing it, the display changes from a preview (e.g., a nonword or low-frequency word) to the actual target, isolating parafoveal preprocessing effects on fixation times. This method has demonstrated that readers extract orthographic and phonological information from upcoming words before direct fixation, influencing initial processing efficiency. Modern , sampling at 1000 Hz, enable high temporal resolution for these analyses, minimizing data loss from blinks or minor head movements. The visual world paradigm extends eye-tracking to comprehension, where participants view a visual containing objects compatible with sentence referents while listening to instructions. Eye movements to depicted objects trace the time course of reference and thematic role assignment, showing how linguistic input guides looks incrementally. Seminal work using this paradigm revealed that visual contexts modulate syntactic ambiguity from the outset, as preferentially fixate contextually appropriate referents during spoken like "Put the apple on the towel," avoiding implausible alternatives. Key findings from eye-tracking underscore processing challenges in ambiguous sentences, where longer fixations and elevated regression rates occur at disambiguating regions in , such as "The horse raced past the barn fell," indicating initial commitment to a misparse followed by reanalysis. Predictive processing is evident in anticipatory looks driven by verb biases; for example, upon hearing "The boy will eat...," listeners fixate a depicted cake more than a toy car, reflecting rapid use of selectional restrictions to forecast upcoming arguments before they are named. These patterns hold across languages and populations, though individual differences in reading skill modulate fixation durations. Recent advances in mobile eye-tracking, including wearable devices and webcam-based systems, have facilitated investigations in naturalistic settings beyond controlled lab environments. Post-2020 studies have replicated core effects like increased fixations during resolution in everyday reading scenarios, such as news articles or conversations, demonstrating the method's portability while maintaining sensitivity to processing dynamics. These tools complement traditional lab-based behavioral measures by enabling ecologically valid on unconstrained .

Neuroimaging and electrophysiological methods

Neuroimaging and electrophysiological methods provide critical insights into the neural underpinnings of sentence processing by offering spatial and of activity during and . These techniques reveal how linguistic elements such as and semantics engage specific regions and unfold over milliseconds, complementing behavioral measures like eye-tracking for validation. Functional magnetic resonance imaging (fMRI) excels in spatial localization, identifying key language areas like Broca's area (in the inferior frontal gyrus) for syntactic processing and Wernicke's area (in the superior temporal gyrus) for semantic processing during sentence comprehension. Studies using blood-oxygen-level-dependent (BOLD) responses demonstrate increased activation in these regions with rising syntactic complexity, such as in object-relative clauses compared to subject-relative clauses, highlighting Broca's role in unifying syntactic and semantic constraints. Semantic processing in Broca's area is also evidenced by fMRI activations during lexical integration in sentences, extending beyond purely syntactic functions. Electroencephalography (EEG) and event-related potentials (ERPs) offer high temporal precision, capturing neural responses on the order of milliseconds during sentence processing. The N400 component, peaking around 400 ms post-stimulus, reflects semantic integration difficulties, such as when a word violates contextual expectations in a sentence (e.g., "The Dutch trains are sour and full of milk"). For syntactic processing, the left anterior negativity (LAN, ~300-500 ms) indicates early phrase structure violations, while the P600 (~600 ms) signals later reanalysis or repair of syntactic anomalies, like garden-path sentences. These components map to a cortical network involving temporal and frontal regions, with the N400 linked to lexical retrieval and the P600 to compositional integration. Magnetoencephalography (MEG) combines temporal resolution with source localization, elucidating predictive mechanisms in sentence comprehension. MEG reveals hierarchical predictive coding across fronto-temporal networks, where early sensory areas anticipate semantic content and higher-order regions integrate syntactic structure incrementally from left to right. For instance, source-localized MEG activity shows pre-activation in temporal lobes for expected words, supporting top-down predictions that facilitate processing efficiency. Recent advances from 2020 to 2025 have leveraged (ECoG) for high spatiotemporal resolution in decoding sentence-level representations, often in clinical settings with patients. ECoG recordings demonstrate shared cortical word representations across and tasks within the , enabling decoding of semantic content from neural signals during naturalistic narrative processing. For example, multivariate pattern analysis of ECoG data has reconstructed syntactic roles and word meanings in real-time sentence , revealing a division of labor where ventral temporal areas handle semantics and dorsal frontal areas manage . Datasets like the "" ECoG collection support modeling of narrative , showing stable decoding of linguistic features over extended discourse. Bibliometric analyses indicate a steady increase in publications on processing in recent years, emphasizing and decoding approaches. Despite these strengths, these methods have notable limitations. fMRI's reliance on hemodynamic responses introduces a 4-6 second lag, obscuring fine-grained temporal dynamics of sentence processing. EEG and MEG suffer from poorer due to volume conduction and challenges, respectively, while ECoG is constrained by ethical and practical issues of invasiveness, limiting its use to surgical populations.

Computational modeling

Computational modeling in sentence processing employs simulation-based approaches to test and refine theoretical predictions about how humans incrementally interpret linguistic input. These models generate quantifiable predictions, such as processing difficulty metrics, that can be compared against human behavioral data to evaluate the plausibility of underlying mechanisms. By implementing rule-based or neural architectures, researchers simulate the of syntactic and semantic integration, allowing for the exploration of factors like ambiguity resolution and in controlled computational environments. Rule-based parsers, such as shift-reduce algorithms, provide a foundational method for syntactic analysis by incrementally building parse trees through operations like shifting input words onto a stack or reducing partial structures according to grammar rules. These parsers model human-like incremental processing by prioritizing efficient derivations, as seen in early disambiguation techniques that mimic native speaker preferences for certain syntactic attachments. A seminal example is Hale's 2001 surprisal model, which uses a probabilistic Earley parser to compute surprisal as the negative log probability of a word given its preceding context, \text{surprisal}(w_i) = -\log P(w_i \mid w_{1:i-1}), linking parsing effort to information-theoretic unexpectedness and predicting increased reading times for low-probability continuations. Connectionist networks extend these capabilities by learning incremental predictions from data, with recurrent neural networks (RNNs) capturing sequential dependencies to simulate how influences word-by-word comprehension. RNNs, including variants, process sentences autoregressively, generating hidden states that encode syntactic and semantic expectations, thereby modeling effects like garden-path through error signals in prediction. Post-2017, transformer-based models have advanced modeling via self-attention mechanisms, enabling parallel computation of long-range dependencies and better simulation of human-like holistic representation without recurrent bottlenecks. These architectures draw brief inspiration from constraint-based theories by integrating multiple probabilistic cues during forward passes. Evaluation of these models focuses on fitting simulated outputs to human data, such as regressing model-derived metrics against empirical reading times from eye-tracking studies to assess predictive accuracy. For instance, surprisal from parsers correlates strongly with fixation durations, explaining variance in processing difficulty beyond simple frequency effects. Integration cost, formalized in Gibson's 1998 Dependency Locality Theory, serves as an embedding depth metric, quantifying the cognitive burden of linking dependents across intervening material; models compute this as the number of new entities or syntactic heads in the local , with higher costs predicting slower comprehension for center-embedded structures. Recent advances from 2020 to 2025 integrate psycholinguistic features like surprisal and embedding depth into pipelines, enhancing prediction in tasks such as text simplification. Transformer-derived surprisal, for example, outperforms traditional formulas in forecasting sentence-level reading ease when calibrated to human corpora, achieving correlations up to 0.4 with eye-movement metrics in multilingual settings. Applications of these models include simulating (L2) lexical access, where lexical effects modulate processing through frequency and overlap in bilingual representations. BiLex, a , demonstrates how L2 age of acquisition and exposure alter lexical competition and retrieval.

References

  1. [1]
    [PDF] 1 Title: Sentence Processing * Fernanda Ferreira University of South ...
    By convention, the term "sentence processing" refers to the subfield of psycholinguistics focusing on the interpretation of sentences. A range of ...
  2. [2]
    [PDF] Chapter 18 Sentence Processing Jesse Snedeker and Yi Ting Huang
    Psycholinguistic studies can provide additional evidence about the levels of representation that are employed during language comprehension. An example of this.
  3. [3]
    [PDF] Syntactic dependency formation in sentence processing
    Sep 6, 2023 · In this chapter, we focus narrowly on how comparative psycholinguistic research can inform theories of syntactic dependency formation in ...
  4. [4]
    [PDF] Sentence processing: A tutorial review
    Numerous experimental studies show that lexical preferences influence some stage of language comprehension (e.g. Clifton, Frazier, & Connine, 1984;. Holmes, ...
  5. [5]
    (PDF) Sentence Processing - ResearchGate
    Fillmore, C. J. (1968). The creation of prosody during sentence production. Psychological Review, 100, 233-253. ... introduction to Psycholinguistics and ...
  6. [6]
    [PDF] Introduction to Psycholinguistics Lecture 3: Sentence Processing
    What makes up Psycholinguistics? □ Computational models of the ... ➡ Sentence processing is the means by which the words of an utterance are combined ...
  7. [7]
    Shared computational principles for language processing in humans ...
    Mar 7, 2022 · We provide empirical evidence that the human brain and autoregressive DLMs share three fundamental computational principles as they process the same natural ...
  8. [8]
    Sentence Processing and Syntactic Theory - Wiley Online Library
    Apr 27, 2021 · This chapter explores some ways in which Chomsky's linguistic work has influenced research on one domain of linguistic performance, sentence processing, over ...
  9. [9]
    Full article: Language acquisition and language processing: Finding ...
    Within adult psycholinguistic work on sentence processing, evidence for phonological, lexical, syntactic, and semantic prediction is growing (e.g., Altmann & ...
  10. [10]
    Neurocognitive Recovery of Sentence Processing in Aphasia
    Sentence processing treatment results in improved comprehension and production of complex syntactic structures in chronic agrammatism and generalization to less ...
  11. [11]
    [PDF] History of Psycholinguistics 1 - gerry altmann
    This article reviews the history of psycholinguistics, from the time of the ancient ... Aside from many published articles on the psycholinguistics of sentence ...
  12. [12]
    Sentence comprehension: A parallel distributed processing approach
    This paper reviews conventional sentence comprehension approaches, then describes a parallel distributed processing approach, and shows how it offers different ...
  13. [13]
    (PDF) On Comprehending Sentences: Syntactic Parsing Strategies
    The Minimal Attachment principle, proposed by Frazier (1978 Frazier ( , 1987, assumes a two-stage parsing model in which initial syntactic structures are built ...
  14. [14]
    Grammatical Competence and Parsing Performance, Pritchett
    Pritchett critiques the major performance-based parsing models to argue that the principles of grammar drive ... 206 pages | 58 line drawings | 6 x 9 | © 1992.Missing: head- | Show results with:head-
  15. [15]
    Incremental Pre-head Attachment in Japanese Parsing
    The present study addresses the question of whether structural analyses of verb-arguments are postponed up until the head verb has been processed (head-driven ...
  16. [16]
    Event-related brain potentials elicited by syntactic anomaly
    We evaluate the speculation that the P600 and N400 effects are elicited as a function of anomaly type (syntactic and semantic, respectively). Previous article ...
  17. [17]
    [PDF] Computational Semantics: Lambda Calculus
    Feb 24, 2010 · λ-calculus and compositionality. The semantics of words based on syntactic category. Analysis problem ... sentence or phrase. We'll need to ...
  18. [18]
    Modelling meaning composition from formalism to mechanism - PMC
    Dec 16, 2019 · It is not clear just how semantic composition is in the standard lambda calculus model, for two main reasons. First, recursive function ...<|control11|><|separator|>
  19. [19]
    [PDF] LOGIC AND CONVERSATION*
    It is a commonplace of philosophical logic that there are, or appear to be, divergences in meaning between, on the one hand, at least.
  20. [20]
    ANAPHORA RESOLUTION AND REANALYSIS DURING L2 ...
    Aug 30, 2016 · Reanalysis during L2 Sentence Processing. Resolving linguistic ambiguities requires the ability to revise an initially assigned interpretation ...
  21. [21]
    Sentence Perception as an Interactive Parallel Process - Science
    The restoration of disrupted words to their original form in a sentence shadowing task is dependent upon semantic and syntactic context variables, ...
  22. [22]
    [PDF] Brain Potentials Reflect Semantic Incongruity - Marta Kutas - Cog Sci
    Jan 12, 2006 · We now report, however, that an occasional semantic deviation of a word in a sentence is fol- lowed by a negative brain wave (N400) that is ...
  23. [23]
    (PDF) The notions of information structure - ResearchGate
    Aug 9, 2025 · This article takes stock of the basic notions of Information Structure (IS). It first provides a general characterization of IS follow-ing Chafe (1976)
  24. [24]
    Thinking ahead: The role and roots of prediction in language ...
    N400 responses to the four word types were graded by cloze probability: smallest to the best completions in the strongly constraining contexts, intermediate to ...
  25. [25]
    Ambiguity in sentence processing - ScienceDirect.com
    Ambiguity in sentence processing · Abstract · Section snippets · Syntactic explanations · Multiple constraints in sentence processing · The role of extra-sentential ...Missing: seminal | Show results with:seminal
  26. [26]
    Lexical access during sentence comprehension - ScienceDirect.com
    The effects of prior semantic context upon lexical access during sentence comprehension were examined in two experiments. ... Swinney, et al, 1979. Swinney D., ...Missing: citation | Show results with:citation
  27. [27]
    The Processing of Lexical Ambiguity: Homonymy and Polysemy in ...
    The present study investigated whether these different types of lexical ambiguity are psychologically real.Missing: seminal papers
  28. [28]
    The sausage machine: A new two-stage parsing model
    It is proposed that the human sentence parsing device assigns phrase structure to word strings in two steps.
  29. [29]
    Resolution of quantifier scope ambiguities - ScienceDirect
    This paper investigates structural principles, that is, those which refer to the syntactic or semantic positions of the quantified phrases.
  30. [30]
    Pronoun resolution in two-clause sentences: Effects of ambiguity ...
    The two experiments reported here were designed to study anaphoric pronoun resolution by measuring the effects of syntactic ambiguity, antecedent location, ...
  31. [31]
    Anaphoric Pronouns and the Computation of Prominence Profiles
    Apr 16, 2022 · They are anaphoric, i.e. they inherit their reference from other referring entities. When processing a pronoun, the mental representation of a ...Anaphoric Pronouns And The... · Introduction · Discussion
  32. [32]
    The lexical nature of syntactic ambiguity resolution. - APA PsycNET
    Small, G. Cottrell, & M. K. Tanenhaus (Eds.), Lexical ambiguity resolution: Computational linguistic, and psychological perspectives (pp. 195-228).Missing: homophone | Show results with:homophone
  33. [33]
    Verb-specific constraints in sentence processing: Separating effects ...
    Trueswell (1993) used the verbs from the present experiment as targets in a cross-modal naming study in which the congruence of the verb as a continuation of an ...Missing: pdf | Show results with:pdf
  34. [34]
    The interpretation of prosodic patterns at points of syntactic structure ...
    Abstract. Two experiments explore listener's use of sentence prosody to identify syntactic structure. Beginning portions of synthesized sentences having a ...Missing: seminal paper
  35. [35]
    Dynamic Integration of Pragmatic Expectations and Real-World ...
    Dynamic Integration of Pragmatic Expectations and Real-World Event Knowledge in Syntactic Ambiguity Resolution. 2009. Bicknell, Klinton;; Rohde, Hannah.
  36. [36]
    Working memory constraints on the processing of syntactic ambiguity
    We propose a model that explains how the working-memory capacity of a comprehender can constrain syntactic parsing and thereby affect the processing of ...
  37. [37]
    Cross-linguistic differences and their impact on L2 sentence ...
    Jan 1, 2009 · The results revealed that the L2 German speakers utilized case-marking information and exhibited a subject-preference similar to German native ...
  38. [38]
    [PDF] Fodor, JA (1983). The modularity of mind. Cambridge, MA
    Fodor, J. A. (1983). The modularity of mind. Cambridge, MA: MIT Press. FACULTY PSYCHOLOGY is getting to be respectable again after centuries of hanging ...Missing: URL | Show results with:URL
  39. [39]
    [PDF] On modularity in syntactic processing - Janet Dean Fodor
    Clifton and Frazier (1986) present the results of two experiments in further support of their earlier claim (Frazier, Clifton, & Randall, 1983).
  40. [40]
    [PDF] Toward an Interactive Model of Reading - Amazon S3
    At this level, I am assuming that features are extracted according to the assumptions of the Rumelhart and Siple (1974) model. Moreover,. I am assuming that ...
  41. [41]
    [PDF] An Interactive Activation Model of the Effect of Context in Perception ...
    Thus, for example, processing at the letter level presumably occurs simultaneously with processing at the word level, and with processing at the feature level.
  42. [42]
    [PDF] An Interactive Activation Model of Context Effects in Letter Perception
    In Part I of this paper (McClelland &. Rumelhart, 1981 ), we combined the fun- damental features of the Rumelhart ( 1977) interactive model with the flow-of- ...
  43. [43]
    [PDF] The Modularity of Sentence Processing Reconsidered
    The idea that the sentence processing system is modular has fallen out of fashion. The proposal got off to a promising start with the publication of results ...
  44. [44]
    Eye movements in the analysis of structurally ambiguous sentences
    In accord with the garden-path theory of sentence comprehension, shorter reading times were found for sentences conforming to certain independently ...
  45. [45]
    Experience and Sentence Processing: Statistical Learning and ...
    The results support claims for experience-based individual differences and an important role for statistical learning in sentence comprehension processes.
  46. [46]
    [PDF] A Probabilistic Earley Parser as a Psycholinguistic Model
    In human sentence processing, cognitive load can be defined many ways. This report considers a defini- tion of cognitive load in terms of the total probability.
  47. [47]
    A methodological synthesis of self-paced reading in second ...
    May 17, 2018 · Self-paced reading (SPR) is an online computer-assisted research technique in which participants read sentences, broken into words or segments, ...Missing: seminal | Show results with:seminal
  48. [48]
    Experimental Syntax | SAGE Publications Inc
    6-day delivery 30-day returnsThis original work provides a concise introduction to methods that linguists may use to describe patterns of sentence "acceptability" in speech communities.
  49. [49]
    Online eye tracking and real-time sentence processing
    Aug 1, 2023 · Webcam-based eye tracking is emerging as a powerful online tool capable of capturing sentence processing effects in real time.
  50. [50]
    Eye movements in reading and information processing - PubMed - NIH
    Basic topics discussed with respect to reading are (a) the characteristics of eye movements, (b) the perceptual span, (c) integration of information across ...
  51. [51]
    Integration of Visual and Linguistic Information in Spoken Language ...
    Visual context influenced spoken word recognition and mediated syntactic processing, even during the earliest moments of language processing.
  52. [52]
    Eye movements in the analysis of structurally ambiguous sentences
    Eye movements were recorded as subjects read sentences containing temporary structural ambiguities. In accord with the garden-path theory of sentence ...
  53. [53]
    Incremental interpretation at verbs: restricting the domain ... - PubMed
    The results suggest that information at the verb can be used to restrict the domain within the context to which subsequent reference will be made.Missing: bias tracking
  54. [54]
    Elucidating relations between fMRI, ECoG, and EEG through ... - PMC
    Both are complementary in that fMRI provides high spatial but low temporal resolution, while the opposite is true for EEG. Invasive methods such as ...
  55. [55]
    Vascular responses to syntactic processing: Event‐related fMRI ...
    Functional neuroimaging studies using PET and fMRI are beginning to provide data regarding the localization of syntactic processing in sentence comprehension.
  56. [56]
    Parametric effects of syntactic–semantic conflict in Broca's area ...
    These results suggest that Broca's area supports sentence processing by mediating between multiple incompatible constraints on sentence interpretation, ...
  57. [57]
    An fMRI study of Broca's area in lexical semantics - ResearchGate
    Aug 6, 2025 · In agreement with previous studies, our findings suggest that Broca's area is involved in semantic processing. Findings of lexico-semantic as ...
  58. [58]
    Semantic and Syntactic Processing During Comprehension
    Dec 24, 2021 · Since event-related potentials (ERPs) contain three particular ERP components, N400 component, left anterior negativity (LAN), and P600 ...
  59. [59]
    A time and place for language comprehension: mapping the N400 ...
    Nov 11, 2013 · We propose a new functional-anatomical mapping of the N400 and the P600 to a minimal cortical network for language comprehension.
  60. [60]
    Functional Role of the N400 and the P600 in Language-related ERP ...
    Jan 11, 2021 · The N400 and P600 components are widely recognized as neural markers of semantic integration and syntactic reanalysis, respectively, with N400 ...<|control11|><|separator|>
  61. [61]
    A hierarchy of linguistic predictions during natural language ... - PNAS
    We analyzed EEG and source localized MEG data ... Fedorenko, fMRI reveals language-specific predictive coding during naturalistic sentence comprehension.
  62. [62]
    MEG Evidence for Incremental Sentence Composition in the Anterior ...
    Nov 4, 2016 · Numerous other studies have provided evidence that human sentence processing is predictive in a way consistent with the left-corner strategy ...
  63. [63]
    Decoding words during sentence production with ECoG reveals ...
    Jun 3, 2025 · Our findings confirm that words share cortical representations across tasks, but reveal a division of labor within the language network.
  64. [64]
    Decoding words during sentence production: Syntactic role ...
    Jan 24, 2025 · Here, we investigate this using high-resolution neurosurgical recordings (ECoG) and an overt production experiment where patients produced ...Missing: comprehension 2020-2025
  65. [65]
    The “Podcast” ECoG dataset for modeling neural activity ... - Nature
    Jul 3, 2025 · We introduce the “Podcast” electrocorticography (ECoG) dataset for modeling neural activity supporting natural spoken narrative comprehension.
  66. [66]
    25-year neuroimaging research on spoken language processing
    Nov 27, 2024 · The aim of this study is to depict the dynamics of publications in the field of neuroimaging research on spoken language processing between 2000 and 2024.
  67. [67]
    Revisiting the role of computational neuroimaging in the era ... - Nature
    Sep 6, 2024 · For example, high equipment costs of neuroimaging devices (e.g., fMRI and MEG) limit access for smaller institutions, potentially hindering ...Missing: sentence | Show results with:sentence
  68. [68]
    [PDF] Sentence Disambiguation by a Shift-Reduce Parsing Technique
    We have developed a parsing algorithm--a variant of the LALR(I} shift.-reduce algorithm--that models the preference behavior of native speakers for a range of ...
  69. [69]
    Recurrent neural networks as neuro-computational models of ...
    While RNNs can simulate several key behavioral observations about human speech and language processing, it is unknown whether RNNs also develop computational ...
  70. [70]
    [PDF] Human Sentence Processing: Recurrence or Attention?
    Jun 10, 2021 · Recurrent Neural Networks (RNNs) are widely used in psycholinguistics and Natural Language. Processing (NLP). Psycholinguists have looked to.
  71. [71]
    Genome-wide analyses of individual differences in quantitatively ...
    Individual differences in reading- and language-related skills are influenced by genetic variation, with twin-based heritability estimates of 30 to 80% ...
  72. [72]
    [PDF] Linguistic complexity: locality of syntactic dependencies - TedLab - MIT
    This paper proposes a new theory of the relationship between the sentence processing mechanism and the available computational resources.
  73. [73]
    Testing the Predictions of Surprisal Theory in 11 Languages
    Dec 14, 2023 · A positive Δ means that the target predictors contribute to psycholinguistic predictive power above the baseline predictor, whereas a Δ of zero ...Missing: readability | Show results with:readability
  74. [74]
    BiLex: A computational approach to the effects of age of acquisition ...
    We developed the BiLex computational model to examine the effects of L2 age of acquisition, language use and exposure on lexical retrieval in bilingual speakers ...
  75. [75]
    Lexical Effects on Second Language Grammar Acquisition: Testing ...
    Sep 13, 2024 · This study finds that lexical variables impact the incidental learning of L2 grammar during L2 sentence comprehension.Missing: computational | Show results with:computational