Fact-checked by Grok 2 weeks ago

Syntax

Syntax is the branch of that studies the rules and principles governing the structure of sentences within a , including how words and morphemes combine to form phrases, clauses, and complete utterances that convey meaning. It focuses on the arrangement of linguistic elements to produce grammatically well-formed constructions, distinguishing between acceptable and unacceptable sequences based on a language's specific conventions. Unlike , which deals with word-internal structure, syntax examines larger units and their , such as subject-verb-object orders or embedding of clauses. The formal study of syntax traces back to ancient grammarians like in around the 4th century BCE, who developed rule-based systems for sentence generation, but it gained prominence in Western traditions through the 19th-century psychologistic approaches that linked syntax to mental processes. The field transformed in the mid-20th century with structuralist methods from linguists like , who emphasized to break down sentences into hierarchical components. Noam Chomsky's (1957) marked a pivotal shift by introducing , which posits that syntax arises from an innate enabling humans to produce and comprehend infinite sentences from finite rules, incorporating transformations to relate underlying "deep structures" to observable "surface structures." Subsequent developments include in the 1970s, which formalized to account for consistent patterns across categories like nouns and verbs, and the framework in the 1980s, suggesting that syntactic variation between languages stems from fixed universal principles combined with language-specific parameters. The , proposed by Chomsky in the 1990s, seeks to simplify these models by deriving from general cognitive constraints, emphasizing and efficiency in computation. Alternative approaches, such as and , prioritize relational hierarchies or holistic patterns over strict rule-based generation. Syntax plays a crucial role in language comprehension and production, enabling the recursive embedding of structures that allows expression of complex ideas with a limited vocabulary, as seen in the principle of compositionality where meaning builds hierarchically from smaller units. It interfaces with semantics to determine how structure influences interpretation, with phonology for prosodic realization, and morphology for inflectional agreement, impacting fields like language acquisition, where children master syntactic rules rapidly, and computational linguistics for natural language processing. Cross-linguistically, syntax reveals both universals, such as headedness in phrases, and diversities, informing theories of language evolution and typology.

Fundamentals

Etymology

The term "syntax" originates from the σύνταξις (syntaxis), denoting "arrangement" or "a putting together," derived from the prefix σύν- (syn-, "together") and τάξις (taxis, "arrangement" or "order"). In classical Greek linguistic and philosophical contexts, it initially encompassed the systematic organization of elements, including rhetorical and logical structures. The term entered Latin as syntaxis through scholarly translations and adaptations of Greek grammatical works, with its first systematic application appearing in Priscian's Institutiones Grammaticae (early ). , a grammarian active in , employed syntaxis in Books 17 and 18 to describe the construction and dependencies of sentences, marking the inaugural comprehensive treatment of and establishing it as a core component of grammatical study. This adoption bridged Greek theoretical foundations with Latin pedagogical needs, influencing medieval grammatical traditions. During the , the meaning of syntaxis underwent a notable , transitioning from a rhetorical emphasis on stylistic arrangement—rooted in classical —to a stricter grammatical focus on the structural rules governing formation across languages. Humanist scholars, drawing on Priscian's while adapting it to emerging national grammars, integrated syntax into broader linguistic analyses, as exemplified in Scaliger's De causis linguae Latinae (), which emphasized logical and morphological interrelations in building. This shift facilitated the development of syntax as an autonomous field, distinct from , in early modern European linguistics.

Definition and Scope

Syntax is the branch of that studies the rules, principles, and processes governing the formation of sentences in a , particularly how words combine to create phrases, clauses, and larger syntactic units. This field examines the structural arrangements that determine whether sequences of words are grammatically well-formed, independent of their sound patterns or meanings. The scope of syntax encompasses key phenomena such as phrase structure, which organizes words into hierarchical units like noun phrases and verb phrases; , where elements like subjects and verbs match in features such as number and person; case marking, which indicates grammatical roles through affixes or ; and , allowing structures to embed within themselves to produce complex sentences. However, syntax explicitly excludes , the study of sound systems and pronunciation, and semantics, the analysis of meaning and . These boundaries ensure that syntactic inquiry focuses on form and arrangement rather than auditory or interpretive aspects of language. Syntax is distinct from , which concerns the internal structure of words and how they are built from smaller units called morphemes. For instance, in English, morphology handles verb conjugation, such as adding the "-s" to form "walks" from "walk" to indicate third-person singular , whereas governs the arrangement of words into sentences, like positioning the subject before the in declarative statements ("The dog walks"). This division highlights morphology's focus on word-level modifications versus syntax's emphasis on inter-word relations and sentence-level organization. Within linguistic theory, syntax plays a central role in distinguishing —innate principles common to all human languages—from language-specific rules that vary across tongues. Noam Chomsky's generative framework posits that syntactic competence is biologically endowed, enabling children to acquire complex structures rapidly despite limited input, as outlined in his seminal works on . This innate perspective underscores syntax's foundational position in the human language faculty, balancing universal constraints with parametric variations in individual languages.

Core Concepts

Word Order

Word order in syntax refers to the linear arrangement of major syntactic elements, such as the (S), (V), and object (O), within a . This sequencing varies systematically across languages and plays a crucial role in conveying grammatical meaning, often interacting with morphological markers like case or to disambiguate roles. Typologically, languages are classified based on the dominant order of these elements in declarative sentences, with six primary patterns possible: SVO, SOV, VSO, VOS, OSV, and OVS, though the last two are rare. The most common word order types are SVO and SOV, which together account for approximately 75% of the world's languages according to the World Atlas of Language Structures (WALS) database. English exemplifies SVO order, as in "The cat (S) chased (V) the mouse (O)," where the precedes the , and the object follows. In contrast, represents SOV order, as seen in "Neko-ga (S) nezu-o (O) oikaketa (V)," with the object appearing before the . VSO order is prevalent in many and Austronesian languages; for instance, uses VSO in sentences like "Chonaic (V) mé (S) an fear (O)" meaning "I saw the man." These basic orders provide a foundation for understanding syntactic variation, though actual usage can be influenced by additional factors. Several factors influence deviations from rigid , including the hierarchy, which prioritizes more animate entities (e.g., humans over inanimates) in prominent positions, and prominence, where elements like topics or foci may front or postpone based on structure. For example, in Turkish (SOV-dominant), animate objects can precede the verb more readily than inanimates to highlight them. Typological tendencies also correlate with other features, such as head-initial (SVO) languages favoring prepositions over postpositions, while head-final (SOV) languages show the reverse pattern. These influences ensure that serves both grammatical and pragmatic functions across language families. Some languages exhibit free or flexible , where the sequence of elements can vary without altering basic meaning, often due to rich case marking that encodes grammatical roles morphologically. Latin is a classic example: the "Puella (S) puerum (O) videt (V)" can be reordered as "Puerum puella videt" or other permutations, with nominative and accusative cases distinguishing subject from object. This flexibility is common in languages with overt case systems, such as or Warlpiri, allowing stylistic or discourse-driven rearrangements while maintaining syntactic coherence through inflection. Historical shifts in illustrate how , simplification, or internal evolution can reshape syntax. , originally SOV in main clauses with subordinate-like embedding, transitioned to SVO around the , influenced by Norman French and the loss of robust case endings, which necessitated fixed positioning for clarity. Similar shifts occur in creoles or scenarios, underscoring word order's adaptability over time.

Grammatical Relations

Grammatical in syntax describe the abstract functional dependencies between constituents in a , primarily involving the and its arguments, such as the , direct object, indirect object, and . The relation typically identifies the primary argument, often encoding the (the initiator of an action) or (the entity undergoing change), as seen in English sentences like "The dog chased the cat," where "the dog" is the subject-agent and "the cat" is the direct object-patient. The direct object relation marks the entity most directly affected by the predicate, while the indirect object specifies a secondary or recipient, as in "She gave him a ." Predicate relations link the to these arguments, and provide optional modifiers like time or without core participation in the event. Identification of these relations relies on multiple criteria, including morphological agreement, government, and behavioral tests. Agreement involves feature matching between the subject and predicate, such as number and person; in Spanish, for instance, a singular subject requires a singular verb form, as in "El perro corre" (the dog runs), where the verb "corre" agrees in third-person singular with "perro," but mismatches like "*El perro corren" are ungrammatical. Government refers to the structural dominance of a head (e.g., a verb) over its dependents, enabling case assignment; verbs govern and assign accusative case to direct objects in languages like German, where "Ich sehe den Hund" (I see the dog) marks "Hund" with accusative "-en" under the verb's government. Behavioral tests further diagnose relations through syntactic operations: in passivization, the direct object of an active clause like "The cat chased the dog" raises to subject position in "The dog was chased by the cat," while the original subject demotes to an oblique; raising constructions similarly promote subjects, as in "The dog seems to chase the cat," where only the subject "the dog" can raise from the embedded clause. Cross-linguistically, exhibit variations in systems, contrasting accusative (where the subject of intransitives aligns with transitive subjects, S=A ≠ O) and ergative (where the subject of intransitives aligns with transitive objects, S=O ≠ A) patterns. In accusative languages like English or , the subject of "The dog runs" patterns with that of "The dog chases the cat" in controlling verb agreement and . Ergative appears in languages like , where the intransitive subject in "Gizonak korrika egiten du" (the man runs) takes absolutive case (unmarked), aligning with the transitive object in "Gizonak mutila ikusi du" (the man saw the boy), while the transitive subject takes ergative "-ak"; this inverts the typical subject-object hierarchy for morphological marking and some syntactic behaviors. These relations play a crucial role in sentence interpretation by projecting semantic content into syntactic structure, particularly through theta roles, which assign thematic interpretations like or to arguments in specific positions. Under the Uniformity of Theta Assignment Hypothesis, theta roles such as (external argument in specifier position) and (internal argument as complement) are systematically mapped to syntactic projections, ensuring that event participants like the in "John broke the window" occupy the subject position to license the thematic structure. This projection facilitates semantic composition while interacting with surface variations like , though relations remain abstract and positional-independent.

Constituency and Phrase Structure

In syntax, constituency refers to the hierarchical grouping of words into larger units known as constituents, such as noun phrases (NPs), verb phrases (VPs), and clauses, which form the building blocks of sentence structure. These groupings are not merely linear sequences but reflect functional and structural relationships that determine how sentences are parsed and interpreted. Linguists identify constituents through specific tests that reveal whether a of words behaves as a cohesive unit. One key method is the substitution test, where a potential constituent can be replaced by a single word or , such as a , without altering the sentence's . For example, in "The big dog barked loudly," the string "the big dog" can be substituted with "it" to yield "It barked loudly," indicating that "the big dog" forms an constituent. Similarly, "barked loudly" can be replaced with "did so" in "The big dog did so," confirming it as a VP. Another test is , which checks if a string can be relocated within the sentence while preserving ; for instance, "The big dog" can be fronted to "The big dog, I saw yesterday," but individual words like "big" cannot move alone in the same way. The coordination test involves joining two identical strings with a like "and"; in "I saw the dog and the cat," both "the dog" and "the cat" can be coordinated, showing they are parallel NP constituents, whereas "dog and the" cannot. These tests collectively demonstrate that constituents exhibit unified behavior in syntactic operations. Phrase structure rules provide a formal way to represent these hierarchical groupings, specifying how categories expand into subconstituents. Introduced in early generative , a basic set of rules for English might include S → NP VP (a consists of a followed by a ), NP → Det N (a consists of a and a ), and VP → V (a consists of a ). These rules generate tree structures that visualize the ; for the "The cat sleeps," the structure is as follows:
      S
     / \
    NP  VP
   /|    |
 Det N    V
  |  |    |
 The cat  sleeps
This tree illustrates how "the cat" branches as an under , distinct from the VP "sleeps." Such rules capture the endocentric nature of phrases, where a head word (e.g., the noun in ) determines the category. A crucial property enabled by is , allowing a category to embed instances of itself indefinitely, which accounts for the creative potential of language. For example, the rule NP → NP PP (a can include another modified by a prepositional phrase) permits nesting, as in "The cat [that chased the mouse [that ate the cheese]] sleeps," where relative clauses embed recursively. This recursive embedding generates sentences of arbitrary complexity from a of rules, a defining of human syntax. While most constituents are continuous spans of words, some languages exhibit non-constituent structures with discontinuous constituents, where elements of a are separated by other material. In , a example is verb-second in main clauses combined with verb-final tendencies in subordinates, as in "Ich habe das Buch gelesen" (I have the book read), where the object "das Buch" and verb "gelesen" form a discontinuous VP split by the auxiliary. Such discontinuities challenge strictly linear models but are handled in phrase structure analyses by allowing gaps or traces in the . These structures relate briefly to grammatical roles, as discontinuous NPs often function as subjects or objects in clause-level relations.

Historical Development

Ancient and Medieval Foundations

The systematic study of syntax in the tradition began with Thrax's Grammatike (Art of Grammar), composed around the 2nd century BCE, which systematically classified words into eight parts of speech—, , , , , , , and —and outlined basic principles of composition as a structured combination of these elements. This work emphasized the role of parts of speech in forming coherent utterances, laying foundational concepts for syntactic analysis by treating as ordered sequences governed by grammatical roles rather than mere word lists. Roman grammarians adapted and expanded these models to in the 4th to 6th centuries , with Aelius Donatus's Ars Grammatica providing an accessible framework through its Ars Minor, which defined the eight parts of speech in a question-and-answer format tailored to Latin's inflectional system, and the more advanced Ars Maior, which addressed construction via morphological and syntactic integration. , in his comprehensive Institutiones Grammaticae (c. 500 ), further developed in Books XVII and XVIII, focusing on constructio ( building) and emphasizing (congruentia) in , number, and case between elements like and verbs, as well as the functional roles of cases in expressing such as nominative for and accusative for objects. Parallel developments in the during the featured Sibawayh's Al-Kitab (The Book), the seminal text of from the Basra school, which analyzed sentence structure through hierarchical relations between words, employing notions akin to where a governing element (e.g., a ) determines the form and function of dependents (e.g., or object via i'rab case endings). This approach treated syntax as a system of operator-operand interactions, distinguishing nominal and verbal sentences and exploring dependencies in complex constructions, influencing subsequent linguistic traditions. In medieval Europe, the 13th-century Modistae, or speculative grammarians, advanced syntactic theory by integrating Aristotelian philosophy with earlier traditions in works like Thomas of Erfurt's Grammatica Speculativa (c. 1300), proposing that syntax arises from universal modi significandi (modes of signifying) shared across languages. These modes divided into essential properties of words (e.g., a noun's signifying essence or quality) and relational aspects (e.g., dependency in concord and government), viewing sentence structure as a reflection of mental modes of understanding (modi intelligendi) that mirror reality (modi essendi), thus providing a metaphysical basis for syntactic universality.

Modern Emergence

The modern emergence of syntax as a distinct field of linguistic inquiry began in the 19th century with the development of the comparative method, pioneered by scholars such as Franz Bopp and Jacob Grimm, who applied systematic comparisons across Indo-European languages to uncover syntactic structures. Bopp's Vergleichende Grammatik des Sanskrit, Zend, Griechischen, Lateinischen, Littlauischen, Gothischen und Deutschen (1833–1852) extended comparative analysis beyond phonology and morphology to syntax, identifying parallels in grammatical forms, including verb placement, where many early Indo-European languages exhibited verb-final tendencies in subordinate clauses. Grimm, in his Deutsche Grammatik (1819–1837), further explored these parallels, noting consistent syntactic patterns in word order and inflectional agreement across Germanic and other Indo-European branches, such as the positioning of finite verbs in main and subordinate clauses. This approach emphasized empirical reconstruction of proto-syntactic features, laying the groundwork for viewing syntax as a historically reconstructible system rather than isolated rules. In the early , American , led by , shifted focus toward descriptive analysis of surface structures, marking a pivotal advancement in syntactic methodology. Bloomfield's (1933) introduced (IC analysis), a technique for sentences into hierarchical binary divisions based on distributional patterns, such as dividing "The man hit the ball" into ("The man") and ("hit the ball"), then further subdividing each. This method prioritized observable form over meaning, treating syntax as the arrangement of forms into meaningful sequences without invoking mentalistic constructs. During the and , Bloomfieldian dominated, influencing fieldwork and corpus-based studies that cataloged syntactic units through segmentation and tests. Post-Bloomfieldian developments refined this distributional paradigm, conceptualizing syntax as patterns of among linguistic elements. Zellig Harris, a key figure in this era, advanced string-based methods in Methods in Structural Linguistics (1951), where sentences were analyzed as linear sequences of classes (e.g., Noun-Verb-Noun patterns), using transformation rules to derive variations from kernel structures without reference to deep meaning. Charles Hockett's item-and-arrangement (IA) model, outlined in "Two Models of Grammatical Description" (1954), formalized syntax as the linear positioning of discrete items (morphemes or words) within slots, contrasting with process-oriented views and emphasizing empirical predictability in arrangements like ordering in (e.g., "cats" as + ). This approach dominated mid-20th-century descriptivism, focusing on verifiable distributions derived from corpora. A transitional figure bridging and later innovations was , whose Language: An Introduction to the Study of Speech (), particularly Chapter VI on "Form in Language," highlighted syntax as patterned formal relations that shape conceptual expression, influencing early ideas on grammatical form in subsequent generations. Sapir's emphasis on drift and formal processes in syntactic evolution provided conceptual depth to distributional methods, paving the way for generative approaches without delving into transformational mechanics.

Major Theoretical Frameworks

Dependency Grammar

Dependency grammar represents syntactic structure as a in which words serve as nodes connected by directed relations, emphasizing head-dependent asymmetries without intermediate constituents. In this framework, every word except the root depends on exactly one head, forming a projective structure that captures the of through direct word-to-word dependencies. For instance, in the "The cat chased the mouse," the "chased" acts as the head, with "cat" as its dependent and "mouse" as its object dependent, illustrating how dependencies encode grammatical roles like and object. The foundational work in dependency grammar was established by Lucien Tesnière in his posthumously published book Éléments de syntaxe structurale (1959), which introduced key concepts such as valency—the number of dependents a head can take—and stemma diagrams to visualize dependency trees. Tesnière argued that syntax is best understood through these asymmetric relations between words, rejecting phrase-based hierarchies in favor of a more linear, relational approach that highlights the verb's central role as the sentence's governor. His theory drew on examples from numerous languages to demonstrate how dependencies account for syntactic connections universally, influencing subsequent developments in . One major advantage of dependency grammar is its suitability for languages with free or flexible word order, such as , where surface linear arrangements vary without altering core relations, as dependencies focus on relational ties rather than fixed positions. This flexibility facilitates cross-linguistic and in diverse languages. Additionally, dependency representations have proven effective in computational applications, particularly , where they enable robust handling of reordering and alignment between source and target languages by preserving syntactic relations independently of word order. Several variants of dependency grammar extend Tesnière's principles while incorporating specific constraints or cognitive emphases. Word Grammar, developed by Richard Hudson in 1984, treats syntax as a network of word-to-word dependencies within a broader cognitive framework, emphasizing and default rules to model linguistic as interconnected lexical items. Another variant, Link Grammar, proposed by Daniel Sleator and Davy Temperley in 1993, formalizes dependencies as typed links between words with a no-crossing constraint to ensure planarity in parses, allowing efficient algorithmic implementation for tasks like English sentences.

Categorial Grammar

Categorial grammar is a formal theory of syntax that models linguistic expressions as categories drawn from a type system, where syntactic combination proceeds through function-argument application akin to logical inference. This approach emphasizes the lexical specification of syntactic behavior, treating words and phrases as functors that saturate arguments to build larger structures. The theory originated in the work of Kazimierz Ajdukiewicz, who in 1935 proposed a system of syntactic categories to ensure the "syntactic connectivity" of expressions, preventing semantically anomalous combinations like treating a noun as a sentence. Ajdukiewicz's framework used basic categories such as N for nouns and S for sentences, along with functor categories like S/N, which denotes an expression that combines with an N on its right to yield an S. Ajdukiewicz's ideas were revived and modernized in the late , particularly through Steedman's development of (CCG), as detailed in his monograph Surface Structure and Interpretation. Steedman extended the theory to directly derive surface structures, incorporating combinatory rules to permit flexible and coordination without relying on abstract underlying representations. In this system, categories are defined directionally: for instance, a like "sees" is assigned the category (S\NP)/NP, indicating it takes a (NP) object on the right to form a verb phrase (S\NP), which then combines with a NP on the left to yield a sentence S. This slash notation—forward slash (/) for rightward argument seeking and backward slash () for leftward—encodes the order-sensitive application of functions, as in the combination: NP + ((S\NP)/NP) + NP → S, exemplified by "Alice sees Bob." A key advancement in came with type-logical variants, notably the Lambek calculus formulated by Joachim Lambek in 1958. Lambek's non-commutative calculus introduces product-free calculi with division operations (/ and ) that respect linear precedence, allowing the theory to generate context-free languages while capturing constraints inherent in natural languages. Unlike commutative logics, the Lambek system enforces directionality, ensuring that arguments combine in the specified sequence, such as deriving S from (S/NP) + NP but not vice versa without additional rules. Categorial grammar's strengths lie in its capacity to model complex phenomena that challenge phrase-structure approaches, including discontinuous constituents and non-constituent coordination. Discontinuous constituents, such as those arising in extraction constructions like "Who does Alice see?", are handled via type-raising, where an raises to a like (S/)\NP to permit crossing dependencies without movement operations. Non-constituent coordination, as in "Alice likes not only syntax but also semantics," is analyzed by assigning coordinators like "" categories that wrap around non-adjacent phrases, such as (X/XP)(YP\YP) for the first conjunct, enabling flat coordination over arbitrary constituents. These mechanisms relate briefly to by specifying argument positions through slash directions, where, for example, the left slash in S\NP identifies the role.

Generative Grammar

Generative grammar, pioneered by Noam Chomsky, posits that human language syntax is governed by an innate universal grammar (UG), a set of formal rules enabling the infinite generation of sentences from finite means. This framework emerged as a response to limitations in earlier structuralist approaches, emphasizing hierarchical structures and recursive rules over mere linear sequences. In his seminal 1957 work Syntactic Structures, Chomsky introduced phrase structure rules to generate underlying syntactic trees and transformations to derive surface forms from them, such as converting active sentences like "John eats the apple" to passive "The apple is eaten by John." These mechanisms captured syntactic regularities while accounting for productivity and creativity in language use. Phrase structure rules, exemplified by productions like S → NP VP and VP → V NP, provided a formal basis for constituency, while transformations, including deletions and movements, explained relations between sentence types. This shift from taxonomic description to explanatory adequacy revolutionized linguistics, positing syntax as a computational system of the mind. Key developments in the 1970s included X-bar theory, which generalized phrase structure into a uniform template applicable across categories. Proposed by Chomsky in 1970, X-bar theory structures phrases hierarchically as XP → Specifier X' and X' → X Complement (or Adjunct), where X represents any head (e.g., N, V), ensuring endocentricity and capturing cross-categorial parallels, such as in noun phrases (NP → Det N') and verb phrases (VP → NP V'). This theory constrained possible grammars by prohibiting flat structures and ad-hoc rules. By 1981, Chomsky's government and binding (GB) theory integrated X-bar with subtheories like binding (governing pronoun-antecedent relations), case (assigning thematic roles), and government (defining head-dependent interactions), unifying diverse phenomena under modular principles within UG. The Minimalist Program, outlined by Chomsky in 1995, streamlined GB by reducing syntax to core operations and economy principles, aiming for maximal simplicity in UG. Central is Merge, a recursive binary operation combining elements to form new sets (e.g., Merge({the, book}) yields {the, book}), building structures bottom-up without extraneous machinery. Transformations are recast as internal Merge (movement), guided by principles like shortest move, which favors minimal distances for efficiency, as in wh-movement where auxiliaries raise only as needed. This approach eliminates language-specific rules, deriving variation from interface conditions with phonology and semantics. Underpinning these models is the principles-and-parameters framework, where UG consists of invariant principles (e.g., structure preservation in movement) and finite parameters set during acquisition. For instance, the head-directionality parameter determines whether heads precede (head-initial, as in English VO order) or follow complements (head-final, as in Japanese OV order), accounting for typological differences with minimal variation. This binary choice exemplifies how children "fix" parameters from sparse input, supporting rapid language acquisition across diverse tongues.

Functional and Usage-Based Grammars

Functional and usage-based grammars emphasize the role of syntax in facilitating communication and emerge from patterns observed in actual language use, rather than abstract formal rules. These approaches view grammatical structures as shaped by their functional contributions to , such as organizing to convey topics and effectively. In Functional Grammar, developed by Simon C. Dik, syntax is analyzed as a resource for performing functions, where elements like topic (the entity about which is provided) and (the new or highlighted ) determine word order and constituent placement to meet communicative needs. Dik's framework, outlined in his book Functional Grammar, posits that sentences are built around predicate frames that incorporate pragmatic functions to structure flow in context, drawing on cross-linguistic data to illustrate how syntax serves interpersonal and textual purposes. Usage-based grammars extend this functional orientation by grounding syntax in empirical patterns derived from language exposure, particularly through as formalized by Adele E. Goldberg. In this model, constructions—conventionalized form-meaning pairings such as the caused-motion construction ("X causes Y to VERB," e.g., "She sneezed the napkin off the table")—are the basic units of grammar, acquired and stored as holistic units based on frequency in input corpora. Goldberg's 1995 work Constructions: A Construction Grammar Approach to Argument Structure argues that these constructions license argument structures beyond what verbs alone predict, enabling speakers to express nuanced meanings through memorized patterns encountered in usage. This approach highlights how syntactic productivity arises from generalizing over specific exemplars, with corpus analysis revealing that high-frequency constructions shape idiomatic and novel expressions alike. Cognitive linguistics integrates these ideas in Ronald W. Langacker's Cognitive Grammar, which treats syntax as an extension of general cognitive processes involving conceptualization and categorization. Langacker's framework features conceptual networks where syntactic structures reflect prototype-based schemas, allowing for graded membership and family resemblances among constructions rather than discrete rules. In Foundations of Cognitive Grammar (1987), Langacker demonstrates how phenomena like nominalization or possession involve symbolic assemblies that profile certain aspects of scenes, with prototype effects explaining variations in syntactic behavior across contexts. This view underscores syntax's embodiment in human cognition, where meaning construction drives form. The empirical foundation of functional and usage-based s lies in the demonstrable impact of and input quality on syntactic acquisition, challenging nativist accounts of innate . shows that ren learn constructions through statistical patterns in ambient , with higher token accelerating the of schemas and type promoting . For instance, studies of child corpora reveal that exposure to varied exemplars of a construction, such as ditransitive patterns, facilitates without positing innate principles. Unlike formal models assuming pre-wired syntactic modules, these grammars attribute to emergent generalizations from usage, supported by longitudinal data on acquisition trajectories.

Probabilistic and Network Models

Probabilistic models in syntax introduce elements to grammatical formalisms, treating as distributions over possible rules or relations rather than deterministic ones. This approach addresses the inherent in by assigning probabilities to parses, enabling systems to select the most likely based on training data. Such models facilitate learning from corpora and improve robustness in handling varied linguistic inputs, contrasting with rule-based systems by emphasizing empirical patterns over innate universals. Probabilistic context-free s (PCFGs) extend standard context-free grammars by associating probabilities with production rules, ensuring that the sum of probabilities for rules expanding a given nonterminal equals 1. The probability of a is then the product of the probabilities of its constituent rules; for instance, a grammar might assign P(S \to [NP](/page/NP) \, VP) = 0.9 to reflect the high likelihood of this sentential expansion in English. PCFGs are foundational for statistical , where algorithms like the Cocke-Kasami-Younger (CKY) parser, augmented with probability computations, efficiently find the maximum-likelihood derivation for a . Collins (1999) advanced this framework through head-driven statistical models that lexicalize rules by incorporating head information, achieving parse accuracies around 88% on Wall Street Journal sections while reducing error rates compared to unlexicalized PCFGs. Network-based models draw from to represent syntax through distributed activations in neural architectures, bypassing explicit symbolic rules in favor of learned patterns from sequential data. Elman (1990) demonstrated this with simple recurrent networks (SRNs), where a hidden layer recurrently processes word sequences, gradually discovering grammatical categories like noun-verb distinctions through prediction tasks on synthetic languages. In one experiment, SRNs trained on sentences from a toy formed internal representations that clustered words by syntactic , with rates dropping below 10% after sufficient exposure, illustrating emergent sensitivity to long-range dependencies. These models highlight how probabilistic weight updates during encode syntactic regularities implicitly. For dependency structures, probabilistic models focus on scoring directed relations between words, often assuming projectivity to enable efficient . Eisner (1996) proposed three such models, including a generative approach that factors probabilities over head-modifier attachments and uses a variant of the inside-outside for parameter estimation and . This yields an O(n^3) algorithm for projective dependency trees, with empirical evaluations on parsed corpora showing attachment accuracies up to 85% for certain languages, outperforming earlier non-probabilistic methods by incorporating lexical and directional biases. Modern extensions integrate these ideas into architectures, particularly models that leverage mechanisms to model syntactic dependencies without recurrence. Vaswani et al. (2017) introduced the , which uses multi-head self- to weigh word interactions globally, implicitly capturing hierarchical syntax in tasks like when fine-tuned on treebanks; for example, variants achieve F1 scores exceeding 95% on universal dependency benchmarks. These networks extend probabilistic and connectionist traditions by scaling to large datasets, where distributions approximate parse probabilities. These approaches underpin contemporary computational syntax applications, such as efficient in systems.

Contemporary Applications and Debates

Computational Syntax

Computational syntax encompasses the algorithmic and data-driven approaches to analyzing and generating in () and . It bridges theoretical syntax with practical computation, enabling machines to sentences, resolve structural ambiguities, and produce syntactically well-formed text. Central to this field are methods for efficiently processing context-free grammars (CFGs) and dependency grammars, supported by annotated corpora that serve as training data for statistical models. These techniques have powered key NLP applications, though they face challenges in handling linguistic variability and computational demands. Since the mid-2010s, transformer-based models like (introduced in 2018) and large language models (LLMs) such as series have revolutionized computational syntax by incorporating contextual embeddings for parsing, achieving near-human accuracy on informal and ambiguous inputs without explicit treebanks. Parsing algorithms form the core of computational syntax, transforming input sentences into syntactic representations such as parse trees or graphs. For CFGs, parsing algorithms like the dynamically build a of partial parses to recognize valid structures, achieving worst-case of O(n^3) for a of length n, where the cubic term arises from spanning subsequences and rules. This efficiency issue limits scalability for long sentences, prompting optimizations like left-corner parsing or approximations. In contrast, for grammars, shift-reduce employs a to incrementally build arcs through actions like shifting words onto the or reducing by attaching dependents to heads, often in linear time O(n) for variants, as formalized in arc-standard systems. These algorithms, frequently enhanced with probabilistic models for disambiguation, enable robust syntactic analysis in diverse pipelines. Neural parsers, trained end-to-end on treebanks, now outperform classical methods, with models like using transformers for real-time as of 2025. Annotated treebanks provide essential empirical foundations for training and evaluating parsers. The Penn Treebank, released in 1993, offers over 4.5 million words of bracketed English sentences from sources like , including part-of-speech tags and phrase structure annotations that capture CFG-based hierarchies. This resource has become a for supervised , enabling data-driven induction of grammars and error analysis, with its influence extending to multilingual extensions like the Universal Dependencies project. Applications of computational syntax include syntax-based (MT) systems prevalent in the pre-neural era, where hierarchical models aligned parse trees across languages to reorder and translate constituents more accurately than word-based methods. For instance, early iterations (2006-2016) incorporated phrase-based MT augmented with reordering models, some inspired by syntactic research, to approximate handling of long-distance dependencies. Grammar checkers also rely on syntactic to detect and suggest corrections for structural errors, such as subject-verb agreement violations, by comparing input against CFG rules or dependency patterns in tools like those using deep parsing for minimal edits. In contemporary , LLMs integrate syntactic understanding implicitly for tasks like and multilingual translation, powering systems like 's neural MT since 2016. Key challenges persist in ambiguity resolution and adapting to informal inputs. Prepositional phrase (PP) attachment ambiguity, exemplified by sentences like "I saw the man with a ," requires deciding whether the PP modifies the verb or , a task addressed via maximum models trained on treebank achieving around 80% accuracy on sets. Handling non-standard syntax in —featuring abbreviations, , and dialectal variations—demands normalization techniques or robust parsers tolerant to noise; for early statistical models (as of 2010s), unnormalized text degraded parsing accuracy by up to 20-30% on platforms like , though neural models as of 2025 reduce this to under 10%. These issues underscore the need for hybrid approaches integrating syntax with broader contextual cues.

Cross-Linguistic Variations

Cross-linguistic variations in syntax reveal profound diversity in how languages structure sentences, reflecting typological patterns documented in large-scale databases. The World Atlas of Language Structures (WALS), edited by Dryer and Haspelmath, compiles data on over 2,600 languages to map structural features, including the distinction between head-marking and dependent-marking systems. In head-marking languages, such as many Native American tongues like , grammatical relations are primarily indicated on the head (e.g., ) through affixes marking subjects and objects, whereas dependent-marking languages like English rely on markers on dependents (e.g., case affixes on nouns). This variation influences how syntactic dependencies are expressed, with head-marking often allowing more flexible due to explicit verbal agreement. Syntactic structures also vary in morphological complexity, as seen in agglutinative and polysynthetic languages. Agglutinative syntax, characteristic of like Turkish, involves chaining multiple suffixes to a root to convey grammatical information without , enabling highly productive . For instance, the Turkish word ev-ler-im-den ('from my houses') derives from the root ev ('') by sequentially adding the -ler, -im ('my'), and ablative -den ('from'), illustrating suffix chaining that builds complex nominals efficiently. In contrast, polysynthetic languages like (an language) incorporate nouns and other elements into expansive complexes, often expressing an entire in a single word. An example is tusaq-tunna-i-sit ('he hears that I am hearing him'), where the root tusaq ('hear') incorporates , object, and aspectual markers, reducing the need for separate words and highlighting the language's high rate. Areal effects further shape syntactic similarities across unrelated languages, as exemplified by the , a convergence zone involving , , , and like Bulgarian. These languages, despite distinct genetic origins, share morphosyntactic traits due to prolonged contact, including a construction using an auxiliary 'have' (or 'be') plus a past participle, as in Bulgarian imam pročeteno ('I have read'). This innovation, absent in their proto-forms, spread through the region, demonstrating how geographic proximity fosters syntactic borrowing beyond genetic inheritance. Amid this diversity, certain syntactic universals impose constraints, as outlined in Greenberg's seminal work on implicational hierarchies. Greenberg identified 45 universals, many implicational, such as Universal 3: if a exhibits dominant verb-subject-object (VSO) order, it invariably uses prepositions rather than postpositions. This pattern holds across sampled s, like Welsh (VSO with prepositions), suggesting underlying tendencies in syntactic organization despite surface variations. Such universals, derived from comparative analysis of 30 s, underscore limits on possible globally. Recent computational analyses of large corpora and databases like have refined these universals and identified new implicational patterns using on over 7,000 s as of 2025.

Interfaces with Semantics and Pragmatics

The syntax-semantics interface concerns the systematic mapping between syntactic structures and their corresponding meanings, ensuring that compositional semantic interpretation aligns with syntactic composition. A foundational approach to this interface is provided by Montague grammar, developed in the 1970s, which treats syntax and semantics as formal algebras interconnected via translation rules that preserve meaning compositionality. In this framework, syntactic categories are mapped to semantic types using intensional logic, often employing lambda calculus to handle phenomena such as quantifier scope ambiguities; for instance, sentences like "Every farmer who owns a donkey beats it" receive interpretations where the quantifier "every" can take wide or narrow scope relative to other elements, determined by syntactic positioning. This direct translation from syntactic trees to logical forms, as in Montague's fragment of English, underscores the principle that semantic values are computed incrementally from syntactic constituents, influencing subsequent formal semantic theories. Contemporary debates explore how LLMs approximate or challenge this compositionality, with studies showing emergent syntactic-semantic alignments in transformer architectures as of 2025. A key challenge at this interface is the projection problem, which addresses how systematically encode and project or theta-roles assigned by predicates to their , maintaining consistency across derivations. In cases like the dative alternation, verbs such as "give" can project either a prepositional dative ("give the to ") or a double-object ("give the "), where the recipient shifts from to object without altering core event semantics, but influencing accessibility to syntactic operations like passivization. This alternation highlights verb sensitivity in role projection, as classified by Levin, where semantic properties of the verb—such as caused —constrain available syntactic realizations, ensuring that theta-criterion satisfaction (one per , all roles assigned) holds across variants. Such projections link to phrasal syntax, resolving how structure is realized without overgeneration or undergeneration of meanings. The syntax-pragmatics interface examines how syntactic operations encode contextual information structure, often leading to mismatches between canonical word order and discourse needs. Topicalization constructions, for example, permit fronting of constituents like "This book, I read" to mark the fronted element as the topic (given information), facilitating discourse continuity while altering prosodic and intonational patterns to signal focus on the remainder. According to Lambrecht's analysis, such structures activate pragmatic presuppositions, where the topic evokes an aboutness relation to prior context, and the comment provides new assertions, thus bridging syntactic displacement with interpretive inferences beyond truth-conditional semantics. This interface reveals how syntax accommodates pragmatic principles like the given-new contract, without altering underlying semantic roles. Debates in the syntax-semantics and syntax- interfaces have evolved within the , emphasizing economy-driven interactions at the computational system's edges with interpretive components. Uriagereka's 1998 exploration posits that functional projections form a cartography-like hierarchy, where syntactic features interface minimally with semantics (via ) and (via licensing), reducing derivational complexity while accounting for and information packaging. This view challenges earlier modular separations by proposing multiple spell-outs that align syntactic derivations directly with interface demands, such as quantifier raising for semantic or for pragmatic prominence, thereby optimizing the language faculty's design. Recent neuro-linguistic studies (2020s) test these interfaces using fMRI, revealing brain areas where syntax, semantics, and converge, informing AI models of human-like interpretation.

References

  1. [1]
    [PDF] An Introduction to Syntactic Analysis and Theory - Linguistics - UCLA
    Our informal characterization defined syntax as the set of rules or princi- ples that govern how words are put together to form phrases, well formed.
  2. [2]
    [PDF] Syntax - Rutgers Center for Cognitive Science
    The Domain of Syntax. At its most basic level, syntax can be defined as the branch of linguistics that studies how the words of a language can be combined ...
  3. [3]
    [PDF] 3. Syntax
    Syntax is the study of these kinds of structures. It is concerned with the ways that words are assembled into phrases (the little trees marked with dots), and ...
  4. [4]
    [PDF] ASPECTS OF THE THEORY OF SYNTAX - Colin Phillips |
    Linguistic theory is concerned primarily with an ideal speaker- listener, in a completely homogeneous speech-community, who knows its language perfectly and is ...Missing: key | Show results with:key
  5. [5]
    [PDF] Syntactic Theory A Formal Introduction - Stanford University
    A third fundamental claim of Chomsky's concerns the bi- ological basis of human linguistic abilities. We will return to this claim in the next section.
  6. [6]
    Linguistics 001 -- Syntax
    This is the principle of compositionality: language is intricately structured, and linguistic messages are interpreted, layer by layer, in a way that depends on ...<|control11|><|separator|>
  7. [7]
    Syntax - Department of Linguistics - University at Buffalo
    Syntax is the study of sentence structure, and how it interacts with other linguistic information, such as phonology, morphology, semantics, and pragmatics.
  8. [8]
    Syntax | Department of Linguistics - University of Maryland
    Syntax seeks to characterize grammars of particular languages and how they differ, to describe the universal properties that human grammars have.
  9. [9]
    Syntax - Etymology, Origin & Meaning
    Originating c.1600 from Greek syntaxis via Late Latin and French, syntax means the arrangement of parts, especially sentence structure and word order in ...
  10. [10]
    None
    ### Summary of Etymology and Historical Development of "Syntax" in Linguistics
  11. [11]
    [PDF] Priscian: A Syntactic Interpretation of a Graeco-Roman World
    My study of Priscian and his work on Latin syntax combines a philological approach with an analysis of the cultural history of Priscian's time. The benefits of ...
  12. [12]
    [PDF] SYNTAX IN BRAZIL: HISTORIOGRAPHICAL NOTES AND ... - SciELO
    ... Institutiones. Grammaticae (Institutions of Grammar) by the grammarian Priscian of Caesarea, written in the 6th Century CE, the most comprehensive source and ...<|control11|><|separator|>
  13. [13]
    [PDF] Grammars and Rhetorics - Loyola eCommons
    Aspects of Latin syntax receive passing notice in other grammars, including those of. Donatus, but the final two books of the Institutiones offered the first ...
  14. [14]
    Part II - Renaissance to Late Nineteenth Century
    Jul 20, 2023 · The art of rhetoric turns into a general theory of language, incorporating other linguistic disciplines (phonetics, grammar, theory of ...
  15. [15]
    8. Analyzing Sentence Structure - NLTK
    3.3 Recursion in Syntactic Structure​​ (10a) involves nested nominal phrases, while (10b) contains nested sentences. a.
  16. [16]
    Analyzing English Grammar (pt.I) - CSUN
    Grammar is traditionally subdivided into two inter-related studies: Morphology and Syntax. Morphology is the study of how words are formed out of smaller units ...
  17. [17]
    Syntax and Morphology | Linguistics - UGA
    Oct 11, 2023 · Morphology is the study of word structure and its relationship both to sentence structure and to meaning.
  18. [18]
    Chapter 2. A Crash Course in Linguistics - CUNY Pressbooks Network
    Morphology and Syntax are closely related fields in English grammar. Syntax studies the structure of sentences, while morphology studies the formation of words.
  19. [19]
    Innateness and Language - Stanford Encyclopedia of Philosophy
    Jan 16, 2008 · Terminological Note: As Chomsky acknowledges (e.g., 1986:28-29), 'Universal Grammar' is used with a systematic ambiguity in his writings.Chomsky's 'Poverty of the... · Other Research Bearing on... · Language Evolution
  20. [20]
    [PDF] ASPECTS OF THE THEORY OF SYNTAX - Colin Phillips |
    Chomsky, 1964, forth- coming). The grammar of a particular language, then, is to be supplemented by a universal grammar that accommodates the creative aspect ...
  21. [21]
    [PDF] Chomsky-1957.pdf - Stanford University
    Syntax is the study of the principles and processes by which senten- ces are constructed in particular languages. Syntactic investigation of a given language ...
  22. [22]
    [PDF] Universal definition of subject - Jean Mark Gawron
    Keenan, Edward L. 1975. Towards a universal definition of “subject”. In Li (Li 1975). Li, Charles, N. (Ed.). 1975. Subject and topic. New York: Academic ...
  23. [23]
    Subject-verb agreement in Spanish and English - ScienceDirect.com
    This paper reports studies of subject-verb agreement errors with speakers of Spanish and English; we used a sentence completion task, first introduced by ...
  24. [24]
    [PDF] Case chapter - MIT
    Case Theory in Government-Binding Syntax​​ By the late 1970s, it had become clear that the distribution of nominals is governed by special laws cross- ...<|separator|>
  25. [25]
    [PDF] Grammatical relations: A functionalist perspective
    For these he applies several behavioral and control tests in order to ascertain which one is the true object. He concludes that in such cases the dative ...
  26. [26]
    [PDF] Ergativity - RMW Dixon
    Oct 7, 2004 · 'Ergativity' is currently an 'in' term in linguistics. It is ... LANGUAGE, VOLUME 55, NUMBER 1 (1979). This paper surveys the various ...
  27. [27]
    [PDF] Thematic Roles and Syntactic Structure* - Sites@Rutgers
    On the nature of grammatical relations . Cambridge, Mass.: MIT Press. McClure, William. 1994. Syntactic projections of the semantics of aspect. Ph.D ...
  28. [28]
    Constituency – The Science of Syntax - KU Libraries Open Textbooks
    By the end of this chapter, you should be able to,. identify and use seven constituency tests for determining structure; understand how to interpret ...
  29. [29]
    2 Syntactic constituenthood - Penn Linguistics
    The most basic test for syntactic constituenthood is the substitution test. The reasoning behind the test is simple. A constituent is any syntactic unit.
  30. [30]
    6.4 Identifying phrases: Constituency tests – ENG 200
    The tests that identify constituents (often called constituency tests) that we'll review in this chapter come in four basic types: Replacement tests; Movement ...
  31. [31]
    6.13 From constituency to tree diagrams – ENG 200
    Phrase structure rules were central to the theory of syntax developed in Chomsky 1957, which kickstarted the modern field of generative syntax. This rule says ...
  32. [32]
    [PDF] Discontinuous NPs in German - Stanford University
    4. Syntactic Aspects. 41. 4.1. Introduction. 41. 4.2. Reanalysis contra Extraction. 41. 4.2.1 The Noun Phrase Constraint (Horn, 1975) 42.
  33. [33]
    [PDF] Is it Really that Difficult to Parse German?
    A third characteristic feature of German syntax that is a challenge for syntactic annotation and for parsing is the treatment of discontinuous con- stituents. ( ...Missing: linguistics | Show results with:linguistics
  34. [34]
    [PDF] Grammar: A Historical Survey - IOSR Journal
    The system of eight parts of speech first appeared in the work of the great Alexandrian grammarian,. Dionysius Thrax( Late 2nd century B.C. ), who is reputed ...
  35. [35]
    The eight parts | What Graeco-Roman Grammar Was About
    The parts as defined by Donatus and by 'Dionysius Thrax'. A noun is defined by 'Dionysius' as a part of an utterance 'subject to case which signifies a ...
  36. [36]
    Donatus (1), Aelius, grammarian | Oxford Classical Dictionary
    The Ars minor, intended for beginners, deals with the eight parts of speech in question-and-answer format; the Ars maior is more comprehensive and includes ...Missing: Grammatica | Show results with:Grammatica
  37. [37]
  38. [38]
  39. [39]
    (PDF) ramzi baalbaki_the legacy of the kitab sibawayhi's analytical ...
    The Kitāb of Sībawayhi is recognized as the most authoritative work in Arabic grammar, establishing foundational analytical methods for future grammarians.
  40. [40]
    (PDF) Syntactic Reconstruction in Indo-European: State of the Art
    Interest in syntactic reconstruction was implicit in the work of the founding fathers of the Comparative Method, including Franz Bopp and his contemporaries.
  41. [41]
    A Reader in Nineteenth Century Historical Indo-European Linguistics
    Grimm has given nine rules, relating the consonants of Germanic with those of Greek and Latin, less commonly with Sanskrit and other Indo-European languages.Missing: syntactic parallels
  42. [42]
    [PDF] Shaping Comparative Linguistics: The Achievement of Franz Bopp
    Abstract: Franz Bopp (1791-1867) is commonly regarded as one of the founding fathers of Indo-European comparative grammar. Bopp's primary interest was.Missing: placement | Show results with:placement
  43. [43]
    [PDF] Leonard Bloomfield - Language And Linguistics.djvu - PhilPapers
    PREFACE TO THE BRITISH EDITION. L. B.. This edition differs from the American form of this book (New York,. 1933) in two respects; the phonetic symbols conform ...
  44. [44]
    Bloomfield, L. (1933). Language. New York Holt, Rinehart & Winston.
    This paper provides a comparative analysis of nominal loanword integration in two different contact in- duced systems of Greek (ie Grico and Capapadocian)Missing: immediate constituent
  45. [45]
    [PDF] American Structuralism Part One: Leonard Bloomfield
    One of the most widely used techniques for displaying sentence structure is the use of its immediate constituents.ICA is a term used in grammatical analysis to ...
  46. [46]
    [PDF] Structural linguistics
    Page 1. 25 (16s net). STRUCTURAL. LINGUISTICS. ZELLIG S. HARRIS. "This book is the most important contribution to descriptive linguistics since . . .
  47. [47]
    [PDF] two models of grammatical description
    The IA picture is therefore potentially somewhat more complex than our original emphasis on the two notions ('item' and 'arrangement') would imply. There ...
  48. [48]
    Two Models of Grammatical Description - Semantic Scholar
    This study compute numeric vectors representing the semantics of pseudowords on the basis of a computational model of lexical processing, ...
  49. [49]
    The Project Gutenberg eBook of Language by Edward Sapir
    Preface. This little book aims to give a certain perspective on the subject of language rather than to assemble facts about it. It has little to say of the ...Missing: Chomsky | Show results with:Chomsky<|control11|><|separator|>
  50. [50]
    Sapir 1921 - Foundations of Linguistics - Rice University
    Aug 25, 2009 · In it, Sapir tells us how languages are all similar in expressive capacity; tells us that they all have phonemes and morphemes, words, and ...
  51. [51]
    Dependency Grammar - Annual Reviews
    Jan 14, 2019 · In this review, we introduce the basic theoretical assumptions of dependency grammar and review some key aspects in which different dependency ...Missing: core | Show results with:core
  52. [52]
    Elements of Structural Syntax - OAPEN Home
    This stream is known today as dependency grammar (DG). Drawing examples from dozens of languages, many of which he was proficient in, Tesnière presents ...
  53. [53]
    [PDF] Dependency Parsing - Stanford University
    Automatic parsing using dependency grammars was first introduced into compu- tational linguistics by early work on machine translation at the RAND Corporation.
  54. [54]
    [PDF] Word Grammar - Richard ('Dick') Hudson
    This theory of language structure is called 'Word Grammar' (WG) because of the central role of the word as the only unit of syntax.
  55. [55]
    [PDF] Parsing English with a Link Grammar - ACL Anthology
    DANIEL SLEATOR AND DAVY TEMPERLEY. {©A-} &: Ds-. &: {©M+ or (C+ &: Bs+)}. &: (J ... Below we show how to express any cat ego rial grammar concisely as a link ...
  56. [56]
    [PDF] Combinatory Categorial Grammar - Stanford University
    The slash notation shown here is used to define the functions in the grammar. ... These examples illustrate the lexical nature of the categorial grammar approach.
  57. [57]
    [PDF] The Mathematics of Sentence Structure Joachim Lambek
    Mar 11, 2008 · The present paper begins with a new exposition of the theory of syntactic types. It is addressed to mathematicians with at most an amateur ...Missing: original | Show results with:original
  58. [58]
    [PDF] chomsky-remarks-1970.pdf - GLOW Linguistics
    REMARKS ON NOMINALIZATION. 187 of the base to an elaboration of the transformational component in such a case as this. Of course this empirical hypothesis is ...
  59. [59]
    Lectures on Government and Binding - De Gruyter Brill
    Chomsky, Noam. Lectures on Government and Binding, Berlin, New York: De Gruyter Mouton, 2010. https://doi.org/10.1515/9783110884166
  60. [60]
    The Minimalist Program - MIT Press
    The Minimalist Program consists of four recent essays that attempt to situate linguistic theory in the broader cognitive sciences.
  61. [61]
    Functional Grammar - Simon Cornelis Dik - Google Books
    Bibliographic information ; Author, Simon Cornelis Dik ; Publisher, North-Holland Publishing Company, 1978 ; ISBN, 0720461804, 9780720461800 ; Length, 230 pages.
  62. [62]
    Simon C. Dik, Functional grammar. North-Holland Linguistic Series ...
    Simon C. Dik, Functional grammar. North-Holland Linguistic Series, 37. Amsterdam: North-Holland Publishing Company, 1978. Pp. xii+230.
  63. [63]
    A Construction Grammar Approach to Argument Structure, Goldberg
    Adele E. Goldberg proposes that grammatical constructions play a central role in the relation between the form and meaning of simple sentences.
  64. [64]
    (PDF) Goldberg's construction grammar - ResearchGate
    Nov 5, 2017 · Constructions: A construction grammar approach to argument structure . University of Chicago Press (1995). Pp. xi+265. The cornerstone of ...
  65. [65]
    Foundations of cognitive grammar volume I. Theoretical ...
    Foundations of cognitive grammar volume I. Theoretical prerequisites: By Ronald W. Langacker. Stanford: Stanford University Press, 1987, xii + 516pp.<|control11|><|separator|>
  66. [66]
    The ubiquity of frequency effects in first language acquisition - PMC
    This review article presents evidence for the claim that frequency effects are pervasive in children's first language acquisition.
  67. [67]
    The role of input revisited: Nativist versus usage-based models
    This article examines the role of input in two contrasting theories of language acquisition: nativist (UG) theory and the usage-based (emergentist) approach.<|control11|><|separator|>
  68. [68]
    [PDF] Probabilistic Context-Free Grammars (PCFGs) - Columbia CS
    A context-free grammar (CFG) is a 4-tuple G = (N,Σ, R, S) where: • N is a finite set of non-terminal symbols. • Σ is a finite set of terminal symbols. • R is a ...Missing: 1999 paper
  69. [69]
    [PDF] Head-Driven Statistical Models for Natural Language Parsing
    The appendices of Collins. (1999) give a precise description of the parsing algorithms, an analysis of their compu- tational complexity, and also a description ...
  70. [70]
    Finding structure in time - ScienceDirect.com
    One approach is to represent time implicitly by its effects on processing rather than explicitly (as in a spatial representation).
  71. [71]
    [PDF] 1990-elman.pdf - Gwern.net
    In this simulation, the goal is to find the temporal structure of the XOR sequence. Simple auto-association would not work, since the task of simply reproducing ...
  72. [72]
    [PDF] Three New Probabilistic Models for Dependency Parsing
    After presenting a novel O(n3) parsing al- gorithm for dependency grammar, we de- velop three contrasting ways to stochasticize.
  73. [73]
    [PDF] arXiv:1706.03762v7 [cs.CL] 2 Aug 2023
    Aug 2, 2023 · We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions.
  74. [74]
    [PDF] Constituency Parsing - Stanford University
    We begin our investigation of the CKY algorithm by examining the requirement that grammars used with it must be in Chomsky Normal Form (CNF). Recall from.
  75. [75]
    Building a Large Annotated Corpus of English: The Penn Treebank
    Marcus, Beatrice Santorini, and Mary Ann Marcinkiewicz. 1993. Building a Large Annotated Corpus of English: The Penn Treebank. Computational Linguistics, 19(2): ...
  76. [76]
    WALS Online - Home
    It is a separate publication, edited by Dryer, Matthew S. & Haspelmath, Martin (Leipzig: Max Planck Institute for Evolutionary Anthropology, 2013). The main ...Languages · Features · Chapters · Download
  77. [77]
    Formulaicity in an agglutinating language: The case of Turkish
    Aug 6, 2025 · In agglutinative languages morphemes are 'glued' together to make up complex words. Durrant (2013) gave the Turkish example olabileceğini ...
  78. [78]
    Polysynthesis: A review - Zúñiga - 2019 - Compass Hub - Wiley
    May 6, 2019 · Complex verb formation revisited: Restructuring in Inuktitut and Nuu-chah-nulth. In M.-A. Mahieu, & N. Tersis (Eds.), Variations on ...
  79. [79]
    [PDF] Lexical polysynthesis - HAL-SHS
    Aug 16, 2012 · Polysynthesis in Inuit appears at both lexical and sentence levels. ... Further problems arise in connection with the extent of polysynthesis in ...
  80. [80]
    [PDF] Friedman VA (2006), Balkans as a Linguistic Area. - Knowledge Base
    In modern terms, a sprachbund is understood as two or more geographically contiguous and genealogical- ly different languages sharing grammatical and lexi- cal ...
  81. [81]
    The Balkan Languages and Balkan Linguistics - Annual Reviews
    Jun 29, 2011 · Abstract. The Balkans were the first linguistic area (sprachbund) to be identi- fied as such. The concept was originally proposed to explain ...
  82. [82]
  83. [83]
    [PDF] Universals of language - Internet Archive
    ... UNIVERSALS of LANGUAGE. EDITED BY JOSEPH H. GREENBERG. PROFESSOR OF ANTHROPOLOGY. STANFORD UNIVERSITY. REPORT OF A CONFERENCE HELD AT. DOBBS FERRY, NEW YORK.
  84. [84]
    [PDF] Greenberg Universals
    The Basic Order Typology *. Joseph H. Greenberg. Linguists are, in general, familiar with the notion that certain languages tend consistently to put modifying ...
  85. [85]
    [PDF] Grammar. - Semantics Archive
    Montague's Universal Grammar: the 'Theory of Reference' (Montague 1970b: 378ff.). Although presented as a development of Frege's (1892) logical semantics ...