Proseminar in Syntax: Architecture, Formalism, and Empirical Challenge
Title: The Architecture of Human Language: Current Contours in Syntactic Theory
Prerequisites: Understanding of GB/Minimalist Program, formal semantics, computational methods, advanced morphology.
Philosophy
This proseminar moves beyond surveying syntax. Its core purpose is to:
Explain the architecture of the human language faculty, not just catalog phenomena.
Identify where current models fail and propose formal solutions.
Bridge theory with mathematical formalization: students formalize syntactic mechanisms using set theory, graph theory, and Minimalist Grammars (MGs).
Produce original, journal-quality contributions: each student’s work is peer-reviewed and grounded in empirical counter-examples.
Overview
1: The Calculus of Merge & Symmetry Breaking
Focus: Internal vs. External Merge; Labeling Algorithm (LA); symmetry crises when two phrases merge.
Exercise: Derive why Subject moves to Spec,TP as a labeling necessity, not because of a feature.
Actionable Task: Students construct formal derivations showing symmetry-breaking with minimal search.
Readings: Epstein, Kitahara, & Seely (2024), Chomsky (2021), Hornstein (2017)
2: Internalization of Locality & Activity Conditions
Focus: Strict Cyclicity, Activity Condition, Minimality, and long-distance dependencies.
Exercise: Analyze a counterexample from a lesser-studied language (e.g., ergative or clitic doubling) that violates standard Minimalist assumptions.
Task: Formal proof in set-theoretic or graph-theoretic terms of how locality constraints derive movement patterns.
3: Linearization & Sensory-Motor Interface
Focus: LCA vs. PF-driven linearization; word order as emergent rather than syntactic.
Actionable Task: Model free word-order languages using Minimalist Grammars (MGs) to predict linearization patterns.
Readings: Collins & Stabler (2024), Kayne (1994), Chesi (2021)
4: Cartography vs. Minimalist Roots
Focus: Fine-structured functional hierarchies (Cinque/Rizzi) vs. economy-driven Minimalism.
Debate: Are rich hierarchies necessary, or can derivational economy explain them?
Exercise: Derive a microvariation dataset using both approaches and formalize predictions.
Readings: Rizzi (2013), Cinque (2005), Boeckx (2017)
5: Phases, Features, and the Death of Agreement
Focus: Dynamic phases, feature checking, and non-mandatory agreement.
Debate: Is agreement a derived phenomenon of strict cyclicity?
Actionable Task: Students construct derivations for D-pronoun incorporation or pro-drop using dynamic phases.
Readings: Gallego (2020), Preminger (2014/2025), Legate (2014)
6: Mathematical & Computational Syntax
Focus: Minimalist Grammars, complexity classes, and the computational limits of syntax.
Exercise: Formalize Swiss-German cross-serial dependencies as an MG; assess if human syntax is Mildly Context-Sensitive.
Readings: Stabler (1997, 2024), Chesi (2021), Collins & Stabler (2024)
7: Evolutionary & Cognitive Frontiers
Focus: Symmetry-breaking in acquisition, interface pressures in processing, recursion, and cognitive constraints on syntactic architecture.
Activity: Whiteboard collaboration deriving formal grammar fragments that accommodate typological or processing anomalies.
Radical Inquiry- The Autopsy of a Miracle
We are not here to memorize rules. We are not here to catalog phenomena. We are here to perform a forensic autopsy on a miracle: the human language faculty.
Ask yourself:
What makes language possible?
Not the words. Not the morphemes. But the computational spine that allows thought to manifest in sound and symbol.
Why does it take the form it does?
Not by accident. Not by cultural drift. But as the necessary consequence of cognitive, evolutionary, and interface constraints.
How is it realized in the mind?
Not as a metaphorical tree. Not as a descriptive map. But as a formal system, a derivational algorithm, a set of precise, lawful operations.
This is not a survey. This is discovery in real time.
Each derivation is a proof, each anomaly a test of necessity.
Every exception is a signal of architecture, not an inconvenience.
Every model we examine is a hypothesis to be dissected, formalized, and stress-tested.
This is a laboratory of thought. Here, theory meets formal rigor, anomalies are opportunities, and the invisible forces of syntax are revealed in every derivation.
You are no longer students of language. You are architects of the computational mind, explorers of possibility itself.
Today, we do not learn.
We interrogate.
We formalize.
We see the architecture of language, not as it appears, but as it must be.
Language Is Not a Set of Rules- Diagnosing the Shadows of Computation
Language is not a checklist of rules, nor a hierarchy of labels, nor a string of morphemes to memorize. Current models, GB, early Minimalism, and even contemporary cartography, treat the surface as causal, and in doing so, they mistake shadows for substance.
Consider the failures:
Overpopulated Features: Syntax is crowded with features, projections, and mechanisms that explain nothing and obscure the underlying logic. These are the patches of descriptive comfort, not the engines of computation.
Hard-Coded Hierarchies: The proliferation of rigid functional layers hides the principle of economy. Derivations are over-determined; the system’s natural efficiency is masked by artificial scaffolding.
Morphology as Cause, Not Reflex: Agreement, case, and inflection are often treated as drivers of syntactic behavior. In truth, they are epiphenomena- surface reflections of deeper computational necessities, like ripples on a river revealing the hidden flow beneath.
Typology as Exception: Cross-linguistic variation is often treated as anomalous rather than informative. These “exceptions” are in fact windows into the constraints of the system, revealing what computation is necessary and what is contingent.
Our proseminar does not catalog. It diagnoses. It does not describe. It formalizes.
Imagine syntax not as a museum of structures, but as a laboratory of constraints, where every apparent irregularity is a signal of underlying necessity.
Every anomaly, every variation, is a proof of architecture, not a footnote of history.
We ask:
What if all previous models are merely shadows of a deeper computation, glimpses of an architecture that obeys mathematical, cognitive, and evolutionary laws?
In this space, you will abandon comfort. You will see the skeleton of possibility, not the flesh of description. You will formalize what was previously implicit, and you will measure the architecture against the mind’s finite, precise, and beautiful computational constraints.
Architecture, Not Inventory
Language is not a collection of labels or a catalog of surface forms. It is a computational system, a structure of constraints and possibilities, a grammar of what can be thought and expressed, not merely what is spoken.
Merge and Symmetry-Breaking: From the simplest binary operation, Merge, emerges the full hierarchy of human thought. Symmetry is not incidental; it is violated by necessity, producing the asymmetries that underlie subject positions, verb placement, and dependency chains.
Locality and Visibility: The brain is a finite processor. Dependencies, movement, and visibility obey cognitive constraints, not arbitrary rules. Locality is the logic of tractable search, the mind’s solution to combinatorial explosion.
Linearization and Externalization: Word order is not a fundamental property of syntax. It is an emergent phenomenon, imposed by the sensory-motor interface. Linearization is the projection of hierarchical thought into the temporal world.
Minimal Structure: Complexity is a consequence, not a design choice. Economy produces rich surface structures from minimal derivations, revealing the efficiency of the system. Bloated hierarchies are illusions of descriptive laziness.
Epiphenomenal Features: Morphology, agreement, and feature-checking are reflexes, not drivers. They emerge from derivations, not from preordained forces. What appears causal is often a signal of derivational inevitability.
Computational Bounds: Human syntax resides in the Mildly Context-Sensitive sweet spot: powerful enough to express recursive thought, yet constrained enough to be learnable, processable, and evolutionarily feasible. The mind negotiates possibility with precision.
Our ambition is not to inventory what languages do.
It is to formalize what they can do, to map the grammar of possibility, and to see language as a window into the architecture of human thought.
The Epistemic Crisis- The Systemic Failure
Slide this into your mind: the Icelandic Dative Subject. Formal notation captures the tension:
This is not a curiosity. It is a falsification.
Traditional GB-theory calls this an “exception.” In this discussion, we recognize it as proof of theoretical collapse.
If our models cannot account for this without ad hoc patches, they are relics of descriptive ambition, not explanatory power.
This is the live anomaly- a pulse in the system that refuses to conform, a challenge to every assumption you hold. It is a laboratory of impossibility, asking not how the system works, but why the architecture permits this configuration at all.
Here, the discussion departs from pedagogy. Let us try to formalize, stress-test, and reconstruct:
Derive the anomaly using Minimalist Grammars
Expose the assumptions hidden in EPP, agreement, and feature-checking
Recast theory so that anomalies are opportunities for universal insight, not exceptions to memorize
The Icelandic Dative Subject is not an obstacle. It is a doorway- a chance to see the limits of orthodoxy and the shape of architecture yet to be formalized.
In this space, every exception is a theorem in disguise, every failure a call to rebuild the theoretical universe.
The Mathematical Core- Mapping the Sweet Spot
Human language is not a free-for-all of symbols, nor a casual sequence of rules. It exists in a precisely tuned computational niche, the Mildly Context-Sensitive (MCS) domain, where expressivity meets tractability.
Context-Free Boundaries: A system too weak cannot capture the intricacies of human syntax. Cross-serial dependencies, nested embeddings, and long-distance relations cannot be realized in a purely context-free system. Limitation begets incompleteness.
Fully Context-Sensitive Chaos: A system without bounds becomes infinite, unmanageable, and evolutionarily impossible. The search space explodes; processing collapses; learning is infeasible. Human cognition cannot sustain unbridled computational freedom.
The Sweet Spot: Language is the mathematical compromise, a system endowed with just enough power to encode recursive, hierarchical thought, yet constrained enough to remain learnable, processable, and evolutionarily viable. This is not accident, it is architecture.
We formalize this architecture rigorously, using the tools that reveal the invisible skeleton of computation:
Set Theory: The simplicity of Merge captures infinite generativity:
From this single operation, hierarchies emerge, complexity unfolds, and derivational possibilities proliferate.
Graph Theory: Phases, visibility, and locality are nodes and edges of cognitive necessity, not arbitrary decorations. They encode the constraints of memory, attention, and interface interaction.
Minimalist Grammars (MGs): The formal machinery that traces human syntax along the MCS frontier, handling cross-serial dependencies, movement patterns, and derivational constraints with precision.
Every derivation is a proof, a formalization of possibility. Every anomaly is a theorem, demanding explanation or adaptation. Syntax ceases to be anecdotal; it becomes law, a reflection of mathematical inevitability within human cognition.
To study syntax is not to catalog words and rules.
It is to witness the architecture of possibility, the boundary where thought and computation converge.
Language Evolution and Cognitive Constraints- The Architecture of Possibility
Why this architecture, and not another? This question is not speculative, it is the hinge of theory itself.
Merge as Hierarchical Cognition: Merge did not arise merely to communicate. It is an expression of the brain’s ability to organize thought hierarchically, to compute relations among abstractions. Language reflects cognitive necessity, not arbitrary design.
Locality as a Cognitive Imperative: Constraints on movement, dependencies, and visibility are not theoretical artifacts, they are manifestations of memory and processing limits. The brain cannot search infinitely; it must optimize.
Linearization as Interface Imposition: Word order is not innate to syntax. It emerges from the demands of the sensory-motor system, a necessary projection of hierarchical computation into linear action.
Phase Dynamics as Optimization: The division of structures into phases is not aesthetic, it is computationally efficient, allowing the mind to incrementally process and transfer information. Complexity is parceled, tractability preserved.
The Sweet Spot of Computation: Human language occupies a precisely tuned niche. Enough power to encode infinite thought, yet constrained enough to be learnable, processable, and evolutionarily feasible. Too little, and expression collapses; too much, and computation becomes intractable.
Language is thus neither arbitrary nor limitless, it is the mathematical consequence of cognition, evolution, and embodied constraints. To understand it is to witness the mind negotiating possibility itself.
The Proseminar as Laboratory
You will not receive knowledge. You will confront it, stress it, and make it yield insight.
Formalize the Unseen: Every derivation in Minimalist Grammar is a trace of computation, a fragment of the mind’s architecture. To formalize it is to translate thought into law.
Critique Orthodoxy: From Chomsky’s foundational models to contemporary Cartography, every system is a hypothesis. You will expose its limits, reveal its hidden assumptions, and measure it against anomalies it cannot explain.
Engineer Solutions: Anomalies are not curiosities, they are challenges to formal inevitability. Your task is to devise solutions, not patches; to render coherent what theory currently cannot.
Produce Knowledge, Not Exercises: Peer review is not a formality. It is the act of entering the arena of intellectual rigor, producing work of journal-ready quality, capable of shaping the field.
This is a laboratory, but not of chemicals or machines. It is a laboratory of the mind, where the architecture of language is stress-tested under the weight of computation, cognition, and cross-linguistic diversity. Here, the mind itself is the experimental apparatus, and every derivation is both a probe and a revelation.
In this space, theory is alive, falsifiable, and malleable.
You are no longer students, you are architects of the possible.
Frontiers of Syntax
The journey through the architecture of language is not a linear progression of topics, nor a checklist of technical milestones. It is a voyage into the very principles that constrain thought itself. Each phase, whether we wrestle with the computational spine, confront the dictates of the interfaces, or formalize the seeming impossibilities of Dyirbal, Icelandic, and Mohawk, reveals more than a linguistic pattern. It exposes the invisible scaffolding of the mind, the mathematical inevitabilities that shape expression and comprehension alike.
To explore the computational spine is not merely to reduce syntax to Merge and Third Factor principles. It is to ask: What makes a hierarchical structure possible at all? The simplicity of recursive Merge belies the profound truth that infinite expressivity can emerge from minimal principles, and that complexity is a byproduct of elegance, not brute addition.
When we turn to interface constraints, we encounter the tension between the abstract and the tangible. Language does not exist in a vacuum; it is sculpted by the capacities of the ear and the vocal tract, by memory limits and processing pressures. Here, syntax reveals itself as a negotiation between the ideal of thought and the demands of embodiment, a system that is both liberated and constrained, a delicate dance of form and function.
The formalization of typologically diverse phenomena, the idiosyncrasies of Dyirbal, the quirks of Icelandic, the subtleties of Mohawk, is a meditation on universality and particularity. Each anomaly is a window into what the mind allows, what it resists, and what it inevitably computes. By seeking a single algebraic model, we are not flattening diversity; we are illuminating the architecture that makes diversity possible.
Finally, the derivational frontier is a call to courage: to confront what current theory cannot yet explain. Here, students are not passive learners; they are architects of insight, probing anomalies as laboratories of discovery. Each formalization is a hypothesis, each counterexample a theorem, each derivation a trace of the invisible constraints that make human thought computationally realizable.
Taken together, these reflections suggest a larger truth: syntax is not a collection of arbitrary rules, but a reflection of the computational, cognitive, and evolutionary necessities of the mind. Our task is not simply to describe, but to understand why the system is as it is, and why it could be no other way.
Language, in its architecture, is both a mirror and a map: a mirror of human cognition, a map of the formal possibilities inherent in finite minds operating under principled constraints. To traverse this frontier is to participate in the creation of knowledge itself, to move beyond description and into the realm of theoretical discovery in real time.
The Mental Posture of the Linguistic Architect
To engage with the architecture of language is to adopt a stance of intellectual vigilance. Knowledge is not received; it is interrogated, formalized, and tested against the limits of computation and cognition.
Skepticism toward authority: No derivation is sacred. Every claim, every tree, every assumed principle is a candidate for falsification. Orthodoxy is a starting point, not a boundary. The mind of the linguist must remain radically independent, seeing anomalies not as obstacles but as illuminations of hidden principles.
Precision in formalism: Thought without implementable form is mere rhetoric. Every insight must withstand translation into set-theoretic, graph-theoretic, or computational representation. Language is a system of computation; understanding it requires mathematical rigor, not narrative description.
Ambition in creativity: Anomalies are treasures, not inconveniences. The irregular, the exception, the counterexample are signposts toward deeper architecture, revealing the constraints and freedoms that underlie all human languages. True discovery lies in embracing the unexpected and deriving universal insight from the particular.
Interdisciplinary thinking: Syntax, computation, cognition, and evolution are inseparable. The architecture of language is not an isolated artifact; it is the intersection of physical brains, abstract computation, and evolutionary possibility. To comprehend it is to navigate these domains simultaneously, allowing each to illuminate the others.
In sum, the linguistic architect’s posture is one of alert curiosity, disciplined rigor, and expansive imagination. It is a stance that transforms anomalies into insight, computation into explanation, and the study of language into a philosophical and scientific act of discovery.
The Architect’s Command- Closing Provocation
Take sentences that defy theory. Do not ask how. Do not search for comfort in existing frameworks.
Treat those sentences as fissures in the edifice of orthodoxy, challenge yourself, and take it as alluring invitations to rewrite what is considered possible.
The solutions to these sentences are not in textbooks, journals, or lectures.
They exist only in the architecture we should build, formalize, and dare to inhabit.
Every day while studying syntactic theory, we should obliterate assumptions, confront anomalies as opportunities, and render visible the invisible constraints of human cognition.
Let us endeavor to discover not merely what language is, but why it must be that way and why no other arrangement could satisfy the mind’s imperatives.
This is the frontier. This is the laboratory. And here, the impossible is merely the unformalized.
'Let us proceed':
1. Merge Is Not Free: Symmetry, Labeling, and the Computational Instability of Syntax
Architectural Question (Non-Negotiable)
Why does the simplest combinatorial operation in human language-Merge-immediately generate objects that are computationally unstable, and what does this reveal about the architecture of the language faculty?
This question is not pedagogical. It is architectural.
If syntax is built from a single operation (Merge), then any systematic pressure that forces movement, labeling, phases, or features must be traceable to the internal consequences of Merge itself, not to stipulative mechanisms layered on top of it.
The central claim here is:
Merge, by itself, systematically creates symmetry.
Symmetry is computationally fatal at the interfaces.
Much of syntax exists to destroy symmetry.
Minimalism’s Hidden Assumption: Merge Is Benign
The Standard Picture
In its strongest minimalist formulation:
Merge = set formation
Merge is:
unbounded
structure-building
cost-free
interface-legible once labeled
Problems (movement, agreement, phases) are often treated as secondary complications.
This picture is misleading.
The Core Problem Minimalism Often Downplays
Merge does not merely build structure.
Merge creates symmetry by default.
The object {XP, YP} contains:
no projection
no asymmetry
no head–dependent relation
no clear CI-interpretive instruction
If syntax interfaces with systems that require asymmetric instructions (predicate–argument, operator–variable, scope, selection), then {XP, YP} is computationally uninterpretable.
This is not a PF problem.
This is a core architectural instability.
Symmetry as a Computational Crisis
Symmetry Defined (Formally)
A syntactic object SO is symmetric iff:
Its immediate constituents are of the same type
No asymmetric dominance relation is available for interpretation
Formally:
Examples:
{NP, NP}
{XP, YP}
{DP, TP} (under some assumptions)
Symmetry is not rare. It is the default outcome of Merge.
Why Symmetry Is Illicit
At the Conceptual–Intentional (CI) interface, interpretation requires:
function–argument mapping
scope relations
predication
thematic asymmetry
But a symmetric object provides no instruction for:
which element projects
which element is selected
which element is interpreted as predicate
Thus:
Symmetry is not a descriptive inconvenience.
It is an interpretive crash.
Labeling as Symmetry Destruction
The Labeling Algorithm (LA) Revisited
Following Chomsky (2013, 2015, 2021):
Labels are not primitives
Labels are computed at CI
A label is recoverable only if:
one element is a head, or
one element has moved, or
the elements share prominent features
Crucially:
Labeling is not about naming structures.
It is about rescuing structures from symmetry.
Why {XP, YP} Cannot Be Labeled
Neither XP nor YP provides a projection
No minimal search yields a unique label
CI receives no asymmetric instruction
Hence {XP, YP} must be repaired.
Movement as a Computational Repair Strategy
Reframing Movement
Standard view:
Movement is triggered by features (EPP, edge features, probes)
Architectural view:
Movement exists because symmetry is otherwise fatal.
Movement removes one term from the symmetric configuration, leaving:
This object is asymmetric and labelable.
Subject Movement to Spec-TP (Derivation)
Consider the vP–TP transition.
Initial Merge:
This is a symmetric {XP, YP} configuration.
No label is recoverable.
Repair:
DP moves to Spec-TP
Lower copy becomes invisible for labeling
Result:
Label = TP
Interpretation proceeds.
Conclusion:
Subject movement is forced, not optional.
It is not feature-driven.
It is a labeling necessity.
Formalizing the Argument
Let Merge be defined as unordered set formation.
Let Label(SO) be a partial function defined only if:
SO contains a unique minimal head, or
SO is asymmetric due to copy invisibility.
Then:
For any {XP, YP}:
Movement transforms:
Now:
Movement is thus a precondition for labeling, not an epiphenomenon.
Empirical Stress Test: When Movement Does Not Happen
Apparent Counterexamples
Languages with:
pro-drop
null expletives
subject-in-situ
Superficially, these seem to violate the argument.
Resolution
The crucial point:
It is not overt movement that matters, but asymmetry at CI.
Possible strategies:
Null operators
Feature sharing
Head–spec unification
Alternative symmetry-breaking mechanisms
The architecture predicts variation in repair strategies, not the absence of repair.
Failure Modes of Competing Accounts
Feature-Driven EPP
Problems:
Circular (features introduced to force movement that already happens)
Non-explanatory
Typologically brittle
Cartographic Projections
Problems:
Encode symmetry resolution into structure
Miss the computational origin
Overgenerate functional layers
The symmetry-based account:
Predicts why structure is needed
Explains where movement is obligatory
Allows constrained variation
Architectural Consequences
This lecture commits us to the following strong theses:
Merge is computationally unstable
Syntax is a repair system
Movement is not optional
Features are derivative, not causal
Variation lies in symmetry-breaking strategies, not in Merge itself
Squib Trajectories
Any of the following is LI-viable if executed rigorously:
“Subject Movement Without EPP: A Labeling-Driven Account”
\“Why {XP, YP} Is Uninterpretable: Symmetry as the Core Problem of Syntax”
“Movement as Interface Repair: Evidence from Pro-Drop Languages”
Reviewer-B pressure points:
Formal clarity of labeling
Cross-linguistic robustness
Avoiding covert stipulation
Exercises
Exercise 1 (Formal)
Provide a derivation where:
{XP, YP} is labeled without movement
Show exactly what breaks or what additional assumptions are required
Exercise 2 (Typological)
Identify a language where subject movement appears optional.
Demonstrate how symmetry is nevertheless broken.
Exercise 3 (Computational)
Translate the derivation into a Minimalist Grammar fragment.
Specify where the asymmetry is encoded.
Merge & Symmetry Breaking
Construct derivations showing symmetry-breaking when two phrases merge.
Demonstrate why subject movement to Spec,TP is necessary for labeling.
Take-Home Architectural Claim
The deepest problem of syntax is not displacement, agreement, or word order.
It is symmetry.
Everything else exists to destroy it.
2. Locality as Internalized Computation: Why Syntax Cannot See Arbitrarily Far
Architectural Question
Why does the human language faculty enforce locality constraints so rigidly, even when long-distance dependencies are semantically and communicatively useful?
Or, stated more sharply:
Is locality a grammatical principle, or a computational necessity imposed by the architecture of Merge-based systems?
This post advances a strong thesis:
Locality is not a syntactic “constraint.”
It is the internalization of computational limits within the derivation itself.
The Illusion of Locality as a Rule
Traditional Framing (GB → Minimalism)
Locality is often introduced as a set of independent constraints:
Subjacency
ECP
Minimal Link Condition (MLC)
Relativized Minimality
Phase Impenetrability Condition (PIC)
These are usually taught as rules restricting movement.
But this framing is theoretically suspicious:
Why do so many distinct constraints converge on the same effects?
Why do violations cluster around similar structural configurations?
Why are they gradient, repairable, and interface-sensitive?
Reframing Locality: A Computational Problem
The Core Hypothesis
Locality emerges because derivations must remain computationally tractable at every step.
Key assumptions:
Syntax is incremental, not global.
Interface interpretation applies cyclically.
Unbounded search is computationally illicit.
Merge creates objects that must be immediately interpretable or sealed off.
Locality is therefore forced, not stipulated.
Strict Cyclicity Revisited
What Strict Cyclicity Really Is
Strict cyclicity is often described descriptively:
Operations apply only at the root.
Architectural reinterpretation:
Only the most recently formed object is visible to the computational workspace.
Older material must be:
spelled out, or
frozen, or
made invisible to further computation
This is not a grammatical preference; it is memory management.
The Activity Condition as Visibility Control
Traditional View
An element can move only if it is “active” (i.e., has unchecked features).
This has always been conceptually murky.
Architectural Reanalysis
Activity = computational visibility.
An element is active if:
it remains accessible in the current workspace
it has not been sealed by cyclic transfer
it is not buried under completed interpretation
In other words:
Inactivity is computational garbage collection.
Minimality as Search Economy
Minimality Is Not a Linguistic Preference
Minimality effects:
Closest goal wins
Skipping is disallowed
This follows if:
Search is local and bounded by workspace limits.
Global search would:
explode computational complexity
violate real-time processing constraints
undermine learnability
Minimality is therefore a complexity-minimizing heuristic, internalized as grammar.
Phases as Computational Firewalls
Why Phases Exist
Phases (vP, CP) are often treated as:
arbitrary domains
remnants of GB bounding nodes
Architectural claim:
Phases are computational firewalls that prevent unbounded derivational growth.
They:
limit memory load
enforce cyclic interpretation
block uncontrolled search
PIC is therefore not a “constraint,” but a firewall rule.
Formalizing Locality
Syntax as a Directed Acyclic Graph (DAG)
Nodes = syntactic objects
Edges = Merge relations
Workspace = active subgraph
Locality emerges if:
search is restricted to the active frontier
sealed nodes are removed from the graph
Long-distance movement is possible only if intermediate nodes remain active (escape hatches).
Apparent Locality Violations
Long-Distance Dependencies
Wh-movement, topicalization, relativization.
Crucial observation:
They are not violations, they are carefully engineered survivals.
Intermediate landing sites:
keep the element visible
prevent phase sealing
maintain computational accessibility
Ergativity, Clitic Doubling, and Atypical Locality
In many ergative or clitic-doubling systems:
locality behaves differently
intervention effects are altered
Prediction:
These systems manipulate visibility, not locality itself.
This is a fertile area for student squibs.
Failure Modes of Alternative Accounts
Constraint Lists
Redundant
Non-explanatory
Typologically brittle
Feature Proliferation
Features encode what computation already enforces
Leads to theoretical inflation
Architectural Consequences
Locality is emergent, not primitive
Syntax internalizes computational limits
Phases are memory-management devices
Minimality = bounded search
Variation lies in visibility mechanisms
Squib Trajectories
LI/NLLT squibs:
“Locality Without Constraints: A Computational Account of Movement”
“Phases as Firewalls: Why Syntax Forgets”
“Activity as Visibility: Rethinking the Activity Condition”
Reviewer will demand:
explicit formalization
typological reach
computational plausibility
Exercises
Exercise 1 (Formal)
Provide a derivation where violating PIC would require global search. Show precisely where complexity explodes.
Exercise 2 (Typological)
Find a language where intervention effects are weak or absent. Explain how visibility is preserved.
Exercise 3 (Computational)
Translate a phase-based derivation into a graph with sealed subgraphs.
Locality as Internalized Computation
Provide a derivation where violating PIC requires global search and show the computational explosion.
Analyze a language with non-standard locality patterns (e.g., clitic-doubling or ergative systems) and formalize the derivation.
Translate a phase-based derivation into a graph showing sealed subgraphs.
Take-Home Architectural Claim
Locality is not a rule that syntax obeys.
It is a fact about what syntax can afford to compute.
3. Linearization and the Sensory–Motor Interface: Why Syntax Is Orderless
Architectural Question
Why does a strictly hierarchical, order-free computational system produce such rigid and typologically diverse word orders at the surface?
Or more provocatively:
If syntax is hierarchical and symmetric, where does linear order come from, and why is it so fragile?
This post advances the following thesis:
Linear order is not computed by syntax.
It is imposed at the Sensory–Motor (SM) interface under severe external constraints.
The Illusion of Word Order as Syntactic
The Traditional View
Much of syntactic theory treats order as internal:
Head–parameter (GB)
Kayne’s Linear Correspondence Axiom (LCA)
Directionality encoded in syntax
Cartographic fixed orders
These approaches share a hidden assumption:
Syntax “knows” left and right.
This assumption is architecturally suspicious.
The Case for Orderless Syntax
Merge Is Order-Free
By definition:
Sets are unordered.
Thus:
No precedence
No adjacency
No directionality
If order exists, it must be added later.
Conceptual Argument
Syntax interfaces with:
CI (meaning)
SM (externalization)
Only SM:
requires temporal sequencing
operates in real time
is modality-specific (speech, sign)
Thus:
Order is a requirement of externalization, not computation.
Kayne’s LCA
What LCA Achieves
The LCA derives linear order from asymmetric c-command.
Merits:
Unifies hierarchy and order
Explains many rigid patterns
The Architectural Cost
LCA requires:
syntax to compute precedence
antisymmetry as a universal law
widespread covert movement
Problem:
It internalizes an interface problem into core syntax.
This violates minimalist economy.
PF-Driven Linearization
Alternative Architecture
Proposed view:
Syntax builds hierarchical objects
Objects are transferred cyclically
PF linearizes structures under:
prosodic constraints
processing limitations
modality-specific requirements
Word order = solution space, not rule.
Free Word Order Is Not Free
Case Study: Scrambling Languages
Languages like:
Urdu
Hindi
Warlpiri
Latin
Appear to allow “free” word order.
But:
Information structure constrains order
Prosody matters
Scope relations remain stable
Prediction:
Linearization is flexible only where SM constraints allow multiple solutions.
Minimalist Grammars and Linearization
Why MGs Matter
MGs:
separate hierarchical generation from linear output
allow multiple surface strings from one derivation
formally capture externalization
An MG can generate:
the same tree
multiple linearizations
This matches typological reality.
Formal Sketch: Linearization as Optimization
Let:
H = hierarchical structure
L = linearization function
C = set of SM constraints
Then:
Why Word Order Changes So Easily
Diachronic Fragility
Word order changes faster than:
case systems
agreement
movement options
Explanation:
Interfaces change faster than core computation.
Externalization is evolutionarily plastic.
Apparent Counterexamples
Rigid Word Order Languages
English, French.
But rigidity correlates with:
loss of morphology
heavier SM reliance on order
processing pressures
Order compensates for lost cues.
Failure Modes of Syntactic Order Accounts
Parameter Proliferation
Directionality parameters multiply
Typology becomes stipulative
Cartographic Rigidity
Encodes interface outcomes as syntactic primitives
Predicts far less variation than observed
Architectural Consequences
Syntax is order-free
Linearization is external
Variation arises at SM
Word order is optimization, not rule
Syntax–PF separation is real
Squib Trajectories
Potential LI/NLLT squibs:
“Order Without Syntax: Externalization as Optimization”
“Why Free Word Order Is Not a Syntactic Property”
“Linearization Variability and the Architecture of Externalization”
Reviewer pressure:
formal clarity
typological depth
computational explicitness
Exercises
Exercise 1 (Formal)
Provide two distinct linearizations from the same hierarchical structure using different SM constraints.
Exercise 2 (Typological)
Analyze a language with scrambling. Identify which cues compensate for flexible order.
Exercise 3 (Computational)
Implement an MG fragment that generates multiple word orders from one derivation.
Linearization & Sensory–Motor Interface
Generate two or more linearizations from the same hierarchical structure using different SM constraints.
Analyze a scrambling language and identify cues that allow multiple surface orders.
Implement a Minimalist Grammar (MG) fragment generating multiple linearizations for one derivation.
Take-Home Architectural Claim
Syntax does not order words.
It builds relations.
Order is imposed only when language leaves the mind.
At this point, we have established three core architectural theses:
Merge creates symmetry → repair
Locality = internalized computation
Order = interface imposition
4. Cartography vs. Minimalist Roots: Do We Need Rich Functional Structure?
Architectural Question
Does the human language faculty encode richly articulated functional hierarchies, or do such hierarchies emerge from derivational economy and interface pressures?
More sharply:
Are functional sequences part of Universal Grammar, or are they epiphenomena of how syntax survives computation and externalization?
This post treats Cartography not as a rival “school,” but as a stress test for Minimalist architecture.
What Cartography Claims
Core Commitments
Cartography (Cinque, Rizzi) claims that:
Functional structure is highly articulated
The hierarchy of projections is universal and invariant
Cross-linguistic variation is largely movement within a fixed spine
Canonical examples:
CP split into Force–Top–Foc–Fin
Adverbial hierarchy (Mood > Tense > Aspect > Voice…)
Cartography’s empirical achievements are undeniable.
The Minimalist Challenge
Minimalism asks a deeper question:
Why would UG encode dozens of ordered functional heads?
From an architectural standpoint:
Each head increases search space
Each projection raises acquisition burden
Each fixed order hard-codes interface outcomes
Two Competing Explanations of Order
Cartographic Explanation
Order reflects innate syntactic hierarchy
Deviations = movement
Stability = evidence for hard-coding
Minimalist Explanation
Order reflects:
labeling resolution
phase boundaries
information-structural pressures
SM optimization
Stability emerges from recurrent repair strategies
The same surface facts, radically different architectures.
Microvariation as a Diagnostic Tool
The Cartographic Prediction
If hierarchies are innate:
microvariation should be limited
deviations should be movement-based
unattested orders should be impossible
The Minimalist Prediction
If hierarchies emerge:
microvariation should cluster around:
interface pressures
economy trade-offs
optional repairs
Empirical question:
Which prediction is borne out?
Derivational Economy vs. Structural Encoding
Economy Argument
If two analyses generate the same interpretive output:
one using:
multiple functional heads
the other using:
fewer heads + derivational operations
Minimalism favors the latter.
Cartography must therefore justify why structure is cheaper than computation.
Formal Pressure: Overgeneration
Rich hierarchies risk:
massive overgeneration
reliance on movement filters
post-hoc licensing
Minimalist roots seek:
minimal structure
maximal explanation
Case Study: Topic–Focus Articulation
Cartographic Account
Dedicated TopP and FocP
Fixed ordering
Cross-linguistic uniformity
Minimalist Root Account
Topic and focus arise via:
edge effects
labeling needs
discourse-driven movement
No need for fixed projections
Prediction:
If discourse context changes, structural reflexes should shift.
Empirically supported in many languages.
Formalizing the Debate
Let:
S = syntactic structure
I = interpretive output
C = interface constraints
Cartography:
Minimalist Roots:
Architectural economy favors minimizing S, not C.
Where Cartography Is Strong
This lecture does not dismiss Cartography.
Cartography excels at:
descriptive precision
cross-linguistic pattern discovery
hypothesis generation
Minimalism depends on cartographic results, but resists reifying them.
Where Minimalism Must Deliver
Minimalism must:
derive hierarchies, not deny them
formally model emergence
explain stability without hard-coding
Failure to do so cedes ground to Cartography.
Hybrid Positions
Some contemporary work:
accepts cartographic results
but seeks minimalist explanations
This is the productive frontier.
Squib Trajectories
Squib topics:
“Are Functional Hierarchies Primitive or Emergent?”
“Microvariation as Evidence Against Fixed Functional Spines”
“Deriving Cartographic Orders Without Cartographic Structure”
Reviewer focus:
explanatory depth
avoidance of strawman arguments
empirical sharpness
Exercises
Exercise 1 (Derivational)
Take a cartographic hierarchy and remove half the projections. Show what must be derived instead.
Exercise 2 (Typological)
Find a language where cartographic ordering is violated. Explain how interpretation survives.
Exercise 3 (Formal)
Model topic/focus without dedicated projections using labeling and movement.
Cartography vs Minimalist Roots
Remove half the functional projections in a cartographic hierarchy; derive the same interpretive output minimally.
Identify a language that violates standard cartographic order; explain interpretive survival.
Model topic/focus distinctions without dedicated functional projections using labeling and movement.
Take-Home Architectural Claim
Cartography maps the surface.
Minimalism must explain the map.
5. Phases, Features, and the Death of Agreement: Why Syntax Does Not Check Anything
Architectural Question
Why does syntax appear to traffic in features that are “checked,” “valued,” or “deleted,” and what would the architecture look like if features were not causal at all?
More sharply:
Is agreement a grammatical operation, or is it a surface reflex of cyclic computation and visibility?
This post defends a strong thesis:
Features do not drive syntax.
They annotate outcomes of computation.
Agreement is epiphenomenal.
Feature-Checking: A Convenient Fiction
The Standard Story
In mainstream Minimalism:
Heads carry uninterpretable features
Probes search for goals
Agreement “licenses” movement
Derivations crash if features remain unchecked
This story has pedagogical appeal, but deep architectural problems.
The Core Problem with Features
Circularity
Features are introduced:
to force movement
to explain agreement
to enforce locality
But these phenomena already follow from:
labeling
cyclicity
visibility
Thus:
Features explain only what they are designed to explain.
Proliferation
Every new phenomenon introduces:
new features
new probes
new stipulations
This violates:
economy
learnability
typological restraint
Phases Reinterpreted
Traditional View
Phases:
define Spell-Out domains
enforce PIC
constrain extraction
Architectural Reanalysis
Phases are computational compression points.
They:
reduce workspace size
seal off completed structures
regulate visibility
No features required.
Visibility Without Features
Activity Revisited
An element is “active” if:
it remains visible
it has not been sealed by phase transfer
it is on the derivational frontier
No need for:
unvalued features
licensing conditions
Visibility alone explains:
movement possibilities
intervention effects
freezing
Agreement Without Agreement
Empirical Cracks
Agreement is:
optional in many languages
partial or defective
absent without interpretive loss
Examples:
default agreement
clitic doubling
non-agreeing subjects
If agreement were causal, these systems should crash.
They don’t.
Preminger’s Challenge
Preminger (2014, 2025):
Agreement failure does not cause ungrammaticality
Agreement is not obligatory
Syntax survives without it
This is devastating for feature-checking.
Reframing Agreement
Agreement is a PF-visible reflex of shared accessibility within a phase.
If two elements:
are simultaneously visible
share structural proximity
Then:
morphological concord may be realized
If not:
default morphology appears
No crash. No checking.
Formal Sketch: Agreement as Correlation
Let:
V(x) = visibility of element x
P = phase domain
Agreement holds if:
No valuation function required.
D-Pronouns, Pro-Drop, and Incorporation
Traditional Feature Story
Pro-drop licensed by agreement
Null pronouns need rich features
Phase-Based Story
Null arguments arise when referential material is recoverable at CI.
Dynamic phases allow:
referential recovery
incorporation
pronominal silence
Agreement may accompany this, but does not license it.
Failure Modes of Feature-Based Accounts
Overgeneration
Predict obligatory agreement
Fail on defective systems
Interface Blindness
Treat morphology as syntactic
Ignore PF variability
Architectural Consequences
Features are descriptive, not causal
Phases regulate visibility
Agreement is optional
Movement is independent of agreement
Syntax computes structure, not morphology
Squib Trajectories
Squibs:
“Agreement Without Features”
“Why Syntax Does Not Check Anything”
“Phases as Visibility Domains”
Reviewer demands:
explicit derivations
serious engagement with Preminger
typological reach
Exercises
Exercise 1 (Formal)
Construct a derivation where agreement fails but syntax converges.
Exercise 2 (Typological)
Analyze a language with default agreement. Explain convergence without features.
Exercise 3 (Theoretical)
Remove feature-checking from a standard Minimalist derivation. What survives?
Phases, Features, and the Death of Agreement
Construct a derivation where agreement fails but the syntax converges successfully.
Analyze a language with default or defective agreement; explain convergence without feature-checking.
Remove feature-checking from a standard derivation; evaluate which operations survive.
Take-Home Architectural Claim
Syntax does not check features.
It manages visibility.
Agreement is what we see when morphology catches up.
6. Mathematical & Computational Syntax: Is Human Language Mildly Context-Sensitive?
Architectural Question
What is the exact computational power of human syntax—and how do Merge, locality, and phases constrain it?
More sharply:
Is the human language faculty strong enough to escape context-freeness, but weak enough to avoid computational intractability?
This post defends a carefully bounded thesis:
Human syntax occupies a narrow computational niche:
stronger than Context-Free Grammars, weaker than Context-Sensitive Grammars.
Why Computational Adequacy Is Not Optional
The Minimalist Obligation
If syntax is:
recursive
hierarchical
generative
Then it must be:
formally characterizable
computationally implementable
learnable under finite resources
Any theory that cannot be computationally modeled is architecturally incomplete.
Context-Free Grammars (CFGs): Too Weak
What CFGs Can Do
CFGs capture:
basic recursion
simple embedding
local dependencies
But they fail on:
cross-serial dependencies
multiple agreement chains
non-nested dependencies
Canonical Counterexample: Swiss German
Pattern:
Jan säit dass mer d’chind em Hans es huus haend wele laa hälfe aastriiche
Verbs and objects align cross-serially, not hierarchically nested.
CFGs cannot generate this pattern.
Context-Sensitive Grammars (CSGs): Too Strong
CSGs can generate:
any dependency
arbitrary copying
global constraints
But:
parsing is PSPACE-complete
acquisition becomes implausible
overgeneration is unconstrained
Human syntax shows discipline, not brute power.
The Mildly Context-Sensitive (MCS) Hypothesis
Defining MCS Languages
A grammar is MCS if it:
Properly contains CFGs
Handles cross-serial dependencies
Has polynomial-time parsing
Preserves locality
This is the Goldilocks zone.
Minimalist Grammars (MGs)
Why MGs Matter
Stabler (1997, 2011, 2024):
MGs:
formalize Merge and Move
derive displacement without transformations
are equivalent in power to Tree-Adjoining Grammars (TAGs)
Crucially:
MGs generate exactly the class of mildly context-sensitive languages.
This is not an accident.
Formal Core: MG Operations (Sketch)
MGs use:
Merge = structure-building
Move = feature-driven reordering (formally encoded)
Lexicalized derivations
Key result:
MG derivations are computable in polynomial time.
Thus:
powerful enough
constrained enough
Reconciling MGs with Architecture
Recall earlier section of this post/discussion:
Merge creates symmetry (1)
Locality restricts search (2)
Linearization is external (3)
Structure is minimal (4)
Features are epiphenomenal (5)
MGs align with this if features are reinterpreted as bookkeeping, not causal triggers.
Swiss German Revisited (MG Sketch)
MG derivation:
multiple verbal projections
parallel Merge operations
controlled interleaving via Move
Key point:
No global search, no unconstrained copying.
The dependency is managed, not computed freely.
Complexity Boundaries
Why Syntax Avoids Hard Problems
Human syntax avoids:
unrestricted copying
mirror dependencies
global counting
These push grammars into:
NP-hard
PSPACE-complete territory
Prediction:
No natural language requires such power.
This is empirically supported.
Apparent Challenges
Multiple Agreement & Reduplication
Often cited as pushing beyond MCS.
But careful analysis shows:
morphology absorbs complexity
syntax remains bounded
Interface again saves the architecture.
Failure Modes of Overpowered Syntax
Theories that allow:
unrestricted movement
global constraints
feature stacking
predict:
unattested languages
acquisition failure
processing collapse
Architectural Consequences
Syntax is computationally bounded
MGs are a faithful formalization
MCS is not arbitrary, it is forced
Interfaces absorb excess complexity
Human language is computationally elegant
Squib Trajectories
Squib topics:
“Why Human Syntax Is Mildly Context-Sensitive”
“Minimalist Grammars as a Model of the Language Faculty”
“Computational Limits as Grammatical Universals”
Reviewer pressure:
formal precision
clear complexity claims
no hand-waving
Exercises
Exercise 1 (Formal)
Provide an MG fragment generating a cross-serial dependency.
Exercise 2 (Comparative)
Show how a CFG fails on the same data.
Exercise 3 (Theoretical)
Identify a hypothetical language that would exceed MCS power. Argue why it is unattested.
Mathematical & Computational Syntax
Provide an MG fragment generating a cross-serial dependency.
Show how a CFG fails to generate the same dependency pattern.
Identify a hypothetical language exceeding mildly context-sensitive power; argue why it is unattested.
Take-Home Architectural Claim
Human language is computationally powerful, but only just enough.
Anything more would be unlearnable.
Anything less would be inexpressive.
Where the proseminar Now Stands
We have now built an internally coherent architecture:
Merge destabilizes structure
Locality internalizes computation
Linearization is external
Structure emerges, not hard-codes
Features annotate outcomes
Syntax is mildly context-sensitive
Architectural Question
Why does the human language faculty exhibit exactly the architecture we have uncovered?
The question is:
Why has evolution produced a faculty with these precise constraints—no more, no less?
Constraints from Evolutionary Plausibility
Cognitive Limitations
Working memory is bounded
Attention is limited
Incremental processing is necessary
Implications:
Locality, phases, and minimal structure are consequences of cognitive architecture, not arbitrary design.
Learnability Constraints
Children acquire language from finite input
Excessive feature proliferation or unrestricted power is unlearnable
Mildly context-sensitive derivational systems maximize learnability
Implication:
Syntax balances expressive power with learnability, a Goldilocks solution.
Symmetry and Merge: Evolutionary Roots
Merge as a Core Innovation
Merge produces structured recursion
Recursion allows complex thought and planning
Symmetry-breaking is required for linearizable output
Evolutionary insight:
Merge may have arisen for general hierarchical cognition, not just language.
Locality as Computational Necessity
Why the Brain Enforces Locality
Global search is computationally expensive
Phases act as memory firewalls
Minimalist derivations are incremental, tractable, and learnable
Observation:
Locality is architecture, not a grammatical whim.
Linearization & the Sensory–Motor Interface
Evolutionary Perspective
Humans process speech incrementally
Externalization channels (speech, sign) impose temporal order
Word order emerges as a cognitive constraint on production
Implication:
Linearization is shaped by modality, not UG itself.
Features and Agreement as Epiphenomena
Morphology reflects interpretive recoverability
Agreement arises when SM interfaces can encode correlations
Features do not drive derivation
Evolutionary reading:
Feature-checking is a cognitive “side effect,” not a selective pressure.
Computational Bounds as Adaptive Constraints
Mild Context-Sensitivity = expressivity without computational overload
Cross-serial dependencies can exist, but arbitrarily complex dependencies cannot
Syntax aligns with polynomial-time parsing in the human mind
The language faculty is optimized, not maximal.
Typological & Cross-Linguistic Evidence
Languages vary in morphology, agreement, and order
Despite variation, all languages obey:
Merge-based hierarchies
Locality constraints
Mild context-sensitivity
Implication:
Constraints reflect cognitive and evolutionary limits, not historical accident.
Cognitive Frontiers
Recursion & Thought
Merge allows nested thought structures
Language may co-opt general hierarchical reasoning
Language is an interface between cognition and communication
Processing Pressure
Incremental parsing requires visibility and phases
Syntax is tuned to real-time comprehension
Evolutionary Hypotheses
Merge-first hypothesis
Hierarchy existed for general cognition
Externalization adapted later
Minimalist efficiency hypothesis
Locality and bounded derivation emerge from memory/processing constraints
Interface-driven architecture
Linearization, agreement, and functional hierarchies are shaped by SM/CI pressures
Squib Trajectories
Squibs:
“Phases as Evolutionary Memory Firewalls”
“Why Human Syntax Is Just Powerful Enough”
“From Merge to Linearization: An Interface-Driven Evolutionary Account”
Reviewer expects:
integration of typology, computation, and evolution
formal clarity connecting derivations to cognitive principles
critical engagement with interface pressures
Exercises
Exercise 1 (Comparative)
Analyze a language with optional agreement. Explain evolutionary advantage of epiphenomenal agreement.
Exercise 2 (Computational)
Model a derivation with multiple long-distance dependencies. Demonstrate tractability constraints.
Exercise 3 (Theoretical)
Formulate an argument why increasing syntax beyond MCS would be evolutionarily maladaptive.
Evolutionary & Cognitive Frontiers
Analyze a language with optional agreement; explain its evolutionary advantage.
Model a derivation with multiple long-distance dependencies; demonstrate tractability constraints.
Argue why increasing syntactic computational power beyond MCS would be maladaptive.
Take-Home Architectural Claim
Human syntax is the product of evolutionary compromise:
Powerful enough to express unbounded hierarchical thought
Constrained enough to be learnable, incremental, and processable
Minimal in structure, externalized flexibly, and free of superfluous features
In short: every quirk of syntax, Merge, locality, phases, linearization, feature-tracking, is an adaptation to cognitive and evolutionary reality.
| Core Insight | |
|---|---|
| 1 | Merge destabilizes structure |
| 2 | Locality is internalized computation |
| 3 | Word order is externalization |
| 4 | Structure emerges from economy, not hard-coded hierarchies |
| 5 | Features and agreement are epiphenomenal |
| 6 | Human syntax is mildly context-sensitive |
| 7 | Architecture reflects cognitive and evolutionary optimization |
We began this proseminar not to learn syntax, but to see it, to expose the invisible architecture that constrains, enables, and defines human language. Over the course of the year, we have probed, formalized, and stress-tested the limits of current theory, moving beyond description into the domain of architectural discovery.
The enterprise of syntax is no longer a catalog of forms; it is a laboratory of necessity. Every derivation we formalized revealed a fundamental truth about cognition: that language is both infinitely expressive and finitely realizable, a delicate compromise dictated by the architecture of the human mind. This is the domain of the Third Factor: universal principles of efficient computation that exist as much in the stars as in the synapse.
We confronted anomalies, Dyirbal ergativity, Icelandic dative quirks, Swiss-German cross-serial dependencies, not as curiosities, but as philosophical provocations. Each anomaly demanded that we formalize the impossible, revealing where the orthodoxy fails and where the computational spine of language must be refined.
From this crucible, the architecture of human language revealed itself:
Merge as the singular recursive engine of structure, its simplicity yielding hierarchical complexity.
Locality and phase dynamics as the inevitable reflections of finite memory, processing limits, and interface constraints.
Linearization as the projection of abstract computation into temporal reality—the moment thought becomes speech.
Epiphenomenal features as signals, not causes, reminders that description is never explanation.
Mildly Context-Sensitive boundaries as the sweet spot where the mind balances expressivity and tractability.
This proseminar has not merely explored language. It has performed an autopsy on the miracle, revealing that what we once called “rules” are, in truth, consequences of universal computational principles. The grammar of possibility is not invented, it is discovered.
As scholars, your work now occupies a space of original contribution. Each derivation, every formal proof, every anomaly resolved is a claim about the architecture of thought itself, capable of enduring scrutiny, capable of influencing theory worldwide.
The lesson is both simple and profound: language exists at the intersection of computation, cognition, and evolution. To understand it is to understand a fundamental architecture of the human mind. There are no sacred derivations, no inviolable hierarchies, only the constraints of possibility.
Every “impossible” sentence is an invitation. Every empirical counterexample is an opportunity. In this seminar, you have not merely mapped phenomena; you have charted the frontiers of human understanding.
We close not with certainty, but with provocation: the architecture of language is not fully known. Our work is the first step in a continuum of inquiry that will stretch theory, computation, and philosophy together.
Language is not a museum. It is a laboratory. It is a universe of constraints and possibilities. And the frontier moves only when we push it.
This is your charge: continue to formalize, challenge, and imagine. Stand where orthodoxy hesitates. Derive what no one thought possible. Let the architecture of language guide you, not merely as scholars, but as architects of thought itself.
The derivation is never truly finished. The anomalies await. The frontier awaits. And so do you.
Bibliography
- Baltin, M., & Collins, C. (Eds.). (1991). The handbook of contemporary syntactic theory. John Wiley & Sons.
- Bellina, A., & Servedio, V. D. (2025). Cognitive Limits Shape Language Statistics. arXiv preprint arXiv:2503.17512.
- Boeckx, C. (2021). Reflections on language evolution: From minimalism to pluralism. Language Science Press.
- Boeckx, C. (2009). Understanding minimalist syntax: Lessons from locality in long-distance dependencies. John Wiley & Sons.
- Berwick, R. C., & Chomsky, N. (2016). Why only us: Language and evolution. MIT Press.
- Borsley, R. (2014). Syntactic theory: A unified approach. Routledge.
- Bošković, Ž. (2021). Merge, move, and contextuality of syntax: The role of labeling, successive-cyclicity, and EPP effects. Ms, University of Connecticut.
- Cai, Z., Duan, X., Haslett, D., Wang, S., & Pickering, M. (2024, August). Do large language models resemble humans in language use?. In Proceedings of the workshop on cognitive modeling and computational linguistics (pp. 37-56).
- Chesi, C. (2024). Is it the end of (generative) linguistics as we know it?. arXiv preprint arXiv:2412.12797.
- Chomsky, N. (2014). The minimalist program. MIT press.
- Chomsky, N. (2011). Current issues in linguistic theory (Vol. 38). Walter de Gruyter.
- Chomsky, N. (1998, 2000) Minimalist Inquiries: The Framework, Preliminary Version in MIT Working Papers in Linguistics No. 15. In: Martin, R., Michaels, D. and Uriagereka, J., Eds., Step by Step: Essays on Minimalist Syntax in Honor of Howard Lasnik, MIT Press, Cambridge, 89-155.
- Chomsky, N. (1959). On certain formal properties of grammars. Information and control, 2(2), 137-167.
- Chomsky, N. (2005) Three Factors in Language Design. Linguistic Inquiry, 36, 1-22.
- https://doi.org/10.1162/0024389052993655
- Chomsky, N., Seely, T. D., Berwick, R. C., Fong, S., Huybregts, M. A. C., Kitahara, H., … Sugimoto, Y. (2023). Merge and the Strong Minimalist Thesis. Cambridge: Cambridge University Press.
- Cinque, G. (2005). Deriving Greenberg's Universal 20 and its exceptions. Linguistic inquiry, 36(3), 315-332.
- Citko, B. (2014). Phase Theory: An Introduction. Cambridge: Cambridge University Press.
- Citko, B. (2014). Variation in phasehood. In Phase Theory: An Introduction (pp. 161–184). chapter, Cambridge: Cambridge University Press.
- Collins, C., & Stabler, E. (2016). A formalization of minimalist syntax. Syntax, 19(1), 43-78.
- Cooper, R. (2013). Quantification and syntactic theory (Vol. 21). Springer Science & Business Media.
- Deming, M., & Helong, T. (2024). The Development of Chomskyan Generative Grammar.
- Drijvers, L., Small, S. L., & Skipper, J. I. (2025). Language is widely distributed throughout the brain. Nature Reviews Neuroscience, 26(3), 189. https://doi.org/10.1038/s41583-024-00903-0
- Epstein, S. D. (1998). A derivational approach to syntactic relations. Oxford University Press.
- Epstein, S. D., Kitahara, H., & Seely, T. D. (2021). A minimalist theory of simplest merge. New York: Routledge.
- Epstein, S. D., & Seely, T. D. (2006). Derivations in minimalism (No. 111). Cambridge University Press.
- Fedorenko, E., Piantadosi, S.T. & Gibson, E.A.F. Language is primarily a tool for communication rather than thought. Nature 630, 575–586 (2024). https://doi.org/10.1038/s41586-024-07522-w
- Fedorenko, E., Ivanova, A. A., & Regev, T. I. (2024). The language network as a natural kind within the broader landscape of the human brain. Nature reviews. Neuroscience, 25(5), 289–312. https://doi.org/10.1038/s41583-024-00802-4
- Fox, D., & Pesetsky, D. (2005). Cyclic linearization of syntactic structure. Theoretical linguistics, 31(1-2), 1-45.
- Frampton, J., & Gutmann, S. (1999). Cyclic computation, a computationally efficient minimalist syntax. Syntax, 2(1), 1-27.
- Fukui, N., & Takano, Y. (1998). Symmetry in syntax: Merge and demerge. Journal of East Asian Linguistics, 7(1), 27-86.
- Gallego, Á. J. (2010). Phase theory.
- Ginsburg, J. (2024). Constraining Free Merge. Biolinguistics, 18, 1-60.
- Henry, A. (2004). Variation and syntactic theory. The handbook of language variation and change, 267-282.
- Hornstein, N., Nunes, J., & Grohmann, K. K. (2005). Understanding minimalism. Cambridge University Press.
- Kalin, L., & Weisser, P. (2021). Minimalism and morphology. Ms., Princeton University and Leipzig University.
- Kayne, R. (1994). The antisymmetry of syntax. Linguistic Inquiry Monograph25/MIT Press.
- Kirby, S., Griffiths, T., & Smith, K. (2014). Iterated learning and the evolution of language. Current opinion in neurobiology, 28, 108-114.
- Lasnik, H. (2000). Syntactic structures revisited: Contemporary lectures on classic transformational theory (Vol. 33). MIT Press.
- Lasnik, H. (2002). The minimalist program in syntax. Trends in cognitive sciences, 6(10), 432-437.
- Legate, J. A. (2021). Noncanonical passives: A typology of voices in an impoverished Universal Grammar. Annual Review of Linguistics, 7(1), 157-176.
- Legate, J. A. (2014). Voice and v: Lessons from Acehnese. MIT Press.
https://doi.org/10.7551/mitpress/9780262028141.001.0001 - LF, P. Principles & Parameters Theory and Minimalism.
- Linz, P., & Rodger, S. H. (2022). An introduction to formal languages and automata. Jones & Bartlett Learning.
- Mahowald, K., Ivanova, A. A., Blank, I. A., Kanwisher, N., Tenenbaum, J. B., & Fedorenko, E. (2024). Dissociating language and thought in large language models. Trends in cognitive sciences, 28(6), 517-540.
- Mair, C., & Leech, G. N. (2020). Current changes in English syntax. The handbook of English linguistics, 249-276.
- Moro, A. (2008). The boundaries of Babel: The brain and the enigma of impossible languages. MIT press.
- Müller, S. (2023). Grammatical theory: From transformational grammar to constraint-based approaches. Language Science Press.
- Newmeyer, F. (2023). Linguistic theory in America. Brill.
- Oliveira, R. P. D., & Quarezemin, S. (2020). Contemporary Issues in Syntax and Semantics. DELTA: Documentação de Estudos em Lingüística Teórica e Aplicada, 36(1), 2020360113.
- Piantadosi, S. T. (2023). Modern language models refute Chomsky’s approach to language. From fieldwork to linguistic theory: A tribute to Dan Everett, 15, 353-414.
- Pierce, A. E. (1992). Language acquisition and syntactic theory. In Language Acquisition and Syntactic Theory: A Comparative Analysis of French and English Child Grammars (pp. 1-17). Dordrecht: Springer Netherlands.
- Pleyer, M., & Hartmann, S. (2019). Constructing a consensus on language evolution? Convergences and differences between biolinguistic and usage-based approaches. Frontiers in Psychology, 10, Article 2537. https://doi.org/10.3389/fpsyg.2019.02537
- Preminger, O. (2014). Agreement and its failures (Vol. 68). MIT Press.
- Pullum, G. K. (1977). Word order universals and grammatical relations. In Grammatical relations (pp. 249-277). Brill.
- Radford, A. (2004). Minimalist syntax: Exploring the structure of English. Cambridge university press.
- Rizzi, L., & Cinque, G. (2016). Functional categories and syntactic theory. Annual Review of Linguistics, 2, 139-163.
- Rizzi, L., & Cinque, G. (2016). Functional categories and syntactic theory. Annual Review of Linguistics, 2, 139-163.
- Rizzi, L., & Si, F. (2021). Introduction: On the comparative basis of cartographic studies. In Current Issues in Syntactic Cartography (pp. 1-12). John Benjamins Publishing Company.
- Sag, I. A., Wasow, T., Bender, E. M., & Sag, I. A. (2003). Syntactic theory: A formal introduction. Stanford: Center for the Study of Language and Information.
- Schätzle, C., Butt, M., & Kotcheva, K. (2015). The diachrony of dative subjects and the middle in Icelandic: A corpus study.
- MIT Lingusitics papers: http://mitwpl.mit.edu/catalog/
