Minimal Computation and The Architecture of Language: Noam Chomsky
Minimal Computation and The Architecture of Language: Noam Chomsky
Noam Chomsky*
Minimal Computation and the Architecture
of Language
Abstract: The article argues that humans are endowed with an inborn language
faculty. The ‘Basic Property’ of language is defined as a finitely-specified
procedure represented in the brain, which generates a discrete infinity of
hierarchically structured expressions. These unordered structures are linked to
two interfaces: (i) the sensorimotor interface and (ii) the conceptual-intentional
interface. The sensorimotor interface externalizes and linearizes internal
structures, usually in the sound modality. Externalization and linearization
account for the structural diversity of the world’s languages. Human language
did not evolve from simpler communication systems. The available evidence
suggests that language is primarily an instrument of thought, not of
communication.
*Corresponding author, Noam Chomsky: Massachusetts Institute of Technology,
E-mail: [email protected]
1 Historical background
From the early days of the modern scientific revolution, there has been intense
interest in human language, recognized to be a core feature of human nature
and the primary capacity distinguishing modern humans from other creatures.
14 Noam Chomsky
[T]he acquisition of the uniquely modern [human] sensibility was instead an abrupt and
recent event. […] And the expression of this new sensibility was almost certainly crucially
abetted by the invention of what is perhaps the single most remarkable thing about our
modern selves: language.
Centuries earlier, Galileo and the seventeenth century Port Royal logicians
and grammarians were awed by the “marvelous invention” of a means to
construct “from 25 or 30 sounds that infinity of expressions, which bear no
resemblance to what takes place in our minds, yet enable us to reveal [to others]
everything that we think, and all the various movements of our soul” (Arnauld
& Lancelot, 1975: 65). Descartes took this capacity to be a primary difference
between humans and any beast-machine, providing a basic argument for his
mind-body dualism. The great humanist Wilhelm von Humboldt (1836)
characterized language as “a generative activity [eine Erzeugung]” (p. 39) rather
than “a lifeless product” [ein todtes Erzeugtes] (p. 39), Energeia rather than
Ergon (p. 41), and pondered the fact that somehow this activity “makes infinite
use of finite means”[…] sie mussvon endlichen Mitteln einen unendlichen
Gebrauch machen] (p. 106) (for sources see Chomsky, 1966/2009). For the last
great representative of this tradition, Otto Jespersen (1924), the central question
of the study of language is how its structures “come into existence in the mind
of a speaker” (p. 19) on the basis of finite experience, yielding a “notion of […]
structure” (p. 19) that is “definite enough to guide him in framing sentences of
his own”, (p. 19) crucially “free expressions” (p. 20) that are typically new to
speaker and hearer. And more deeply, to go beyond to unearth “the great
principles underlying the grammars of all languages” (p. 344) and by so doing
to gain “a deeper insight into the innermost nature of human language and of
human thought” (p. 347) – ideas that sound much less strange today than they
did during the structuralist/behavioral science era that came to dominate much
of the field through the first half of the twentieth century, marginalizing the
leading ideas and concerns of the tradition (see Jespersen, 1924).
Throughout this rich tradition of reflection and inquiry there were efforts to
comprehend how humans can freely and creatively employ “an infinity of
expressions” to express their thoughts in ways that are appropriate to
circumstances though not determined by them, a crucial distinction. However,
tools were not available to make much progress in carrying these ideas forward.
That difficulty was partially overcome by mid-twentieth century, thanks to the
work of Gödel, Turing, and other great mathematicians that laid the basis for
Minimal Computation and the Architecture of Language 15
The study of the finite means that are used in linguistic behavior – the puppet
and the strings – has been pursued very successfully since the mid-twentieth
century in what has come to be called the “generative enterprise” and
“biolinguistic framework”, drawing from and contributing to the “cognitive
revolution” that has been underway during this period. The kinds of questions
that students are investigating today could not even have been formulated not
16 Noam Chomsky
many years ago, and there has been a vast explosion in the languages of the
widest typological variety that have come under investigation, at a level of
depth never before contemplated in the long and rich history of investigation of
language since classical Greece and ancient India. There have been many
discoveries along the way, regularly raising new problems and opening new
directions of inquiry. In these respects the enterprise has had considerable
success.
Departing from the assumptions of the structuralist/behaviorist era and
returning to the spirit of the tradition in new forms, the generative/biolinguistic
enterprise takes a language to be an internal system, a “module” of the system
of human cognitive capacities. In technical terms, a language is taken to be an
“I-language” – where “I” stands for internal, individual, and intensional
(meaning that we are concerned with the actual nature of the biological object
itself rather than with some set of objects that it generates, such as a corpus of
expressions or set of behaviors). Each I-language satisfies the Basic Property of
human language, formulated above. Jespersen’s “great principles underlying
the grammars of all languages” are the topic of Universal Grammar (UG),
adapting a traditional term to a new framework, interpreted now as the theory
of the genetic endowment for the faculty of language, the innate factors that
determine the class of possible I-languages.
There is by now substantial evidence that UG is a species property, uniform
among humans apart from severe pathology, and with no close analogue, let
alone anything truly homologous, in the rest of the animal world. It seems to
have emerged quite recently in evolutionary time, as Tattersall concluded,
probably within the last 100,000 years. And we can be fairly confident that it
has not evolved at least since our ancestors began to leave Africa some 50-60
thousand years ago. If so, then the emergence of the language faculty – of UG –
was quite sudden in evolutionary time, which leads us to suspect that the Basic
Property, and whatever else constitutes UG, should be very simple. Furthermore,
since Eric Lenneberg’s pioneering work in the 1950s, evidence has been
accumulating that the human language faculty is dissociated from other
cognitive capacities – though of course the use of language in perception
(parsing) and production integrates the internal I-language with other
capacities (see Lenneberg, 1967; Curtiss, 2012).That too suggests that whatever
emerged quite suddenly (in evolutionary time) should be quite simple.
As the structuralist and behavioral science approaches took shape through
the first half of the twentieth century, it came to be generally assumed that the
field faced no fundamental problems. Methods of analysis were available,
notably Zellig Harris’s Methods in Structural Linguistics, which provided the
means to reduce a corpus of materials to an organized form, the primary task of
Minimal Computation and the Architecture of Language 17
the discipline. The problems of phonology, the major focus of inquiry, seemed
to be largely understood. As a student in the late 1940s, I remember well the
feeling that “this is really interesting work, but what happens to the field when
we have structural grammars for all languages?” These beliefs made sense
within the prevailing framework, as did the widely-held “Boasian” conception
articulated by theoretical linguist Martin Joos (1957: 96) that languages can
“differ from each other without limit and in unpredictable ways” (p. 96) so that
the study of each language must be approached “without any preexistent
scheme of what a language must be” (Joos, 1957: v).
These beliefs collapsed as soon as the first efforts to construct generative
grammars were undertaken by mid-twentieth century. It quickly became clear
that very little was known about human language, even the languages that had
been well studied. It also became clear that many of the fundamental properties
of language that were unearthed must derive in substantial part from the innate
language faculty, since they are acquired with little or no evidence. Hence there
must be sharp and determinate limits to what a language can be. Furthermore,
many of the properties that were revealed with the first efforts to construct rules
satisfying the Basic Principle posed serious puzzles, some still alive today,
along with many new ones that continue to be unearthed.
In this framework, the study of a specific language need not rely just on the
behavior and products of speakers of this language. It can also draw from
conclusions about other languages, from neuroscience and psychology, from
genetics, in fact from any source of evidence, much like science generally,
liberating the inquiry from the narrow constraints imposed by strict
structuralist/behavioral science approaches.
In the early days of the generative enterprise, it seemed necessary to
attribute great complexity to UG in order to capture the empirical phenomena of
languages. It was always understood, however, that this cannot be correct. UG
must meet the condition of evolvability, and the more complex its assumed
character, the greater the burden on some future account of how it might have
evolved – a very heavy burden in the light of the few available facts about
evolution of the faculty of language, as just indicated.
From the earliest days, there were efforts to reduce the assumed complexity
of UG while maintaining, and often extending, its empirical coverage. And over
the years there have been significant steps in this direction. By the early 1990s it
seemed to a number of researchers that it might be possible to approach the
problems in a new way: by constructing an “ideal solution” and asking how
closely it can be approximated by careful analysis of apparently recalcitrant
data, an approach that has been called “the minimalist program”. The notion
“ideal solution” is not precisely determined a priori, but we have a grasp of
18 Noam Chomsky
raising rule of a type familiar in many languages, yielding which book will John
read.
It is important to note that throughout, the operations described satisfy MC.
That includes the deletion operation in the mapping to SM, which sharply
reduces the computational and articulatory load in externalizing the Merge-
generated SO. To put it loosely, what reaches the mind has the right semantic
form, but what reaches the ear has gaps that have to be filled by the hearer.
These “filler-gap” problems pose significant complications for
parsing/perception. In such cases, I-language is “well-designed” for thought
but poses difficulties for language use, an important observation that in fact
generalizes quite widely and might turn out to be exceptionless, when the
question arises.
Note that what reaches the mind lacks order, while what reaches the ear is
ordered. Linear order, then, should not enter into the syntactic-semantic
computation. Rather, it is imposed by externalization, presumably as a reflex of
properties of the SM system, which requires linearization: we cannot speak in
parallel or articulate structures. For many simple cases, this seems accurate:
thus there is no difference in the interpretation of verb-object constructions in
head-initial or head-final constructions.
The same is true in more complex cases, including “exotic” structures that
are particularly interesting because they rarely occur but are understood in a
determinate way, for example, parasitic gap constructions. The “real gap” RG
(which cannot be filled) may either precede or follow the “parasitic gap” PG
(which can be filled), but cannot be in a dominant (c-command) structural
relation to the PG, as illustrated in the following:
grammars were undertaken. On the surface, it seems to conflict with the quite
natural and generally operative principles of MC.
To illustrate, consider the following sentences:
The structures of (6) and (7) are, roughly, as indicated by bracketing in (6’)
and (7’) respectively:
We might proceed to entertain another bold but not implausible thesis: that
generation of CI – narrow syntax and construal/interpretive rules – is uniform
among languages, or nearly so. In fact, realistic alternatives are not easy to
imagine, in the light of the fact that the systems are acquired on the basis of
little or no evidence, as even the few simple examples given earlier illustrate.
The conclusion also comports well with the very few known facts about origin of
language. These appear to place the emergence of language within a time frame
that is very brief in evolutionary time, hardly more than an instant, and with no
evolutionary change since. Hence we would expect what evolved – UG – to be
quite simple.
22 Noam Chomsky
3 Concluding remarks
I remarked earlier that in my student days in mid-twentieth century, it seemed
as though the major problems of the study of language had been pretty much
solved and that the enterprise, though challenging, was approaching a terminal
point that could be fairly clearly perceived. The picture today could not be more
different.
References
Arnauld, A., & Lancelot, C. (1975). General and rational grammar: The Port Royal Grammar. (J.
Rieux & B. Rollin, Trans.). The Hague: Mouton.
Berwick, R., Pietroski, P., Yankama, B., & Chomsky, N. (2011). Poverty of the stimulus revisited,
Cognitive Science, 35, 1–36.
24 Noam Chomsky
Bizzi, E., & Ajemian, R. (2015). A hard scientific quest: Understanding voluntary movements,
Daedalus, 144(1), 83–95.
Chomsky, N. (1966). Cartesian linguistics: A chapter in the history of rationalist thought. New
York: Harper & Row. Third edition revised and edited with an introduction by J. McGilvray
(2009); ebooks, Cambridge University.
Chomsky, N. (1995). The minimalist program. Cambridge, MA: MIT Press.
Curtiss, S. (2012). Revisiting modularity: Using language as a window to the mind. In R.
Berwick & M. Piattelli-Palmarini (Eds.), Rich languages from poor inputs (pp. 68–90).
Oxford, UK: Oxford University Press.
Humboldt, W. von (1836). Ueber die Verschiedenheit des menschlichen Sprachbaues und ihren
Einfluss auf die geistige Entwicklung des Menschengeschlechts. Berlin: F. Dümmler.
Jespersen, O. (1924). The philosophy of grammar. Chicago: The University of Chicago Press.
Joos, M. (Ed.) (1957). Readings in linguistics. Washington, DC: American Council of Learned
Societies.
Lenneberg, E. (1967). Biological foundations of language. New York: John Wiley and Sons.
Musso, M., Moro, A., Glauche, V., Rijntjes, M., Reichenbach, J., Büchel C. & Weiller, C. (2003).
Broca’s area and the language instinct. Nature Neuroscience,6, 774–781.
Smith, N. (1999). Chomsky: Ideas and ideals. Cambridge, UK: Cambridge University Press.
Smith, N., & Tsimpli, I.-M. (1995). The mind of a savant: Language learning and
modularity.Oxford, UK: Blackwell.
Tattersall, I. (2012). Masters of the planet: The search for our human origins. London: Palgrave
Macmillan.
Bionote
Noam Chomsky
Noam Chomsky (b. 1928) is Institute Professor & Professor of Linguistics (Emeritus) at the
Massachusetts Institute of Technology. His research areas include linguistic theory, syntax,
semantics, and philosophy of language.