Astral Codex Ten - Book Review, How Languages Began (2024)
Astral Codex Ten - Book Review, How Languages Began (2024)
[This is one of the finalists in the 2024 book review contest, written by an ACX
reader who will remain anonymous until after voting is done. I’ll be posting about
one of these a week for several months. When you’ve read them all, I’ll ask you to
vote for a favorite, so remember which ones you liked]
I. THE GOD
You may have heard of a field known as "linguistics". Linguistics is supposedly the
"scientific study of language", but this is completely wrong. To borrow a phrase
from elsewhere, linguists are those who believe Noam Chomsky is the rightful
caliph. Linguistics is what linguists study.
I'm only half-joking, because Chomsky’s impact on the study of language is hard
to overstate. Consider the number of times his books and papers have been cited,
a crude measure of influence that we can use to get a sense of this. At the current
time, his Google Scholar page says he's been cited over 500,000 times. That’s a
lot.
Yes, fields vary in ways that make these comparisons not necessarily fair: fields
have different numbers of people, citation practices vary, and so on. There is also
probably a considerable recency bias; for example, most biologists don’t cite
Darwin every time they write a paper whose content relates to evolution. But
500,000 is still a mind-bogglingly huge number.
Not many academics do better than Chomsky citation-wise. But there are a few,
and you can probably guess why:
…well, okay, maybe I don’t entirely get Foucault’s number. Every humanities
person must have an altar of him by their bedside or something.
:
Chomsky has been called “arguably the most important intellectual alive today” in
a New York Times review of one of his books, and was voted the world’s top
public intellectual in a 2005 poll. He’s the kind of guy that gets long and gushing
introductions before his talks (this one is nearly twenty minutes long). All of this is
just to say: he’s kind of a big deal.
:
This is what he looks like. According to Wikipedia, the context for this picture is:
Is language for communicating? No, it’s mainly for thinking: (What Kind of
Creatures Are We? Ch. 1, pg. 15-16)
It is, indeed, virtual dogma that the function of language is communication. ...
there is by now quite significant evidence that it is simply false. Doubtless
language is sometimes used for communication, as is style of dress, facial
expression and stance, and much else. But fundamental properties of
language design indicate that a rich tradition is correct in regarding language
as essentially an instrument of thought, even if we do not go as far as
Humboldt in identifying the two.
Should linguists care about the interaction between culture and language? No,
that’s essentially stamp-collecting: (Language and Responsibility, Ch. 2, pg. 56-
57)
Again, a discipline is defined in terms of its object and its results. Sociology is
the study of society. As to its results, it seems that there are few things one
can say about that, at least at a fairly general level. One finds observations,
intuitions, impressions, some valid generalizations perhaps. All very valuable,
no doubt, but not at the level of explanatory principles. … Sociolinguistics is, I
suppose, a discipline that seeks to apply principles of sociology to the study of
language; but I suspect that it can draw little from sociology, and I wonder
whether it is likely to contribute much to it. … You can also collect butterflies
and make many observations. If you like butterflies, that’s fine; but such work
must not be confounded with research, which is concerned to discover
explanatory principles of some depth and fails if it has not done so.
Did the human capacity for language evolve gradually? No, it suddenly appeared
around 50,000 years ago after a freak gene mutation: (Language and Mind, third
edition, pg, 183-184)
The simplest account of the “Great Leap Forward” in the evolution of humans
would be that the brain was rewired, perhaps by some slight mutation, to
provide the operation Merge … There are speculations about the evolution of
language that postulate a far more complex process … A more parsimonious
speculation is that they did not, and that the Great Leap was effectively
instantaneous, in a single individual, who was instantly endowed with
intellectual capacities far superior to those of others, transmitted to offspring
and coming to predominate. At best a reasonable guess, as are all
speculations about such matters, but about the simplest one imaginable, and
not inconsistent with anything known or plausibly surmised. It is hard to see
what account of human evolution would not assume at least this much, in one
or another form.
I think all of these positions are kind of insane for reasons that we will discuss
later. (Side note: Chomsky’s proposal is essentially the hard takeoff theory of
human intelligence.)
Most consequential of all, perhaps, are the ways Chomsky has influenced (i) what
linguists mainly study, and (ii) how they go about studying it.
The words “expected” and “persuaded” appear in the same location in each
sentence, but imply different ‘latent’ grammatical structures, or ‘deep structures’.
One way to show this is to observe that a particular way of rearranging the words
produces a sentence with the same meaning in the first case (1a = 1b), and a
different meaning in the second (2a != 2b):
In particular, the target of persuasion is “John” in the case of (2a), and “the
specialist” in the case of (2b). A full Chomskyan treatment of sentences like this
would involve hierarchical tree diagrams, which permit a precise description of
deep structure.
You may have encountered the famous sentence: “Colorless green ideas sleep
furiously.” It first appeared in Chomsky’s 1957 book Syntactic Structures, and the
point is that even nonsense sentences can be grammatically well-formed, and
that speakers can quickly assess the grammatical correctness of even nonsense
sentences that they’ve never seen before. To Chomsky, this is one of the most
important facts to be explained about language.
The problem for the linguist, as well as for the child learning the language, is to
determine from the data of performance the underlying system of rules that
has been mastered by the speaker-hearer and that he puts to use in actual
performance. Hence, in the technical sense, linguistic theory is mentalistic,
since it is concerned with discovering a mental reality underlying actual
behavior. Observed use of language or hypothesized dispositions to respond,
habits, and so on, may provide evidence as to the nature of this mental reality,
but surely cannot constitute the actual subject matter of linguistics, if this is to
be a serious discipline.
Moreover, he claims that grammar captures most of what we should mean when
we talk about speakers’ linguistic competence: (Aspects of the Theory of Syntax,
Ch. 1, pg. 24)
Chomsky claims that the history of generative linguistics shows a shift from an
E-language to an I-language approach; ‘the shift of focus from the dubious
concept of E-language to the significant notion of I-language was a crucial
step in early generative grammar’ (Chomsky, 1991b, pg. 10). … Indeed
Chomsky is extremely dismissive of E-language approaches: ‘E-language, if it
exists at all, is derivative, remote from mechanisms and of no particular
empirical significance, perhaps none at all’ (Chomsky, 1991b, pg. 10). 1
The opposition between these two approaches in linguistics has been long and
acrimonious, neither side conceding the other’s reality. … The E-linguist
despises the I-linguist for not looking at the ‘real’ facts; the I-linguist derides
the E-linguist for looking at trivia. The I-language versus E-language
distinction is as much a difference of research methods and of admissible
evidence as it is of long-term goals.
So much for what linguists ought to study. How should they study it?
The previous quote gives us a clue. Especially in the era before Chomsky (BC),
linguists were more interested in description. Linguists were, at least in one view,
people who could be dropped anywhere in the world, and emerge with a tentative
:
grammar of the local language six months later. (A notion like this is mentioned
early in this video.) Linguists catalog the myriad of strange details about human
languages, like the fact that some languages don’t appear to have words for
relative directions, or “thank you”, or “yes” and “no”.
After Chomsky's domination of the field (AD), there were a lot more theorists.
While you could study language by going out into the field and collecting data,
this was viewed as not the only, and maybe not even the most important, way to
work. Diagrams of sentences proliferated. Chomsky, arguably the most influential
linguist of the past hundred years, has never done fieldwork.
In summary, to Chomsky and many of the linguists working in his tradition, the
scientifically interesting component of language is grammar competence, and real
linguistic data only indirectly reflects it.
All of this matters because the dominance of Chomskyan linguistics has had
downstream effects in adjacent fields like artificial intelligence (AI), evolutionary
biology, and neuroscience. Chomsky has long been an opponent of the statistical
learning tradition of language modeling, essentially claiming that it does not
provide insight about what humans know about languages, and that engineering
success probably can’t be achieved without explicitly incorporating important
mathematical facts about the underlying structure of language. Chomsky’s ideas
have motivated researchers to look for a “language gene” and “language areas”
of the brain. Arguably, no one has yet found either—but more on that later.
Where are we going with this? All of this is context for understanding the ideas of
:
a certain bomb-throwing terrorist blight on the face of linguistics: Daniel Everett.
How Language Began is a book he wrote about, well, what language is and how it
began. Everett is the anti-Chomsky.
Daniel Everett is the boy in this story. The woman he married, Keren Graham, is
the daughter of Christian missionaries and had formative experiences living in the
Amazon jungle among the Sateré-Mawé people. At seventeen, Everett became a
born-again Christian; at eighteen, he and Keren married; and over the next few
years, they started a family and prepared to become full-fledged missionaries like
Keren’s parents.
First, Everett studied “Bible and Foreign Missions” at the Moody Bible Institute in
Chicago. After finishing his degree in 1975, the natural next step was to train more
specifically to follow in the footsteps of Keren’s parents. In 1976, he and his wife
enrolled in the Summer Institute of Linguistics (SIL) to learn translation techniques
and more viscerally prepare for life in the jungle:
They were sent to Chiapas, Mexico, where Keren stayed in a hut in the jungle
with the couple’s children—by this time, there were three—while Everett
underwent grueling field training. He endured fifty-mile hikes and survived for
several days deep in the jungle with only matches, water, a rope, a machete,
and a flashlight.
Everett apparently had a gift for language-learning. This led SIL to invite Everett
and his wife to work with the Pirahã people (pronounced pee-da-HAN), whose
unusual language had thwarted all previous attempts to learn it. In 1977, Everett’s
family moved to Brazil, and in December they met the Pirahã for the first time. As
:
an SIL-affiliated missionary, Everett’s explicit goals were to (i) translate the Bible
into Pirahã, and (ii) convert as many Pirahã as possible to Christianity.
But Everett’s first encounter with the Pirahã was cut short for political reasons:
(Don’t Sleep There Are Snakes, Ch. 1, pg. 13-14)
I went to the Pirahãs when I was twenty-six years old. Now I am old enough to
receive senior discounts. I gave them my youth. I have contracted malaria
many times. I remember several occasions on which the Pirahãs or others
threatened my life. I have carried more heavy boxes, bags, and barrels on my
back through the jungle than I care to remember. But my grandchildren all
know the Pirahãs. My children are who they are in part because of the Pirahãs.
And I can look at some of those old men (old like me) who once threatened to
:
kill me and recognize some of the dearest friends I have ever had—men who
would now risk their lives for me.
Everett did eventually learn their language, and it’s worth taking a step back to
appreciate just how hard that task was. No Pirahã spoke Portuguese, apart from
some isolated phrases they used for bartering. They didn’t speak any other
language at all—just Pirahã. How do you learn another group’s language when you
have no languages in common? The technical term is monolingual fieldwork. But
this is just a fancy label for some combination of pointing at things, listening,
crude imitation, and obsessively transcribing whatever you hear. For years.
It doesn’t help that the Pirahã language seems genuinely hard to learn in a few
different senses. First, it is probably conventionally difficult for Westerners to
learn since it is a tonal language (two tones: high and low) with a small number of
phonemes (building block sounds) and a few unusual sounds 3. Second, there is
no written language. Third, the language has a variety of ‘channels of discourse’,
or ways of talking specialized for one or another cultural context. One of these is
:
‘whistle speech’; Pirahãs can communicate purely in whistles. This feature
appears to be extremely useful during hunting trips: (Don’t Sleep, There Are
Snakes, Ch. 11, pg. 187-188)
My first intense contact with whistle speech came one day when the Pirahãs
had given me permission to go hunting with them. After we’d been walking for
about an hour, they decided that they weren’t seeing any game because I, with
my clunking canteens and machete and congenital clumsiness, was making
too much noise. “You stay here and we will be back for you later.” Xaikáibaí
said gently but firmly. …
Fourth, important aspects of the language reflect core tenets of Pirahã culture in
ways that one might not a priori expect. Everett writes extensively about the
‘immediacy of experience principle’ of Pirahã culture, which he summarizes as the
idea that: (Don’t Sleep, There Are Snakes, Ch. 7, pg. 132)
One way the language reflects this is that the speaker must specify how they
know something by affixing an appropriate suffix to verbs: (Don’t Sleep, There
Are Snakes, Ch. 12, pg. 196)
Perhaps the most interesting suffixes, however (though these are not unique
to Pirahã), are what linguists call evidentials, elements that represent the
:
speaker’s evaluation of his or her knowledge of what he or she is saying. There
are three of these in Pirahã: hearsay, observation, and deduction.
To see what these do, let’s use an English example. If I ask you, “Did Joe go
fishing?” you could answer, “Yes, at least I heard that he did,” or “Yes, I know
because I saw him leave,” or “Yes, at least I suppose he did because his boat is
gone.” The difference between English and Pirahã is that what English does
with a sentence, Pirahã does with a verbal suffix.
Everett also convincingly links this cultural principle to the lack of Pirahã number
words and creation myths. On the latter topic, Everett recalls the following
exchange: (Don’t Sleep, There Are Snakes, Ch. 7, pg. 134)
I sat with Kóhoi once and he asked me, after hearing about my god, “What else
does your god do?” And I answered, “Well, he made the stars, and he made
the earth.” Then I asked, “What do the Pirahãs say?” He answered, “Well, the
Pirahãs say that these things were not made.”
And all of this is to say nothing of the manifold perils of the jungle: malaria,
typhoid fever, dysentery, dangerous snakes, insects, morally gray river traders,
and periodic downpours. If Indiana Jones braved these conditions for years, we
would consider his stories rousing adventures. Everett did this while also learning
one of the most unusual languages in the world.
:
People on the bank of the Maici river. (source)
By the way, he did eventually sort of achieve his goal of translating the Bible.
Armed with a solid knowledge of Pirahã, he was able to translate the New
Testament’s Gospel of Mark. Since the Pirahã have no written language, he
provided them with a recorded version, but did not get the reaction he expected:
(Don’t Sleep, There Are Snakes, Ch. 17, pg. 267-268)
When we returned to the village, I recorded Mark’s gospel in my own voice for
the Pirahãs to listen to. I then brought in a wind-up tape recorder to play the
recording, and I taught the Pirahãs how to use it, which, surprisingly enough,
some of the children did. Keren and I left the village and returned a few weeks
later. The people were still listening to the gospel, with children cranking the
recorder. I was initially quite excited about this, until it became clear that the
only part of the book that they paid attention to was the beheading of John the
Baptist. “Wow, they cut off his head. Play that again!”
:
One reaction to hearing the gospel caught Everett even more off-guard: (Don’t
Sleep, There Are Snakes, Ch. 17, pg. 269)
"Because last night he came to our village and tried to have sex with our
women. He chased them around the village, trying to stick his large penis into
them."
Kaaxaóoi proceeded to show me with his two hands held far apart how long
Jesus's penis was—a good three feet.
But the Pirahã had an even more serious objection to Jesus: (Don’t Sleep, There
Are Snakes, Ch. 17, pg. 265-266)
"The Pirahã men then asked, "Hey Dan, what does Jesus look like? Is he dark
like us or light like you?" I said, "Well, I have never actually seen him. He lived a
long time ago. But I do have his words." "Well, Dan, how do you have his words
if you have never heard him or seen him?"
They then made it clear that if I had not actually seen this guy (and not in any
metaphorical sense, but literally), they weren't interested in any stories I had to
tell about him. Period. This is because, as I now knew, the Pirahãs believe only
what they see. Sometimes they also believe in things that someone else has
told them, so long as that person has personally witnessed what he or she is
reporting.
In the end, Everett never converted a single Pirahã. But he did even worse than
:
converting zero people—he lost his own faith after coming to believe that the
Pirahã had a good point. After keeping this to himself for many years, he revealed
his loss of faith to his family, which led to a divorce and his children breaking
contact with him for a number of years afterward.
But Everett losing his faith in the God of Abraham was only the beginning. Most
importantly for us, he also lost his faith in the God of Linguistics—Noam
Chomsky.
But the article ignited a firestorm of controversy that follows Everett to this day.
Praise for Everett and his work on recursion in Pirahã:
You, too, can enjoy the spotlight of mass media and closet exoticists! Just find
a remote tribe and exploit them for your own fame by making claims nobody
will bother to check!
Andrew Nevins, UCL professor and linguist (Harvard professor at quote time)
:
I think he knows he’s wrong, that’s what I really think. I think it’s a move that
many, many intellectuals make to get a little bit of attention.
bedobi, Redditor
Apparently he struck a nerve. And there is much more vitriol like this; see Pullum
for the best (short) account of the beef I’ve found, along with sources for each
quote except the last. On the whole affair, he writes:
I’m not going to rehash all of the details, but the conduct of many in the pro-
Chomsky faction is pretty shocking. Highly recommended reading. Substantial
portions of the books The Kingdom of Speech and Decoding Chomsky are also
dedicated to covering the beef and related issues, although I haven’t read them.
What’s going on? Assuming Everett is indeed acting in good faith, why did he get
this reaction? As I said in the beginning, linguists are those who believe Noam
Chomsky is the rightful caliph. Central to Chomsky’s conception of language is
the idea that grammar reigns supreme, and that human brains have some
specialized structure for learning and processing grammar. In the writing of
Chomsky and others, this hypothetical component of our biological endowment is
sometimes called the narrow faculty of language (FLN); this is to distinguish it
from other (e.g., sensorimotor) capabilities relevant for practical language use.
A paper by Hauser, Chomsky, and Fitch titled “The Faculty of Language: What Is
It, Who Has It, and How Did It Evolve?” was published in the prestigious journal
:
Science in 2002, just a few years earlier. The abstract contains the sentence:
We hypothesize that FLN only includes recursion and is the only uniquely
human component of the faculty of language.
Some additional context is that Chomsky had spent the past few decades
simplifying his theory of language. A good account of this is provided in the first
chapter of Chomsky’s Universal Grammar: An Introduction. By 2002, arguably not
much was left: the core claims were that (i) grammar is supreme, (ii) all grammar
is recursive and hierarchical. More elaborate aspects of previous versions of
Chomsky’s theory, like the idea that each language might be identified with
different parameter settings of some ‘global’ model constrained by the human
brain (the core idea of the so-called ‘principles and parameters’ formulation of
universal grammar), were by now viewed as helpful and interesting but not
necessarily fundamental.
Hence, it stands to reason that evidence suggesting not all grammar is recursive
could be perceived as a significant threat to the Chomskyan research program. If
not all languages had recursion, then what would be left of Chomsky’s once-
formidable theoretical apparatus?
Everett’s paper inspired a lively debate, with many arguing that he is lying, or
misunderstands his own data, or misunderstands Chomsky, or some combination
of all of those things. The most famous anti-Everett response is “Pirahã
Exceptionality: A Reassessment” by Nevins, Pesetsky, and Rodrigues (NPR),
which was published in the prestigious journal Language in 2009. This paper got
a response from Everett, which led to an NPR response-to-the-response.
To understand how contentious even the published form of this debate became, I
reproduce in full the final two paragraphs of NPR’s response-response:
We began this commentary with a brief remark about the publicity that has
been generated on behalf of Everett's claims about Pirahã. Although reporters
and other nonlinguists may be aware of some ‘big ideas’ prominent in the field,
the outside world is largely unaware of one of the most fundamental
:
achievements of modern linguistics: the three-fold discovery that (i) there is
such a thing as a FACT about language; (ii) the facts of language pose
PUZZLES, which can be stated clearly and precisely; and (iii) we can propose
and evaluate SOLUTIONS to these puzzles, using the same intellectual skills
that we bring to bear in any other domain of inquiry. This three-fold discovery
is the common heritage of all subdisciplines of linguistics and all schools of
thought, the thread that unites the work of all serious modern linguists of the
last few centuries, and a common denominator for the field.
In our opinion, to the extent that CA and related work constitute a ‘volley fired
straight at the heart’ of anything, its actual target is no particular school or
subdiscipline of linguistics, but rather ANY kind of linguistics that shares the
common denominator of fact, puzzle, and solution. That is why we have
focused so consistently on basic, common-denominator questions: whether
CA’s and E09’s conclusions follow from their premises, whether contradictory
published data has been properly taken into account, and whether relevant
previous research has been represented and evaluated consistently and
accurately. To the extent that outside eyes may be focused on the Pirahã
discussion for a while longer, we would like to hope that NP&R (and the
present response) have helped reinforce the message that linguistics is a field
in which robustness of evidence and soundness of argumentation matter.
Two observations here. First, another statement about “serious” linguistics; why
does that keep popping up? Second, wow. That’s the closest you can come to
cursing someone out in a prestigious journal.
Polemics aside, what’s the technical content of each side’s argument? Is Pirahã
recursive or not? Much of the debate appears to hinge on two things:
what one means by the statement “All natural human languages have
recursion.”
Everett generally takes recursion to refer to the following property of many natural
:
languages: one can construct sentences or phrases from other sentences and
phrases. For example:
“The cat died.” -> “Alice said that [the cat died].” -> “Bob said that [Alice said that
[the cat died.]]”
In the above example, we can in principle generate infinitely many new sentences
by writing “Z said X,” where X is the previous sentence and Z is some name. For
clarity’s sake, one should probably distinguish between different ways to generate
new sentences or phrases from old ones; Pullum mentions a few in the context of
assessing Everett’s Pirahã recursion claims:
Everett reports that there are no signs of no multiple coordination (It takes
[skill, nerve, initiative, and courage]), complex determiners ([[[my] son’s]
wife’s] family), stacked modifiers (a [nice, [cosy, [inexpensive [little
cottage]]]]), or—most significant of all—reiterable clause embedding (I thought
[ you already knew [that she was here ] ]). These are the primary
constructions that in English permit sentences of any arbitrary finite length to
be constructed, yielding the familiar argument that the set of all definable
grammatical sentences in English is infinite.
Everett takes the claim “All natural human languages have recursion.” to mean
that, if there exists a natural human language without recursion, the claim is false.
Or, slightly more subtly, if there exists a language which uses recursion so
minimally that linguists have a hard time determining whether a corpus of
linguistic data falsifies it or not, sentence-level recursion is probably not a
bedrock principle of human languages.
‘John, the man that put people in the water in order to clean them for God, that
lived in a place like a beach with no trees and that yelled for people to obey
God’.
‘John cleaned people in the river. He lived in another jungle. The jungle was like
a beach. It had no trees. He yelled to people. You want God!’
The non-recursive structure was accepted readily and elicited all sorts of
questions. I subsequently realized looking through Pirahã texts that there were
no clear examples involving either recursion or even embedding. Attempts to
construct recursive sentences or phrases, such as ‘several big round barrels',
were ultimately rejected by the Pirahãs (although initially they accepted them
to be polite to me, a standard fieldwork problem that Jeanette Sakel and I
discuss).
He does explicitly claim (in the aforementioned paper and elsewhere) that Pirahã
probably has no longest sentence, which is about the most generic anti-recursion
statement one can make.
For generative linguistics the recursive function is Merge, which combines two
:
words or phrases to form a larger structure which can then be the input for
further iterations of Merge. Any expression larger than two words, then,
requires recursion, regardless of whether there is embedding in that
expression. For instance the noun phrase “My favourite book” requires two
iterations of Merge, (Merge(favourite, book)= [Favourite book], Merge(my,
[favourite book])= [my [favourite book]]) and therefore is an instance of
recursion without embedding.
To be clear, this usage of ‘recursion’ seems consistent with how many other
Chomskyan linguists have used the term. And with all due respect to these
researchers, I find this notion of recursion completely insane, because it would
imply (i) any language with more than one word in its sentences has recursion,
and that (ii) all sentences are necessarily constructed recursively.
The first implication means that “All natural human languages have recursion.”
reduces to the vacuously true claim that “All languages allow more than one word
in their sentences.” 5 The second idea is more interesting, because it relates to
how the brain constructs sentences, but as far as I can tell this claim cannot be
tested using purely observational linguistic data. One would have to do some kind
of experiment to check the order in which subjects mentally construct sentences,
and ideally make brain activity measurements of some sort.
NS: But there are critics such as Daniel Everett, who says the language of the
Amazonian people he worked with seems to challenge important aspects of
universal grammar.
Chomsky: It can't be true. These people are genetically identical to all other
humans with regard to language. They can learn Portuguese perfectly easily,
just as Portuguese children do. So they have the same universal grammar the
:
rest of us have. What Everett claims is that the resources of the language do
not permit the use of the principles of universal grammar.
That's conceivable. You could imagine a language exactly like English except it
doesn't have connectives like "and" that allow you to make longer expressions.
An infant learning truncated English would have no idea about this: they would
just pick it up as they would standard English. At some point, the child would
discover the resources are so limited you can't say very much, but that doesn't
say anything about universal grammar, or about language acquisition.
Chomsky makes claims like this elsewhere too. The argument is that, even if there
were a language without a recursive grammar, this is not inconsistent with his
theory, since his theory is not about E-languages like English or Spanish or Pirahã.
His theory only makes claims about I-languages, or equivalently about our innate
language capabilities.
But this is kind of a dumb rhetorical move. Either the theory makes predictions
about real languages or it doesn’t. The statement that some languages in the
world are arguably recursive is not a prediction; it’s an observation, and we didn’t
need the theory to make it. What does it mean for the grammar of thought
languages to be recursive? How do we test this? Can we test it by doing
experiments involving real linguistic data, or not? If not, are we even still talking
about language?
To this day, as one might expect, not everyone agrees with Everett that (i) Pirahã
lacks a recursive hierarchical grammar, and that (ii) such a discovery would have
any bearing at all on the truth or falsity of Chomskyan universal grammar. Given
that languages can be pretty weird, among other reasons, I am inclined to side
with Everett here. But where does that leave us? We do not just want to throw
bombs and tell everyone their theories are wrong.
Okay, so far, so good. To the uninitiated, it looks like Everett is just listing all of the
different things that are involved in language; so what? The point is that language
is more than just grammar. He goes on to say this explicitly: (How Language
Began, Ch. 1, pg. 16)
His paradigmatic examples here are Pirahã and Riau Indonesian, which appears to
lack a hierarchical grammar, and which moreover apparently lacks a clear
noun/verb distinction. You might ask: what does that even mean? I’m not 100%
sure, since the linked Gil chapter appears formidable, but Wikipedia gives a pretty
good example in the right direction:
For example, the phrase Ayam makan (lit. 'chicken eat') can mean, in context,
anything from 'the chicken is eating', to 'I ate some chicken', 'the chicken that
is eating' and 'when we were eating chicken'
Is “chicken” the subject of the sentence, the object of the sentence, or something
:
else? Well, it depends on the context.
There is a wide and deep linguistic chasm between humans and all other
species. … More likely, the gap was formed by baby steps, by homeopathic
changes spurred by culture. Yes, human languages are dramatically different
from the communication systems of other animals, but the cognitive and
cultural steps to get beyond the ‘language threshold’ were smaller than many
seem to think. The evidence shows that there was no ‘sudden leap’ to the
uniquely human features of language, but that our predecessor species in the
genus Homo and earlier, perhaps among the australopithecines, slowly but
surely progressed until humans achieved language. This slow march taken by
early hominins resulted eventually in a yawning evolutionary chasm between
human language and other animal communication.
But Everett’s work goes beyond taking the time to bolster common sense ideas
on language origins. Two points he discusses at length are worth briefly exploring
here. First, he offers a much more specific account of the emergence of language
than Chomsky does, and draws on a mix of evidence from paleoanthropology,
:
evolutionary biology, linguistics, and more. Second, he pretty firmly takes the
Anti-Chomsky view on whether language is innate: (Preface, pg. xv)
… I deny here that language is an instinct of any kind, as I also deny that it is
innate, or inborn.
These two points are not unrelated. Everett’s core idea is that language should
properly be thought of as an invention rather than an innate human capability. You
might ask: who invented it? Who shaped it? Lots of people, collaboratively, over a
long time. In a word, culture. As Everett notes in the preface, “Language is the
handmaiden of culture.”
In any case, let’s discuss these points one at a time. First: the origins of language.
There are a number of questions one might want to answer about how language
began:
What aspects of human biology best explain why and how language
emerged?
Symbols are conventional links to what they refer to. They … need not bear any
resemblance to nor any physical connection to what they refer to. They are
agreed upon by society.
There are often rules for arranging symbols, but given how widely they can vary in
practice, Everett views such rules as interesting but not fundamental. One can
have languages with few rules (e.g., Riau) or complex rules (e.g., German); the
key requirement for a language is that symbols are used to convey meaning.
:
Where did symbols come from? To address this question, Everett adapts a theory
due to the (in his view underappreciated) American polymath Charles Sanders
Peirce: semiotics, the theory of signs. What are signs? (Ch. 1, pg. 16)
A sign is any pairing of a form (such as a word, smell, sound, street sign, or
Morse code) with a meaning (what the sign refers to).
Everett argues that symbols did not appear out of nowhere, but rather arose from
a natural series of abstractions of concepts relevant to early humans. The so-
called ‘semiotic progression’ that ultimately leads to symbols looks something like
this:
indexes (dog paw print) -> icons (drawing of dog paw print) -> symbols (“d o g”)
This reminds me of what little I know about how written languages changed over
time. For example, many Chinese characters used to look a lot more like the
things they represented (icon-like), but became substantially more abstract
(symbol-like) over time:
:
Eight examples of how Chinese characters have changed over time.
:
(source)
For a given culture and concept, the icon-to-symbol transition could’ve happened
any number of ways. For example, early humans could’ve mimicked an animal’s
cry to refer to it (icon-like, since this evokes a well-known physical consequence
of some animal’s presence), but then gradually shifted to making a more abstract
sound (symbol-like) over time.
Is there a clear boundary between indexes, icons, and symbols? It doesn’t seem
like it, since things like Chinese characters changed gradually over time. But
Everett doesn’t discuss this point explicitly.
Why did we end up with certain symbols and not others? Well, there’s no good a
priori reason to prefer “dog” over “perro” or “adsnofnowefn”, so Everett attributes
the selection mostly to cultural forces. Everett suggests these forces shape
language in addition to practical considerations, like the fact that, all else being
equal, we prefer words that are not hundreds of characters long, because they
would be too annoying to write or speak.
The feats Everett talks about are things like traveling long distances across
continents, possibly even in a directed rather than random fashion; manufacturing
nontrivial hand tools (e.g., Oldowan and Mousterian); building complex
settlements (e.g., the one found at Gesher Benot Ya'aqov); controlling fire; and
using boats to successfully navigate treacherous waters. Long before sapiens
:
arose, paleoanthropological evidence suggests that our predecessors Homo
erectus did all of these things. Everett argues that they might have had language
over one million years ago 6.
Everett’s anatomy arguments relate mainly to the structure of the head and larynx
(our ‘voice box’, an organ that helps us flexibly modulate the sounds we produce).
Over the past two million years, our brains got bigger, our face and mouth
became more articulate, our larynx changed in ways that gave us a clearer and
more elaborate inventory of sounds, and our ears became better tuned to hearing
those sounds. Here’s the kind of thing Everett writes on this topic: (Ch. 5, pg. 117)
What aspects of biology best explain all of this? Interestingly, at no point does
Everett require anything like Chomsky’s faculty of language; his view is that
language was primarily enabled by early humans being smart enough to make a
large number of useful symbol-meaning associations, and social enough to
perpetuate a nontrivial culture. Everett thinks cultural pressures forced humans to
evolve bigger brains and better communications apparatuses (e.g., eventually
:
giving us modern hyenoid bones to support clearer speech), which drove culture
to become richer, which drove yet more evolution, and so on.
He mostly seems to take issue with the idea that some region of our brain is
specialized for language. Instead, he thinks that our ability to produce and
comprehend language is due to a mosaic of generally-useful cognitive
capabilities, like our ability to remember things for relatively long times, our ability
to form and modify habits, and our ability to reason under uncertainty. This last
capability seems particularly important since, as Everett points out repeatedly,
most language-based communication is ambiguous, and it is important for
participants to exploit cultural and contextual information to more reliably infer the
intended messages of their conversation partners. Incidentally, this is a feature of
language Chomskyan theory tends to neglect 7.
Can’t lots of animals do all those things? Yes. Everett views the difference as one
of degree, not necessarily of quality.
What about language genes like FOXP2 and putative language areas like Broca’s
and Wernicke’s areas? What about specific language impairments? Aren’t they
clear evidence of language-specific human biology? Well, FOXP2 appears to be
more related to speech control—a motor task. Broca’s and Wernicke’s areas are
both involved in coordinating motor activity unrelated to speech. Specific
language impairments, contrary to their name, also involve some other kind of
deficit in the cases known to Everett.
I have to say, I am not 100% convinced by the brain arguments. I mean, come on,
look at the videos of people with Broca’s aphasia or Wernicke’s aphasia. Also, I
:
buy that Broca’s and Wernicke’s areas (or whatever other putative language areas
are out there) are active during non-language-related behavior, or that they
represent non-language-related variables. But this is also true of literally every
other area we know of in the brain, including well-studied sensory areas like the
primary visual cortex. It’s no longer news when people find variable X encoded in
region Y-not-typically-associated-with-X.
Still, I can’t dismiss Everett’s claim that there is no language-specific brain area.
At this point, it’s hard to tell. The human brain is complicated, and there remains
much that we don’t understand.
As I’ve tried to convey, Everett is well-qualified to write something like this, and
has been thinking about these topics for a long time. He’s the kind of linguist
most linguists wish they could be, and he’s worth taking seriously, even if you
don’t agree with everything he says.
V. THE REVELATIONS
I want to talk about large language models now. Sorry. But you know I had to do
this.
Less than two years ago at the time of writing, the shocking successes of
ChatGPT put many commentators in an awkward position. Beyond all the
quibbling about details (Does ChatGPT really understand? Doesn’t it fail at many
tasks trivial for humans? Could ChatGPT or something like it be conscious?), the
brute empirical fact remains that it can handle language comprehension and
:
generation pretty well. And this is despite the conception of language underlying
it—language use as a statistical learning problem, with no sentence diagrams or
grammatical transformations in sight—being somewhat antithetical to the
Chomskyan worldview.
Chomsky has frequently criticized the statistical learning tradition, with his main
criticisms seeming to be that (i) statistical learning produces systems with serious
defects, and (ii) succeeding at engineering problems does not tell us anything
interesting about how the human brain handles language. These are reasonable
criticisms, but I think they are essentially wrong.
Statistical learning also allows these systems to appreciate context and reason
under uncertainty, at least to some extent, since both of these are crucial factors
in many of the conversations that appear in training data. These capabilities
would be extremely difficult to implement by hand, and it’s not clear how a more
Chomskyan approach would handle them, even if some kind of universal-
grammar-based latent model otherwise worked fairly well.
:
Chomsky’s claim that engineering success does not necessarily produce
scientific insight is not uncommon, but a large literature speaks against it. And
funnily enough, given that he is ultimately interested in the mind, engineering
successes have provided some of our most powerful tools for interrogating what
the mind might look like.
The rub is that artificial systems engineered to perform some particular task well
are not black boxes; we can look inside them and tinker as we please. Studying
the internal representations and computations of such networks has provided
neuroscience with crucial insights in recent years, and such approaches are
particularly helpful given how costly neuroscience experiments (which might
involve, e.g., training animals and expensive recording equipment) can be. Lots of
recent computational neuroscience follows this blueprint: build a recurrent neural
network to solve a task neuroscientists study, train it somehow, then study its
internal representations to generate hypotheses about what the brain might be
doing.
When we peer inside language-competent LLMs, what will we find? This is a topic
Everett doesn’t have much to say about, and on which Chomsky might actually be
right. Whether we’re dealing with the brain or artificial networks, we can talk
about the same thing at many different levels of description. In the case of the
brain, we might talk in terms of interacting molecules, networks of electrically
active neurons, or very many other effective descriptions. In the case of artificial
networks, we can either talk about individual ‘neurons’, or some higher-level
description that better captures the essential character of the underlying
algorithm.
:
Maybe LLMs, at least when trained on data from languages whose underlying
rules can be parsimoniously described using universal grammar, effectively
exploit sentence diagrams or construct recursive hierarchical representations of
sentences using an operation like Merge. It’s still possible that formalisms like
Chomsky’s provide a useful way of talking about what LLMs do, if anything like
that is true. Such descriptions might be said to capture the ‘mind’ of an LLM,
since from a physicalist perspective the ‘mind’ is just a useful way of talking about
a complex system of interacting neurons.
Regardless of who’s right and who’s wrong, the study language is certainly
interesting and we have a lot more to learn. Something Chomsky wrote in 1968
seems like an appropriate summary of the way forward: (Language and Mind, pg.
1)
1 Chomsky 1991b refers to “Linguistics and adjacent fields: a personal view”, a chapter
of The Chomskyan Turn. I couldn’t access the original text, so this quote-of-a-quote
will have to do.
:
2 Chomsky’s domination of linguistics is probably due to a combination of factors.
First, he is indeed brilliant and prolific. Second, Chomsky’s theories promised to
‘unify’ linguistics and make it more like physics and other ‘serious’ sciences; for
messy fields like linguistics, I assume this promise is extremely appealing. Third, he
helped create and successfully exploited the cognitive zeitgeist that for the first time
portrayed the mind as something that can be scientifically studied in the same way
that atoms and cells can. Moreover, he was one of the first to make interesting
connections between our burgeoning understanding of fields like molecular biology
and neuroscience on the one hand, and language on the other. Fourth, Chomsky was
not afraid to get into fights, which can be beneficial if you usually win.
3 One such sound is the bilabial trill, which kind of sounds like blowing a raspberry.
5 Why is this vacuously true? If, given some particular notion of ‘sentence’, the
sentences of any language could only have one word at most, we would just define
some other notion of ‘word collections’.
7 A famous line at the beginning of Chomsky’s Aspects of the Theory of Syntax goes:
“Linguistic theory is concerned primarily with an ideal speaker-listener, in a
completely homogeneous speech community, who knows its language perfectly and
is unaffected by such grammatically irrelevant conditions as memory limitations,
distractions, shifts of attention and interest, and errors (random or characteristic) in
applying his knowledge of the language in actual performance.”
Previous Next
678 Comments
Write a comment...
:
Oldest first
Thanks for that link. I never quite understood the distinctions (or non-
distinctions) between self-embedding and MERGE, but the intro explains it
rather well.
REPLY SHARE
Why does AI research have so many citations? Particularly when compared with
(other branches of?) Computer Science?
REPLY (8) SHARE
Yes, what is going on here? And not just AI research but Deep Learning in
particular (Bengio, Hinton). This area has now captured a significant majority of
researchers for all time?
REPLY SHARE
Because there are so many AI researchers. I wonder how the numbers would
look like if normalized by the size of respective research communities.
REPLY (1) SHARE
:
Kenny Easwaran Jul 21
In some of these cases it’s hard to say what “the research community” is. I
would have expected Marx to be higher, since the research community of
all Soviet scientists working on any subject was larger than any particular
modern research community.
REPLY (1) SHARE
delesley Jul 20
Because the number of researchers has exploded in the past 10 years, driven
by massive investment from industry, rather than academia. Big AI conferences
:
now get more than 10,000 submissions a year. Other areas of computer
science are primarily academic, and are at least two orders of magnitude
smaller.
Hinton's been around for a long time, but if you look at citations prior to 2014
or so, when neural nets were still mostly an academic curiosity, the number will
look much more reasonable.
REPLY SHARE
I did a little experiment. I was browsing some of the papers that cite Chomsky's
most cited book, so I picked up one paragraph from the abstract of one of the
top results that Google Scholar produced, fed it to an AI chatbot (Grok) and
asked for reference suggestions:
MugaSofer Jul 21
Did it suggest any papers actually cited by the allegedly human author?
Were they even real papers (hallucinated citations are a real issue with
LLMs)?
REPLY (1) SHARE
5 out of 8 were actually cited by the human author of the book (it was
a book). Another one was the original book.
REPLY SHARE
:
Arby Jul 21
Wasserschweinchen Jul 19
There is a big difference between the small amount of language a baby hears when
learning to speak and the large amount of language artificial intelligence needs. It is
not clear that artificial intelligence disproves Chomsky's theories.
REPLY (3) SHARE
B_Epstein Jul 20
This, indeed, doesn’t disprove Chomsky- but it means the poverty of stimulus
argument is far from decisive, as the stimulus is not all that poor.
REPLY SHARE
Bugmaster Jul 20
I am guessing that the baby has most of the deep layers already hard-coded
(by evolution), and only requires some equivalent of a LoRA to get up to speed
in his native language. The LLM has to be taught from scratch.
:
REPLY (2) SHARE
Bugmaster Jul 20
No, not language itself (that'd be the LoRA), but rather the facility to
learn language (however that might function, which I suspect no one
knows).
REPLY (3) SHARE
LLMs don't have that luxury: they need to learn a language that
evolved not for ease of understanding by them, but by some
aliens with completely different quirks in their underlying
systems.
REPLY (2) SHARE
Bugmaster Jul 20
Freedom Jul 22
Continue Thread →
Laplace Jul 22
The human genome doesn't really have enough room to provide all that
much hard coding. The entire thing is a couple of gigabytes, and most of it
doesn't have anything to do with how particular synapses in the brain are
wired. You've got basic reflexes, the learning system for training up the
bulk of the mind mostly from scratch, and I'd guess not much else.
1) Get a lot more training signal than just language. You've got to count all
the visual information, audio information, etc. as well. This data is probably
:
lower quality, but it still helps a lot.
Bugmaster Jul 22
Leppi Jul 22
pozorvlak Jul 22
Laplace Jul 22
Laplace Jul 22
And the point here is that x<<2^11, or ca. 3 gigabytes. And I think
that's likely not enough to encode all the possible different
pieces of deep language heuristics.
Now in reality, the genes don't actually need to encode the entire
blueprint for the language faculties. They just need to encode
some initial setup that will predictably become a language
faculty once the neural network starts training. This requires
much less description length. You just need to encode the right
optimiser and the right sort of architecture.
Laplace Jul 24
Continue Thread →
vindication Jul 21
Just rowing in here with my usual comment that we have no idea if humans are the
only animals with language.
REPLY (3) SHARE
At some point, this becomes a matter of definition, but for most intents and
purposes, we kinda do. (My go-to book for this is Zorina & Smirnova's "What
speaking animals taught us", but it is in Russian.)
REPLY (1) SHARE
It just seems astounding unlikely that we are so unique ... but if you have a
book with good arguments otherwise in a language I can read (just
English, sorry!) then I'll gladly read it
REPLY (3) SHARE
DanielLC Jul 21
We already know we're unique. We've built a giant civilization and all
sorts of crazy technology. Our uniqueness including our language
isn't that surprising.
That said, we also know we're not unique in terms of sheer brain
power. I don't know what pilot whales and the like need all that brain
power for, but it's at least plausible that they could use language.
1123581321 Jul 20
I’ve long suspected marine mammals are quite intelligent, and it’s a small leap
of imagination to think they have a language. Water is an excellent sound
conductor; natural underwater waveguides allow sound propagation for literally
thousands of miles. Perfect environment to develop a language.
REPLY (1) SHARE
This is very far from a proof that whale communication is language. I think
it’s absolutely right to remain open-minded about whether whale
communication is or isn’t like human language in whatever sense you
mean. There might be a few bird species whose communication is
possibly like human language.
1123581321 Jul 21
Did you read the whole piece? Where, for example, the researchers
describe “[…]the sperm whale "phonemes" could be used in a
combinatorial fashion, allowing the whales to construct a vast
repertoire of distinct vocalisations. The existence of a combinatorial
coding system, write the report authors, is a prerequisite for "duality
of patterning" – a linguistic phenomenon thought to be unique to
human language – in which meaningless elements combine to form
meaningful words.”
They are pretty far along the road toward “yes whales have a
language”. The help of machine learning pattern recognition is useful,
but far from the only thing pointing toward it.
REPLY (1) SHARE
Here is the key chart, where they classify all 8000 or so codas in
their database on the combinatorial dimensions of rhythm,
tempo, and ornamentation:
https://www.nature.com/articles/s41467-024-47221-8/figures/3
1123581321 Jul 22
Yeah I’m with you on science coverage, and not just BBC of
:
course. This one seems reasonable enough.
Note - it's not that they have 600 distinct "words" - it's
that they have recordings of about "600 words" worth
of conversation total. In any human language, about a
hundred of those words would likely be repeats.
Continue Thread →
vindication Jul 21
>The objection then becomes that grammar is more than just structure
but a structure of a certain complexity defined to separate human
language from other animals, which seems very post-hoc to me.
You have a good point, but there are fairly obvious measures of the
complexity of human language which I'd be surprised to see any other
animal's language reach.
:
For instance, human sentences are sufficiently varied that, barring quotes,
repeated sentences (except exceedingly short ones, "Damn!" is common)
are rare. Does any other animal match this? This doesn't require any
specific properties of the language's syntax.
REPLY SHARE
The patterns by which these elements are combined are recursive (e.g.,
any two sentences can be combined with “and”, noun phrases can
combine with a preposition to form a modifier within a bigger noun
phrase).
REPLY SHARE
Does anyone know what the Chomsky followers would say about the recent findings
that non-human animals (a large number of them) clearly have methods of
communication?
REPLY (2) SHARE
I'm not sure what important recent findings you're referring to; the fact of non-
human communication is ancient knowledge. What Chomskyans would say is
that human language is evolutionarily discontinuous (it is not an outgrowth of
the same phenomenon in other species) because there are categorical
differences. Nonhuman communication systems are syntactically basic
(certainly non-recursive), goal-orientated, lack reference, conceptual labelling,
symbol arbitrariness and so on. Claims to the contrary based on e.g. training
apes to sign have always anthropomorphised the data and there is not wide
acceptance that any species has demonstrated human-like capabilities.
REPLY (2) SHARE
Ryan L Jul 19
Can you summarize, for a complete lay person in this area, what the
evidence is to support this claim?
REPLY (1) SHARE
Ryan L Jul 19
Continue Thread →
jumpingjacksplash Jul 19
Continue Thread →
Ryan L Jul 19
https://www.reddit.com/r/self/comments/3yrw2i/i_never_thought_with
_language_until_now_this_is/
The person who wrote that claims that they used language to
communicate but not to think. I realize it's one, self-reported Reddit
post, so not exactly high-quality data, but I wonder if there are other
claims like this, and if they have been investigated by linguists?
REPLY (2) SHARE
There are other claims like this, and they are true at some level
and false at another. We often indeed don't think in language in
the sense of "build a sentence, up to and including how it
sounds/what the articulatory commands are" (the
acoustic/articulatory debate is its own bag of worms), but the
idea is that to formulate _complex_ thoughts specifically, we use
:
a method to chain thoughts which is so similar to syntax it
probably _is_ syntax, even if it never gets sent to the systems of
articulation ("is never sent to spell-out", in jargon-y terms). Note
how using language was tremendously helpful for the guy,
because it helped organize thought.
REPLY (4) SHARE
Ryan L Jul 19
Continue Thread →
luciaphile Jul 20
Continue Thread →
Continue Thread →
(I use inner speech a lot too, and it is very useful for certain
kinds of thought. It is but one modality of thought.)
First,
Second,
Continue Thread →
Xpym Jul 20
Ryan L Jul 19
https://www.nature.com/articles/s41559-024-02420-w.epdf
REPLY (2) SHARE
They may communicate, but they don't seem to do language well. All the ape-
language studies from the 70s and 80s failed miserably (the most famous of
these research projects, Coco the gorilla, made a good mascot but terrible
science, particularly as the researcher refused to share or release unedited
video of Coco communicating). It seems like even smart animals, like chimps,
can't grok language the way humans can.
https://www.psychologytoday.com/us/blog/the-origin-words/201910/why-
chimpanzees-cant-learn-language-1
REPLY (2) SHARE
Thank you, this is very informative. I'm thinking about results from whales
and elephants. Do you have thoughts on whether this counts as
"language" from a Chomskian view?
Having a name for each elephant that they use to get each other’s
attention is more sophisticated than a finite list of alarm calls, but it
does not mean that there is productive recombination of utterances
to express novel meanings. One of the major features of language is
that you can (largely) understand novel expressions that you have
never heard before, and the limitations on your understanding aren’t
due to the novelty, but due to inherent uncertainties of context and
intention.
REPLY SHARE
Bldysabba Jul 20
They can't grok human language the way humans can. I don't understand
why this is sufficient to decide that they can't grok language? Surely one
would need to study how they communicate with each other to decide
that?
REPLY (1) SHARE
:
Kenny Easwaran Jul 21
mst Jul 19
One interesting note: apparently Ken Hale, late well respected field linguist who did
a lot of work on the Australian language Walpiri among many other things,
uncontroversially described Walpiri and some other Australian indigenous
languages as lacking clausal embedding in the 70s. This Language Log blog post
(http://itre.cis.upenn.edu/~myl/languagelog/archives/004592.html) muses on
various possible reasons this basically went unnoticed whereas Everett immediately
got in an academic flamewar, but the main reason seems to have basically been
that recursion hadn't been proposed as a linguistic universal until the introduction
of Chomsky's Minimalist Program in 1995, earlier Chomskyan models of universal
grammar did not include it.
So the idea that this may not even be unique to Pirahã seems like an important
counterpoint to some claims I've seen that Pirahã in fact probably does have
syntactic recursion and the real problem is that its sole expert isn't describing the
language correctly. I've seen Everett comment somewhere that a problem may be
that field linguists are essentially trained to fit grammar into a certain framework
and so may be trying to square peg a round theoretical hole (as he himself says he
did in his original PhD thesis description of Pirahã), thus failing to properly
document the true linguistic diversity that really exists in the world.
Disclaimer that I don't have any expert understanding of the technical issue here,
just a longstanding hobbyist fascination with linguistics. I acknowledge that my
amateur understanding is causing me to miss something important here. But overall
my reactions to the Chomskyan team response to this, where Everett is probably
:
wrong but even if he's right it actually doesn't matter for his theory, strikes me
Expand full comment
similarly to you: it feels like dragging a major hypothesis into an unfalsifiable
REPLY (1) SHARE
The claim that all human languages have sentences containing more than one word
isn’t vacuously true, it’s profoundly true. Just compare this fact with the norm in the
communications of other species. No equivalent of “the [color] bird” is known in
how any other species communicates. That, is Merge.
REPLY (2) SHARE
The claim _is_ profoundly true, but note the dolphin comment above.
REPLY (1) SHARE
Does, as the author of the article suggest, this mean that any
language that has more than one word thereby has recursion?
:
REPLY (1) SHARE
Say what??
Watch out for Jesus! His Chomsky rating is off the charts.
REPLY (1) SHARE
I've done editing in the past. I just can't remember how to do it.
REPLY (1) SHARE
LLMs pick up language in super interesting ways, for example (for any of my
Hinglish speakers out there), try asking ChatGPT:
"Bilkul, main aapse Hindi mein baat kar sakta hoon. Aapko kis cheez ki madad
chahiye?"
:
REPLY (3) SHARE
"Mary saw a snake in front of her" _can_ mean "Mary saw a snake in
front of herself".
And, finally, "Mary knows that Penny saw herself" cannot mean "Mary
knows that Penny saw her" (and vice versa, but the vice versa is
uninteresting because it's the same as the first).
There are very pecuilar and technical laws surrounding both when a
reflexive _can_ appear and when it _must_ appear.
REPLY (2) SHARE
B_Epstein Jul 20
Continue Thread →
(Note that nobody explicitly teaches children the laws, and, judging
by CHILDES corpora, they don't get enough data to arrive at the very
generalizations they tend to arrive at.)
REPLY SHARE
immergence Jul 19
Oh wow, ChatGPT does not realize that this is written using the Latin script.
After directly asking it, it says it's written in Devanagari.
Is that what you meant with this example? I do not speak Hindi
REPLY (1) SHARE
Sorry, should've been clearer! I've written Hindi words in English, and
ChatGPT is responding in Hindi written in English after "recognizing" that
that's what I'm doing. Its ability to mimic that form of Hindi is what I was
trying to highlight.
REPLY (1) SHARE
Cremazie Jul 20
Jake Jul 20
skaladom Jul 21
immergence Jul 25
ChatGPT(-4o) can do much more impressive stuff than that with transliterated
language. Try asking it novel questions in romanized Hindi, or in English
transliterated into devanagari. (e.g. "Who was Prime Minister of Australia on
the day Björn Þórðarson became Prime Minister of Iceland?") I can't say for
sure whether it will be able to handle these sorts of questions using romanized
Hindi or devanagari English, but it seems to do pretty well when I ask questions
in katakana English or romanized Japanese.
"If John is standing to the right of Jane, who is standing to the left of
Jane?"
"Who was Taoiseach of Ireland during the time when John Major was
Prime Minister?"
Even here, ChatGPT gets the answer right at least some of the time.
But the answers seem inconsistent across separate chat sessions. It
has no problem naming Gordon Brown as the next PM after Blair, but
gets confused trying to make the connection to Ireland. Sometimes it
Expand full comment
correctly answers "Bernie Ahern"; sometimes it names Brian Cowen
REPLY SHARE
:
Callum Hackett The Great Exhibition Jul 19
I think this is a shallow take on a complex field. I appreciated that it's well-
researched but the sources have not been well understood, though admittedly the
review is from a non-linguist, and admittedly the discourse in the research literature
is often unprofessional and opaque.
The objections raised against Chomsky's research program are mostly intuitive and
therefore trivial - most were dealt with in the 1950s rejection of structuralism and
behaviorism (which is foundational to almost all of cognitive science, which is why
Chomsky is hugely cited and not as controversial as this makes out), while others
betray a misunderstanding of what generative linguistics studies (this is a
consequence of 'language' itself acceptably meaning many things to many people,
though the review didn't adequately tease these meanings apart).
LearnsHebrewHatesIP Jul 19
The field is large and most working Chomskyan linguists have zero interest
in Everett, which exactly matches the expectation that a secure field
would not take note of him. Popular understanding of his impact is
overestimated because media (even science media) loves drama.
tup99 Jul 19
Hello there :)
Catmint Jul 23
REPLY SHARE
jumpingjacksplash Jul 19
Isn't, "your theory predicts always ¬X, look, I've found some X,"
basically how taking on a scientific hypothesis is meant to work?
REPLY (1) SHARE
Bldysabba Jul 20
Continue Thread →
I agree. For me, this is why it's important that these people are a
small minority. In the wider context of quality research being
done in the field (imo), it makes more sense to conclude that
these people are a type who wilfully indulge in undignified
behaviour - I think there are people in lots of fields who have this
trait even when they are broadly right on the technical detail.
REPLY (1) SHARE
:
Dave Madeley Jul 19
Thanks for the reply. It’s the human condition for sure. I
don't have much stake in this, except we have a problem
with how 'elites' are viewed, and although a lot of that
discourse is toxic, I have some sympathy with this having
done a humanities degree and seeing what "cultural
marxism", for want of a better word, can do to the study of
beautiful things. Can it do the same thing to the true as to
the beautiful? Well here you have a Marxist who apparently
totally dominates an emerging science and his defenders
will go beyond the confines of typical academic discussion
to defend him and that raises my eyebrow, I suppose. A
person can be completely on top of the technical detail and
still be an ideologue. In fact that's mostly how the left has
come to dominate academia. But I'm getting into
unparsimonious
Expand ideas
full comment of my own!
REPLY (1) SHARE
What
Expandfield are you in?
full comment
Continue Thread →
Thanks!
-----------
Expand full comment
Seperately, there's another route of defense available, and
REPLY (1) SHARE
TGGP Jul 19
Did you actually read the linked reddit post? That was a Chomsky skeptic
who still thought Everett was unscientific. There isn't the necessary
connection between the things that you think there is.
REPLY SHARE
...Have you heard of a guy called Fomenko and his "New Chronology"?
Most historians ignore him, but those who do very rarely pretend that this
is a respectable opponent.
Sokow Jul 20
Continue Thread →
Continue Thread →
:
Dmitrii Zelenskii Jul 21
Continue Thread →
FrancoVS Jul 20
I'd argue anticolonialist politics was probably a big factor: Everett was an
American that travelled to Brazil to study a remote tribe and made an
ethnographic account, at a time it was becoming increasingly clear that
this had happened before and the result was questionable methodology
and disastrous consequences for the natives. (Napoleon Chagnon and the
Yanomami)
Also, Everett the charming missionary certainly look the part of the gringo
with an agenda.
Well, to give credit where credit's due, Everett specifically rejects the
idea of their being mentally deficient. (Whether it is actually logically
consistent with his ideas is a tougher question, but people are known
to espouse inconsistent or not-proven-to-be-consistent ideas.)
REPLY SHARE
The mental capacity for language? Seemingly so. And then, by the usual
genetic drift (probably the version for weakly-beneficial mutations,
because being able to do complex thoughts is beneficial, including for
intra-group competition), spread to the rest of the population in one of the
bottleneck events.
REPLY (3) SHARE
Why does he think that? Do you agree with him? What kind of
evidence is there that this is true?
REPLY (1) SHARE
Partly but not fully. The problem is, their brain size
_and_ their brain _zones_, to the extent they're
recognizable from the cranium, are the same as in
modern humans. So, we wouldn't expect huge
difference in "horsepower" of the mind.
REPLY (1) SHARE
Continue Thread →
Exactly.
REPLY (1) SHARE
Continue Thread →
Can you clarify? They haven't done much in what sense and
in what timeline compared to what other oganisms? And
what is the claim evidence for and why is it evidence for that
claim?
*The claim is that the crucial capacity for syntax and its
interpretation appeared first, in what amounts to a
single mutation, and then that system got hooked up to
communicative system in what is plausibly a
_different_, later mutation. So, two mutations: to think
and to articulate the thoughts.
Continue Thread →
So, that 200000 year old skull may well have not had
that _second_ mutation yet.
REPLY SHARE
Pedantry alert: Genetic drift is when it's random (e.g. you kill off half
the population, turns out this gene was more common in one half
than the other, so now it's more common in descendant populations).
Language reaching fixation would be directional selection since, as
you note, it would be beneficial.
REPLY (1) SHARE
There was no major bottleneck even in the last 150,000+ years and
the Khoisan split off before 150,000 years and they do language just
fine. There has been minimal gene flow between them and other
humans since then and the genes that have been selected for when
they interbred with some Cushitic pastoralists 2000 years ago aren't
related to language.
Continue Thread →
(There are a lot of components in how the entire edifice of language arose:
for instance, it required greater control of mouth and/or head movements,
and the hominid excavations suggest that this arose earlier, which
confuses scores of researchers into thinking that's when language arose,
including even such brilliant minds as Svetlana Burlak - but it is quite
obvious that better control of fine motorics would benefit tasks beyond
language.)
REPLY SHARE
Continue Thread →
The 1950s rejection of behaviorism was important and powerful, but it relied on
a lot of arguments about what could and couldn’t in principle be done that have
since been falsified by large language models and other work. Many of
Chomsky’s arguments early on involve the idea that you can’t make sense of
probability as anything but a frequency or some simple observable feature like
pitch or volume, which is an appropriate refutation of Skinner, but not of any of
the many constructions of Bayesian probability in the second half of the 20th
century. He also argues that you need to study language separately from the
world, but modern large language models seem to do a good job of creating a
model of the language-world system, whether or not they adequately separate
it into the two components.
REPLY (1) SHARE
I'm sorry, but since I've recently learned French, the whole time I was reading this I
was thinking about this: https://www.instagram.com/reel/C9hlm5Du71u/?
igsh=MWV0bmRsM3dya3RhMQ==
REPLY (1) SHARE
SurvivalBias Jul 23
English is also full of this kind of thing if learned as a second language. E.g.
write/right/rite/riot, not to mention such wonderful pairs as come/cum,
sheet/shit, beach/bitch, can't/cunt, fox/fucks.
REPLY (1) SHARE
Kevin M. Jul 19
I'm reminded of Max Planck's line, "Science progresses one funeral at a time."
REPLY (1) SHARE
:
Forrest Jul 20
Oh, stop with the pearl-clutching! The man is 95. His death has been
imminent for years.
REPLY (1) SHARE
Forrest Jul 24
Still, I don't think it's in good taste to eagerly root for someone to die
because you disagree with his ideas about linguistics.
REPLY SHARE
> These two points are not unrelated. Everett’s core idea is that language should
properly be thought of as an invention rather than an innate human capability. You
might ask: who invented it? Who shaped it? Lots of people, collaboratively, over a
long time. In a word, culture. As Everett notes in the preface, “Language is the
handmaiden of culture.”
Yup.
REPLY SHARE
I don't understand the relevance of the question of whether it's one or the
other. Obviously on some level, there is [information encoding of perceived real
world things] occurring in the human (or animal) brain, dependent on the [info
:
encoding - "brain"] capabilities of that creature. Is that an internal language? I
don't know, does the categorization matter? I'm a computer programmer and
certainly it sounds like a language to me, for whatever that's worth.
This means that you could reverse the process and do something in the real
world (sounds, calls, whistles) that mirrors that encoding in your brain - in a
way that other brains could interpret in the same, real-world verifiable way, for
evolutionarily beneficial reasons like coordination on hunting. This seems to of
course be both invented/learned (i.e. I point at the tiger and make the same
sound, and you understand that it is a reference to it) and innate (people all
around the world seem to understand that the word "squirrel" is more likely to
describe a small fluffy thing than "hippopotamos").
But I don't get why it matters. And by matters I mean "in a way that, if we prove
it's one or the other, that's a thing that can be objective proven, rather than
just vague category definitions, and then informs our understanding of the
science such that we can apply it to productive new insights."
> But I don't get why it matters. And by matters I mean "in a way that, if
we prove it's one or the other, that's a thing that can be objective proven,
rather than just vague category definitions, and then informs our
understanding of the science such that we can apply it to productive new
insights."
If it's just brain power + culture, that implies you can just keep sticking
more neurons in a dog's brain, and at 86 billion it'll be able to use a full
language. If there's some specialized structures, you'll either need a lot
more than 86 billion neurons (so the resultant creature can use
nonspecialized bits, like how we solve math problems) or to introduce
:
analogous structures.
Hi David Brin! :)
"If it's just brain power + culture, that implies you can just keep
sticking more neurons in a dog's brain, and at 86 billion it'll be able to
use a full language"
So replace "dog" with "Chatgpt" and obviously we've got a thing that
is pretty language competent, right? I feel like LLMs are the
example/thing we're all implicitly referring to?
" If there's some specialized structures, you'll either need a lot more
than 86 billion neurons (so the resultant creature can use
nonspecialized bits, like how we solve math problems) or to introduce
analogous structures."
But here's where the categorization breaks down for me. Obviously
for human language there ARE specialized bio/neuro structures that
are necessary for actual spoken human language: the parts of your
brain that encode information, voice boxes, those parts of the brain
the article above talks about, etc... for a LLM, though, do we
categorize them as having those things or not? Does the screen and
text output sub-module and hard drive structure count? Why does it
matter,full
Expand atcomment
least in an important way?
REPLY (1) SHARE
See the other comment of mine on LLMs :) LLMs are the pinnacle
of "every time I fire a linguist, productivity goes up", because,
well, constraints of natural language learning by a child are quite
different from constraints on computers. Look up CHILDES
corpus: it is small and unrepresentative (poverty of stimulus
:
argument), if you try to teach LLMs on what children actually
hear, rather than on what gazillions on balanced and generally-
more-complicated data OpenAI feeds them, you won't get as
good results as children routinely get, and you won't get biases
toward learnability of existing structures over non-existing. This
suggests that LLMs do just brute-force the problem.
REPLY (1) SHARE
Continue Thread →
>But I don't get why it matters. And by matters I mean "in a way that, if we
prove it's one or the other, that's a thing that can be objective proven,
rather than just vague category definitions, and then informs our
understanding of the science such that we can apply it to productive new
insights."
This has been made moot by LLMs' linguistic success, but, if that
_hadn't_ happened, then the question of "Does language competence
require some special structure, inborn in humans, or is it purely a cultural
artifact, learnable by the same general purpose learning algorithms that
we use for everything else?" would have mattered to AI design. AFAIK, we
now get to ask the same sort of question, but probably about e.g.
executive function, instead.
REPLY (1) SHARE
Many Thanks!
REPLY SHARE
jumpingjacksplash Jul 19
I don't see how humans couldn't have evolved language faculties once they
had language, but developing a cludgy capacity to be able to communicate
with one another, then gradually evolving dedicated capacities to massively
upscale it to the point where it's helpful for thinking seems way more likely
than developing full-on linguistic thought in one step then using it to
:
communicate (presumably originally just among the kids of whatever super-
hominid developed this ability).
REPLY (1) SHARE
People start burying their dead around the same time that they
start painting on cave walls and carving figurines, as well as
other stuff. https://en.wikipedia.org/wiki/Behavioral_modernity
TheGreasyPole Jul 22
Then .... one day.... one human or another discovers he can make
sounds to share this "intuitive personnal i-langauge" with someone
else, and the "someone else" groks it !!! Between the two of them
they are now able to *share* the highly complex thought patterns that
have been evolving for generations for the first time.
But then once they can *actually communicate ideas* it can move
explosively fast as it is now cultural! You can go from 0 to "burying
Expand full comment
your dead and painting deers on caves" in a few generations as
REPLY SHARE
DannyTheSwift Jul 19
Sharkey Jul 20
You don't think those kids' parents ever pointed at things or gestured to them
in any way?
I don't think their home signs included an actual grammar or anything, no.
And keep in mind that only the young kids developed full fluency.
REPLY SHARE
Agreed. Just as, after the initial invention of cooking, our digestion and
cuisine co-evolved to become more and more dependent on cooking, and
to more thoroughly exploit the advantages of that capability, if a linguistic
"MERGE" capability resulted from a single mutation, I would expect that it
would have e.g. increased the advantage that a larger vocabulary
conferred, and, more generally, increased the frequencies of genes that
provided complementary linguistic capabilities. So, in the end, many
genes may be involved in our current linguistic capabilities, above what
were present pre-"MERGE".
REPLY SHARE
1.
Why then:
>He does explicitly claim (in the aforementioned paper and elsewhere) that Pirahã
probably has no longest sentence, which is about the most generic anti-recursion
statement one can make
2.
>I find this notion of recursion completely insane, because it would imply (i) any
language with more than one word in its sentences has recursion
From the quotes, it sounds like the linguisists in question emphasize the role of
recursion as an iterative - not isolated phenomenon. A sentence with two words
might be a primitive instance of recursion, but lack the iterative flexibility that
characterizes it.
:
3.
>But this is kind of a dumb rhetorical move. Either the theory makes predictions
about real languages or it doesn’t
This objection doesn't seem fully compelling. All models are wrong, but some are
useful. The mere existence of a supposed language doesn't prove that the model
Expand full comment
isn't generally useful. Did not the vast majority of langauges discovered since the
REPLY (2) SHARE
Regarding:
I think this was just a typo, honestly. (Two other threads on this comment page
think the same.)
REPLY SHARE
The example that I've seen cited is that thermodynamics owed more to heat
engines than the other way around. I would guess that this is probably true of
mechanical engineering preceding the science of classical mechanics as well.
I'd guess that electromagnetism is probably the other way around: There was a
lot of scientific investigation before the first motor.
REPLY (1) SHARE
Metallurgy is probably the most extreme version of this. Smiths knew how
to create metals with very specific properties with precision without
understanding anything about the underlying chemistry or physics. The
actual understanding of the chemistry of how to make steel and bronze
and, why the specific heatings, coolings, hammerings and ratios of
materials used were necessary, came far later.
REPLY (1) SHARE
This was a well written overview of a fascinating topic, but I don't think it's a very
good book review. Like so many of these reviews, the author uses it primarily as an
opportunity to review a topic they are interested in, rather than a book about that
topic. Obviously, one needs to have some background knowledge of the topic, but
when the the actual book under review only takes up one out of five sections, it's
not really about the book or the ideas it discusses.
This is neatly illustrated by this sentence: "There’s a lot of interesting stuff in there I
haven’t talked about..." But isn't talking about the interesting stuff the main point of
a book review?
Yep, plenty of "book reviews" are just guest posts in disguise. Since those
aren't allowed in the pure form, this kind of thing is the result.
REPLY SHARE
Jordan Aug 4
Exactly my thoughts, great article, not a great book review. Also think some
editing on subject ordering would have been good (define recursion when it
was mentioned, maybe move the chomsky stuff a little later)
REPLY SHARE
I’m afraid Chomsky wins this one because of the work in computation theory that
he did which the reviewer doesn’t reference at all. Chomsky’s biggest insight was
what is now known as the Chomsky Hierarchy of languages or grammars:
Recursive (in the Godelian sense) languages, recognized by Turing machines and
other automata which can use expanding amounts of rewritable memory to keep
track of things, attained by humans with the invention of writing (pre-literate
humans were not Turing Complete and the Church-Turing thesis is not actually
about what humans can compute but about what technology-aided humans can
compute).
Pirahã people (or at least the young children in it) have the capability to learn other
human languages and the same brain improvements that the rest of us have. Their
language is restricted in interesting ways but the only really important question
about it, at the very top level, is whether they are capable of doing things that
people in other societies can do, or whether their language makes this impossible.
Expand full comment
"Context-Free Languages" add "memory" but it seems (naively, to me) like it's
just more complex "memory."
... how do we know it's not just straight memory capacity? Wouldn't
that be exactly what we see when you say:
IExpand
meanfull
oncomment
some level, the brain kinda has to work that way right? I
could see if the "quality" was an LLM-like capability of
REPLY (1) SHARE
You didn’t read the part where I said that unaided human minds are NOT
Turing-complete. They are only Turing-complete when combined with the
TECHNOLOGY of writing.
Laplace Jul 22
> We know for sure that a finite neural net like an LLM *cannot* reproduce
a context-free grammar
I don't know much about linguistics, but modern LLMs are basically Turing
complete. At least in the sense that a normal home computer, or a human
with a finite amount of memory and paper to write on are basically Turing
complete.
REPLY SHARE
> To be clear, this usage of ‘recursion’ seems consistent with how many other
Chomskyan linguists have used the term. And with all due respect to these
researchers, I find this notion of recursion completely insane, because it would
imply (i) any language with more than one word in its sentences has recursion, and
that (ii) all sentences are necessarily constructed recursively.
:
No. The claim is actually somewhat more technical. Dolphins have the equivalent of
sentences, but they are linear, not recursively hierarchical. An illustration: in English,
"I saw a girl with a telescope" has two readings, "I [saw [a girl with a telescope]]"
and "I [[saw a girl] with a telescope]" (i.e. using a telescope). Such structural
ambiguity is not possible in dolphinspeak, where linear position _is_ hierarchical
position but is possible in human languages.
More generally, Everett is exactly as much of a buffoon as Nevins claims he is, and
"common sense" is a poor guide to linguistics.
REPLY (4) SHARE
And more generally, "the lowest level" is neurons firing electrons and ions,
which is no more useful than "silicon pieces firing electrons" for computer,
we don't actually describe what computer does that way. (I am reading
"Gödel, Escher, Bach" now, and there's a related discussion.)
REPLY (1) SHARE
"And more generally, "the lowest level" is neurons firing electrons and
ions, which is no more useful than "silicon pieces firing electrons" for
computer, we don't actually describe what computer does that way."
Sure, if you know how to build - aka if you have a blueprint - aka
if you have a higher-level structure. I do not think the lowest level
is informative on itself, for either human language, other aspects
of human cognition, or computers. At the lowest lowest level,
we're all just a bunch of interacting quarks. We build models of
higher levels because they actually reflect useful generalizations.
Your mileage of what's a useful generalization may vary, but just
like I speak of a processor as an abstract machine that performs
calculations not as of bunch of quarks organized into a bunch of
atoms and electrons interacting in such a way that they happen
to correspond to said calculations, I speak of a function calling
itself not of a line of 1s and 0s that encodes a piece of memory
that... you get the idea.
"A map is not the territory, but you cannot put the territory into
your pants' pocket".
REPLY (1) SHARE
Continue Thread →
Why is it that so many critics of Everett and defenders of Chomsky that I see
don't simply critique Everett, but call him names and say nasty things? You do
that here when you call him a "buffoon." Whatever the merits of either position,
I find it really unappealing and off-putting and it makes me not like people who
defend Chomsky. At least from the outside, it gives me the vibe that people
who defend Chomsky are dogmatic and insecure, that raises a big red flag for
me and probably for others that make us more sympathetic to Everett than we
might otherwise be.
REPLY (1) SHARE
:
Dmitrii Zelenskii Jul 19
Another thread here goes into detail on this. In short, it's because Everett
(unlike, say, Kenneth Hale with the non-configurational suggestion or Mary
Dalrymple with Lexical-Functional Grammar) "started it first", in a sense,
by playing the big, Fomenko-style iconoclast who has seen past
Chomskyans' delusion. (All the more surprising given that in personal -
well, Facebook - communication he doesn't seem an unpleasant person
you'd expect from this characterization.) Think biologists that have to
respond to creationists. Could one, in principle, be a bigger person and
respond the way you respond to normal scientific disagreement? Sure. Is it
the least bit surprising that this isn't what actually happens? No.
REPLY (1) SHARE
What in particular did he do that you see as "starting it"? You say
"playing the big, Fomenko-style iconoclast." What does that involve?
REPLY (1) SHARE
Read: "Every one of you who hasn't done what I did, living with
the group for many years, is a charlatan who doesn't understand
the language they study".
Once again but with passion: "You, yes you who wrote your
thesis on binding violations or person-case constraint or
whatever, are fundamentally untrustworthy".
How did the review put it? "This is the closest you come to
insulting someone in a published paper"?
:
A bit later:
Continue Thread →
Arbituram Jul 20
I was not aware we had that level of knowledge or confidence regarding the
structure of dolphin communication; could you please point me in the correct
direction to learn more?
REPLY (1) SHARE
Well, level of confidence varies, of course, but the researcher to cite would
be James T. Fulton.
REPLY SHARE
Schneeaffe Jul 24
Hey, Ive read a lot of your comments on this article and wanted to collect my
thoughts into a single response.
First, I think a lot of the clashes that you have here are due to different
priorities. While I dont know too much about linguistics, Ive seen those
grammar tree diagramms, and they seem like a reasonable thing for linguists to
do, and like something that could produce real insight about "grammar". But
Chomskyanism also includes a lot of general claims, and often the defense of
these looks like a continuous retreat from falsifiability. For someone not
interested in getting deeper into linguistics, these general claims with potential
takeaways for elsewhere will be the main thing they care about, and they will
give a lot more weight to arguments against them from other fields.
I think your examples of Everett "starting it" by being an iconoclast are pretty
weak. The idea that there could be more relevant evidence than a papers worth
of analysis examples, and that the chomskyian "reanalysis from a distance"
only makes sense in terms of fitting them into their scheme and not in terms of
what makes sense for studying that language - seems like the sort of thing that
is generally accepted without outrage in other sciences, and I suspect it would
have gone over much better in this case if Everett wasnt the sole authority on
Piraha.
Why do you think homonymies and synonymies are evidence that language is
for thinking more than communication? I agree that these are suboptimal for
:
communication, but arent they even worse for thinking? If you have two
concepts that are homonymous in thought, you just... wont be able to
Expand full comment
distinguish them, period. And if you have a synonymie, you can believe things
REPLY (1) SHARE
> First, I think a lot of the clashes that you have here are due to different
priorities.
Absolutely.
> often the defense of these looks like a continuous retreat from
falsifiability
> I suspect it would have gone over much better in this case if Everett
wasnt the sole authority on Piraha.
Very much so. (My bet is that we wouldn't be having this _discussion_ if
he weren't - someone else would just come out and show their own work.)
> I agree that these are suboptimal for communication, but arent they
even worse for thinking?
Schneeaffe Jul 24
Even if we took the argument at face value, I dont think this works as
an excuse, because it doesnt explain why the issue would appear in
linguistics more so than other sciences. In fact, the arguments apply
to linguistics less than to physics, and theres not a general
impression that physicist are slippery.
As for the argument about things like precision and accuracy and
similar — that is a reasonable point, but... There's a difference
between homonymy (two concepts with the same reflection, in
each specific case only one is meant) and vagueness (a
reflection that fails to specify details that could distinguish two
related concepts). Vagueness is inevitable, homonymy is not. It
doesn't invalidate your point but mitigates it. (Example of
vagueness in speech — Russian и fails to specify difference
between BIT vowel and BEET vowel, so the actual pronunciations
are across the whole continuum.)
REPLY (1) SHARE
:
Schneeaffe Jul 24
Continue Thread →
Schneeaffe Jul 24
Continue Thread →
Turtle Jul 19
He also never admitted he was wrong - a diehard “true communism has never been
tried,” he doubled down on the Venezuela situation, saying “Capitalists were free to
undermine the economy in all sorts of ways, like massive export of capital.”
(As per Chomsky, if you have enough resources to flee a failing state, the fact that
it’s failing is now your fault.)
REPLY (2) SHARE
ContemplativeMood Jul 19
This is a clear ad hominem. You wouldn’t infer from Henry Ford’s adoration of
fascism that he made crappy cars. The two arenas are completely unrelated.
REPLY (8) SHARE
LearnsHebrewHatesIP Jul 19
Emphasis mine.
REPLY (1) SHARE
:
ContemplativeMood Jul 19
Well if we’re using caveats then let me just add my own, that if you
***were*** to use Chomsky’s political views to disparage his
academic record ***that would be*** a baseless ad hominem attack.
REPLY (2) SHARE
Ryan L Jul 19
ContemplativeMood Jul 19
Continue Thread →
TGGP Jul 19
Continue Thread →
Turtle Jul 20
Nah, I agree with this. Heisenberg was literally a Nazi, but that
doesn’t invalidate the uncertainty principle.
REPLY (1) SHARE
Seconded!
REPLY SHARE
Not trying to argue at all, simple honest question: does Chomsky ever talk
about how his understanding of linguistics underpins/justifies his political
views? I am by no means an expert on the guy, but I do remember reading
some of his stuff where he makes the connection. (Could be wrong!)
:
REPLY (3) SHARE
You're welcome!
REPLY SHARE
Haven't read the entire corpus of his works, but I don't think so. There
is an often-made joke that Chomsky the linguist and Chomsky the
anarchist are two different people, and in Russian it is reinforced by
the fact that his surname was traditionally translated as Хомсĸий
(with a [x] sound and declinable) for linguistic works but as Чомсĸи
(with the same-ish sound as in English and usually indeclinable) for
political works.
In his debate with Foucault this comes out quite a bit. Universal
grammar suggests a universal human need for things like creativity
which feed into a more university moral and political stance. The
debate is cool. You can watch it on YouTube or read the transcript
online.
:
REPLY SHARE
TGGP Jul 19
TGGP Jul 19
JamesLeng Jul 20
https://en.wikipedia.org/wiki/Henry_Ford#Honors_and_recognition
> In 1938, Ford was awarded Nazi Germany's Grand Cross of the
German Eagle, a medal given to foreigners sympathetic to Nazism.
[145]
REPLY (1) SHARE
TGGP Jul 21
That seems more like evidence of them liking him (as they did
Thomas J. Watson of IBM) than of his "adoration of fascism".
REPLY (1) SHARE
TGGP Jul 22
Continue Thread →
The analogy doesn't work because cars are a physical product that can be
personally tested and used unlike his ideas and theories.
ContemplativeMood Jul 20
anon123 Jul 20
Turtle Jul 20
I was just uncritically accepting the thesis of the reviewer that Chomsky
the linguist is a) overrated and b) dogmatic.
REPLY SHARE
Cars are the kinds of things that either run or don't, as are assembly lines.
Theories of linguistics, very much not.
:
Further, imagine if Chomsky had been a die-hard fascist. Think anyone
would care about his linguistic theories today?
REPLY (1) SHARE
anon123 Jul 20
Melvin Jul 20
TGGP Jul 19
Typo:
> He does explicitly claim (in the aforementioned paper and elsewhere) that Pirahã
probably has no longest sentence
luciaphile Jul 21
I thought maybe he meant, all Piraha sentences are roughly equal in length.
REPLY SHARE
I am wondering why the question of whether Piraha has a longest sentence got so
much attention. English has a longest word, German doesnt. Piraha has a longest
sentence (say), English doesnt. What should matter is that no language has a
longest story. The focus on sentences seems arbitrary to me.
REPLY (1) SHARE
The claim is that sentence is the biggest object built by grammar, while it is not
even clear what a word is, that's a hot debate: for instance, if you claim that
"child seat" is a word that just happens to be written with a space (or, vice
versa, that Kinderstuhl is two words that just happen to be written together),
English suddenly becomes a lot like German.
The example of words was just meant for illustration, apologies. My point
was:
One can easily extend any grammar for sentences to a grammar for
stories. The latter grammar will always be recursive. So what is the point in
claiming that the grammar for Pidaha sentences is not recursive when the
more important grammar for Pidaha stories is recursive?
:
REPLY (1) SHARE
StrangeBanana Jul 19
"To Everett, the most important feature of language is not grammar or any
particular properties of grammar, but the fact that it involves communication using
symbols. "
"Can’t lots of animals do all those things? Yes. Everett views the difference as one
of degree, not necessarily of quality. "
Before today, I had never heard of all this tribal language beef. Nevertheless (being
an inveterate prescriptivist, though untrained in formal linguistics), I arrived at
identical versions of the above conclusions independently some years ago.
I'm still not sure whether this points to a sign of weakness in Everett's argument,
but I'm calling it a small win.
REPLY SHARE
I can't help feeling that being somewhat more open to Aristotelian 'folk'
metaphysics, purely pragmatically, might take the edge off a lot of academic
rancour. The human mind clearly has a great potential for language, which is
nevertheless realised in act only in certain contexts. I'm sure you all have excellent
jokes about bowling balls tied together with string but I'm looking here at an
interminable debate about what "innate" means, and I do think at least having the
vocabulary to talk about "non-being which in some way is" is a strength of
Aristotelianism (and Thomism). It's not sufficient but it seems necessary, or just
helpful, useful. Anyway, interesting review.
REPLY SHARE
Can someone please explain why Chomsky continues to have a role in linguistics
like Marx in communism. What he says has been proved to be wrong again and
:
again yet he is still deified
REPLY (3) SHARE
TGGP Jul 19
Marx was disproven (in practice) after he died. Chomsky is still alive.
REPLY SHARE
The situation is not as black and white as there are Chomskyan linguists and
non-Chomskyan ones, although people do organise themselves institutionally
in these ways for contingent reasons. Few people agree with everything that
Chomsky has claimed but not everything Chomsky has said is false. The useful
stuff sticks around while people reapply it in new frameworks in an improved
form but his personality casts a shadow because people like drama and it gives
outsiders the wrong impression of where the field is intellectually.
REPLY SHARE
I think Newton was a lot like this as well. And in paleontology at the turn of the
20th century, Cope and Marsh were two competing poles that similarly
dominated a field by sheer force of will, developing useful theories that led to
major progress despite being fundamentally flawed in ways that were known
early on (in Newton’s case, by Leibniz).
REPLY (1) SHARE
Erythrina Jul 19
> To borrow a phrase from elsewhere, linguists are those who believe Noam
Chomsky is the rightful caliph.
Me too, kinda, I hear this crap far too often :D But there's a kernel of truth in it.
REPLY SHARE
Ran Jul 19
This review got off on the wrong foot, for me, by initially pretending that all of
linguistics has gone Chomskyan, which is the same order of mistake as pretending
that all Americans are Republican. Chomskyans do like to pretend that they're the
whole field (or at least the "theoretical core" of the field), just as politicians like to
pretend that all Americans agree with them; but the rest of us don't have to believe
it.
(In fairness, the review later implicitly backs off of this pretense, both by using
phrases like "Chomskyan linguists" and by citing various non-Chomskyan linguists;
but I don't think there was any reason for the pretense to begin with, and I think that
an uninformed reader would still come away from this review with a *massive*
overestimate of what proportion of the field is Chomskyan.)
REPLY (4) SHARE
Ran Jul 19
In the 2020 election, 74.2 million Americans voted for Donald Trump,
which is (by a large margin) more people than had ever voted for any
candidate in any U.S. Presidential election. It's far more people than ever
voted for Trudeau, Sunak, Macron, Merkel, Putin, Stalin, etc., etc., etc.
If you were given no other information besides that paragraph, you might
conclude that Trump is overwhelmingly popular and everyone supported
him.
But in fact, his main opponent, Joe Biden, won *even more* votes!
Somehow that just didn't come up before. ;-)
But either way, what the citation counts show (at best) is the size of
Chomskyan linguistics, not the proportion of linguistics that is Chomskyan.
REPLY (1) SHARE
Ran Jul 20
Sure -- and I'll gladly stipulate that there are more Chomskyan
linguists than Marxist economists -- but are you really under the
impression that most economists are Marxists??
REPLY (1) SHARE
Ran Jul 20
Continue Thread →
Chomsky has had a huge impact on linguistics in the sense that his work has
established central debates and research questions. So many linguists who are
“non / anti Chomskyan” are in some ways working within the bounds of those
questions (eg, about innateness vs empiricism, gradualism vs saltatory
development of language, communication vs cognitive function of language,
etc). Of course you’re right that the field is much broader than followers of
Chomsky, but it really is helpful to understand his paradigmatic impact to
contextualize Everett’s work, since some of what Everett argues feels like
common sense to outsiders (eg, language is “for” communication).
REPLY (1) SHARE
Ran Jul 20
anon123 Jul 20
I think that whole part was supposed to be "the hook", an amusing little bit at
the start to get readers interested. It ran too long for me though.
REPLY (1) SHARE
Ran Jul 20
:
I'm sure you're right, but I think the reviewer really does expect us to take
claims like this one at face value:
I agree. It made me worry that this was going to be written from a pro-
Chomskyan point of view that pretended anti-Chomskyans don’t exist.
REPLY SHARE
TGGP Jul 19
The first thing bedobi says in that post is "I'm sympathetic to criticisms of
Chomsky's universal grammar"
REPLY SHARE
toki pona is a minimalist constructed language, made by Sonja Lang about 20 years
ago. It consists of about 130 words and a small grammar. Despite these limitations,
it works surprisingly well for communication, and there are friendly, active online-
communities.
toki pona seems similar to Pirahã in that it has very little recursion. Many things are
broken down in parts. For example, a sentence like "Alice said that the cat died"
would be translated as "jan Ali li toki e ni: soweli li moli" -- literally "person Ali says
this: mammal die".
Interestingly, toki pona *is* an invention. One that has sparked a unique culture,
complete with meetings, music videos, penpals, literature. The language has
:
attracted many people who are neurodiverse or part of the LGBTQ+ culture, maybe
because toki pona comes without some of the cultural baggage of older languages.
In this case, one can truly say that language is the handmaiden of culture.
REPLY (1) SHARE
luciaphile Jul 21
:)
Venkat Jul 19
I'm not a .linguist, but having met linguists, I get the impression that Chomsky is not
universally well-regarded, despite his massive citation count. Sp this review may be
overestimating Chomsky's hold on the field.
It strokes me that there are weaker claims than full Choimskiansism that re still
pretty interesting:
- pretty much all humans can learn all human languages, provided they start early
and put in effort. (language learning easier in childhood). The specifics of
particulars langauages seem noit to be genetically hard-wited at all.
- it sems that this group of people can learn languages like Portuguese with
recursive structures even f (arguably) tgheir own language doesn't use them. So
something like the potential to understand recursive langauges is bioliogically
embedded in all humans
:
- on the other hand, many non-human animals cant learn our languages, So its
clearly not trivial.
REPLY SHARE
I agree that the FOXP2 mutation is not particularly related to language—I think it
causes a general deficit in learned rapid precise motor control, and as it happens,
speech is the area most saliently impacted by a general deficit in learned rapid
precise motor control. See my blog post here for details including two textbooks
that come to the same conclusion:
https://www.lesswrong.com/posts/smgwRt5MbY9dwhecL/is-foxp2-speech-and-
language-disorder-really-foxp2-forebrain
REPLY (1) SHARE
I got the impression that FOXP2 was one' of the steps on the way to language
... conveys the ability to make a wider ranger of squeaky noises (in mice)
without grammar. Presumably, in humans, some other adaptation built upon it.
REPLY SHARE
Chomsky consistently repeats the claim that his linguistics and his politics are
utterly separate. I disagree. The through-line is in Chomsky’s implicit belief that the
Good Life is one of mass democratic decision-making. Anarcho-syndicalism, a kind
of radical egalitarian belief that all humans are created equal, would require that all
people/workers/citizens be informed about and actively participate in the collective
decision-making of their workplace and society. Basically, everyone needs to
become middle-class, literate, and non-violent.
Theories of Universal Grammar also posit that all humans are created equal. But, by
centering grammar as the core of language, they privilege the relationship that
educated, articulate people have with language. Grammar just isn’t that important
to the working class and the underclass, people who didn’t get their grammar
corrected in school until they’ve learned to follow the rules.
Now, I know that Chomsky means something peculiar when he uses “grammar”:
Deep Structure. But what evidence do we have for deep structure? Similar people
:
probably have similar mental representations. But beyond that, can we really say
with confidence that words and sentences encode meaning? No, meaning is
pragmatic, and depends heavily on contextual information. The position that words
have precise meanings, that we can literally evaluate sentences compositionally, is
itself a social construct, and in particular, a hallmark of academic and quasi-
academic discourse, such as this discussion.
eldomtom2 Jul 19
The book Decoding Chomsky may of interest as it argues the opposite; that
Chomsky's linguistic theories serve the interests of the American
establishment.
REPLY (2) SHARE
I contend that Chomsky’s leftism also serves the interest of the American
establishment. It’s an intellectual stance that is of no threat to the ruling
class; a particularly erudite form of champagne socialism/anti-imperialism,
inaccessible to the masses required for a real revolutionary movement.
Chomsky is of course more rationalistic than the critical theorists /
continental philosophers he called obscurantist, eg Foucault and other
pomo theorists that paved the way for 21st century identity politics and
intersectionality — though they too are of course part of the Western elite
discourse, and no threat to the world order; indeed, today, LGBTQ+ rights
and racial justice are a justification and tool of Western power.
For all the criticisms of mass media as a means of elite control over
democratic processes, we haven't yet found a better means of social
:
coordination. The chaotic populism of social media — rampant
misinformation and conspiracy theories, declining trust in authoritative
institutions, a race to the bottom attention economy, incentivizing
everyone to become an influencer — seems to be in many ways worse
Expand full comment
than old-school manufactured consent.
REPLY SHARE
So yes, thank you for the recommendation. I just got the book.
REPLY SHARE
I think it is important to recognize that the way both Chomsky and Everett use
the word grammar is quite different from the way laypeople use the word
grammar. The kind of grammar that linguists are talking about is simply the
organizational structure of speech. This means that Day Labourism professor
is having exactly the same grammar, and it is equally important to all of them.
Conscious knowledge of grammar that is learned in school is not talking about,
it is certainly true that that kind of explicit knowledge is not terribly important
to the average man in the street, but the average man in the street speaks
coherent structure language, which is no different from that of the professor,
perhaps for the number of fancy words being used.that is point on which
Chomsky and Everett would agree
REPLY (1) SHARE
“Grammar just isn’t that important to the working class and the underclass”
That said, I agree with a lot of the other criticism, that people aren’t using
language for the same kind of abstract thinking that Chomsky thinks it is
essentially for.
REPLY (2) SHARE
I'm not saying that working- and underclass people don't have any
linguistic regularity. That would be preposterous. Certainly, how you
communicate signals your affinity with the group, your insider/outsider
status, etc., but this is as much or more a matter of pragmatics and
sociolinguistics as it is about syntax.
Victualis Jul 25
Tony Jul 19
This was one of the 10s I gave in the book review primary season, and also my pick
for secret undercover Scott review
REPLY (1) SHARE
Bldysabba Jul 20
>A more parsimonious speculation is that they did not, and that the Great Leap was
effectively instantaneous, in a single individual, who was instantly endowed with
intellectual capacities far superior to those of others, transmitted to offspring and
coming to predominate.
Isn't this basically what Catholics believe about Adam? That all modern humankind
is descended from a single male (and a single female) ancestor, and that God
granted Adam the gift of rationality, a mind made in his own image?
REPLY SHARE
"He does explicitly claim (in the aforementioned paper and elsewhere) that Pirahã
probably has no longest sentence, which is about the most generic anti-recursion
statement one can make."
Was this meant to say that it probably DOES have a longest sentence? From what I
read up to this point I'd have thought that all languages having no longest sentence
is the standard Chomskyian view.
REPLY (2) SHARE
Zamua Jul 19
Ah I misread this! I thought it was saying he was conceding the one most
generic *pro*-recursion point, that it doesn’t have a longest sentence, even
though it doesn’t use any of the standard recursive methods for extending
them.
REPLY SHARE
taking this account at face value, it seems like it would be *extremely valuable* to
the field of linguistics to have one more person actually interact with the Piraha and
learn their language, which would presumably dissolve the questions about
Everett's description of their language. I realize that getting in slow-burning
asynchronous fights is more fun than that but it's kind of embarrassing to read
:
about as an outsider
REPLY (1) SHARE
That would be cool. This comes at a high cost though, since there are under
400 native speakers and they live in a very remote place.
People are studying recorded material though, and doing smaller experiments
(which do not involve spending years to learn the language).
https://en.m.wikipedia.org/wiki/Pirah%C3%A3_language
REPLY SHARE
Anon_NLP_Enthusiast Jul 19
Your note about parse trees in the "mechanism" of LLM's is no hypothetical - such
trees have actually been found, as early as 2019 (in BERT, a precursor to GPT)! This
work finds that the trees are linearly embedded in a certain fixed subspace of the
latent vectors. It's powerful evidence that such an algorithm emerged with no
explicit programming, as simply a necessary mechanism to predict the statistics of
language.
https://nlp.stanford.edu/~johnhew/structural-probe.html
REPLY SHARE
Excellent review! I'm gonna go out on a limb and say that it's Scott.
And call this a non-central objection if you must, but I'd like to respond to the joke
in the link in 4th footnote. Friedman would have won the exchange if he'd simply
used a binary search: ln(2^n) = n so he still would have been able to pin Yessenin-
Volpin down in linear time.
REPLY (3) SHARE
Fintan Jul 20
I took the joke to connect more to the finitist (even ultrafinitist) view of
:
mathematical objects, which is that only objects that can actually be calculated
(that is, in a finite time) can exist. This means that the larger the number, the
longer it takes to find out if it exists (hence the increasing wait). Greg Egan has
a great short story called “luminous” that uses this idea.
REPLY (2) SHARE
Oh I get it - it's very clever! But Friedman missed the chance to out-clever
him. Just sayin'.
REPLY SHARE
>Greg Egan has a great short story called “luminous” that uses this idea.
> I find this notion of recursion completely insane, because it would imply (i) any
language with more than one word in its sentences has recursion, and that (ii) all
sentences are necessarily constructed recursively.
I had to ctrl+f 'Goptjaam' and 'glossable' to see if anyone beat me to this, but it
looks like no. Granted, Goptjaam is a *deliberately* cursed conlang
(https://www.youtube.com/watch?v=ze5i_e_ryTk), but I think it's still an interesting
example of how this statement needn't necessarily follow! Goptjaam isn't glossable
:
- its words depend entirely on the context of the full sentence. The moment you put
the 'same' word into a different context, it's a different word.
To be clear, this isn't meant as an argument; actual human languages aren't like
this! I just find it fun to look at how one can break languages if one tries.
REPLY SHARE
I think Wittgenstein never thought language was inherent, but he tried to invent a
logically consistent language to make truth statements about the world (an "I-
Language")
He then realised he missed the point, language is something that is practised - and
the practise improves it over time rather than one perfect design
Maybe, but I don't think the distinction between how it's used and how it's
constructed is the relevant difference
The "E-language" [Everett] part seems to me saying that "how it's used is
how it is constructed" while Chomsky [I-Language] says "mental
properties construct how it's used"
REPLY (1) SHARE
There's clearly some interaction but I think syntax and semantics are
distinct. Building the car vs driving the car would be the rough
analogy I'd use.
REPLY SHARE
:
Kenny Easwaran Jul 21
smilerz Jul 20
Great review. Linguistics is such an interesting topic area I appreciate the work that
went into creating a layman's summary.
On a side note: can we create a corollary to Godwin's law stating that the longer an
argument goes on the probability of calling someone a racist approaches 1? And
that anyone that does so automatically loses the argument?
REPLY (3) SHARE
Vampyricon Jul 20
No because there are arguments where calling someone else a racist is correct
and pertinent.
REPLY (1) SHARE
smilerz Jul 20
Xpym Jul 20
Calling someone a racist isn't an argument, it's a status claim. By doing that
you assert that he's a Bad Person, and you're entitled to make that judgement.
This maneuver succeeds or fails mainly based on your relative social standings.
REPLY (2) SHARE
smilerz Jul 20
Xpym Jul 20
Benjamin Jul 20
Some other people have pointed out parts of this, but it's bizarre how this book
review leads in by pretending that non-Chomskyan linguistics doesn't exist. Firstly,
his influence was a lot less in many subdisciplines of linguistics
(phonetics/phonology, historical linguistics, etc.) than in his core
syntax/morphology/child language acquisition specialties. Secondly, even in the
disciplines where Chomsky has had more to say, there's always been a sizable
number of linguists opposing him (see Randy Allen Harris's The Language Wars).
So yeah, Everettian ideas are also building on an established tradition (as should be
:
clear by the many references to pre-Everett field linguistics)
REPLY (1) SHARE
Yes, this seems right and important. But I also think it’s interesting to me to see
an Everettian anti-Chomskyan take, because I’ve come to this from a different
angle.
REPLY SHARE
FrancoVS Jul 20
> He does explicitly claim ( . . . ) that Pirahã probably has *no* longest sentence ( . .
.)
There are a bunch of cryptographers and linguists who are still trying to decipher
the Voynch manuscript. A pattern observed in that community....
a) Someone will point out a statistical feature of the Voynich manuscript which they
claim never occurs in any human language ever. This, they claim, is conclusive
evidence that the Voynich manuscript is a meaningless forgery, not a simple cypher
of some real language.(*)
b) Someone else will point out an example of an absolutely definitely real language
that also exhibits the feature in question. Those guys have samples of lots of
languages in machine-processable format.
(*) Historical context would suggest that if its a real language, its got to be either a
real language known to Europeans around the time of Rudold II of Bavaria, or a
conlang invented by them. This really cuts down the space of likely possibilities.
Had the Voynich manuscript been discovered much later, like in say 1960, after the
widespread use of machine ciphers and with he existence of widely travelled
linguists, our priors for "what kind of cipher of which natural language might this
be" would be so, so much worse.
REPLY (2) SHARE
Exotic possibilities for the Voynich manuscript include: its a simple cipher
padded with nulls, i.e. most of the ciphertext is random junk padding, but
subset of it deciphers to something cohetrent.
REPLY SHARE
mst Jul 24
> (*) Historical context would suggest that if its a real language, its got to be
either a real language known to Europeans around the time of Rudold II of
Bavaria, or a conlang invented by them. This really cuts down the space of
likely possibilities.
How much does that really cut it down? It would rule out American and Oceanic
languages sure, but isn't all of Afro-Eurasia still plausible in principle?
REPLY SHARE
Just want to note that I really enjoyed Kingdom of Speech and found it highly
rewarding to read, though I had a suspicion it started after Wolfe and Chomsky met
each other at a NYC cocktail party and intensely despised each other.
"…well, okay, maybe I don’t entirely get Foucault’s number. Every humanities person
must have an altar of him by their bedside or something. "
delesley Jul 20
It is worth pointing out that the standard transformer architecture used in modern
LLMs is not recursive (or recurrent). That is to say, an LLM cannot construct a
grammatically correct sequence of tokens that corresponds (parses) to an abstract
syntax tree of unlimited depth. This is obvious on theoretical grounds from the
basic architecture, but you can also demonstrate it quite trivially by training a
transformer on toy languages described by context-free grammars. The maximum
:
depth of recursion is limited by the number of layers in the transformer.
There are some variants (i.e. feedback transformer, universal transformer) that are
fully recursive. We don't use those variants because in practice it doesn't matter.
Humans can't process grammatical structures of unlimited depth either; once the
complexity of a sentence starts to overflow working memory, the grammatical
structure gets lost. So the difference between English and Pirahã is merely one of
degree -- Pirahã has a maximum recursive depth of 1, while English has unlimited
depth in theory, but in practice has a maximum recursive depth that is slightly
higher, but still small (single digits).
Note that humans can still write code of unlimited length, even though the abstract
syntax trees for human-written programs have finite (and typically small) depth.
The trick is to do exactly what Everett did when translating the Bible to Pirahã --
break big expressions into smaller ones, and use variables (names/pronouns)
instead of nested clauses. An extreme example is assembly language, where each
expression (sentence) has exactly 3 "words" -- e.g. (operation register1 register2).
Recursive
Expand grammar
full comment is unnecessary to represent complex ideas.
REPLY (2) SHARE
B_Epstein Jul 20
It’s also worth noting that half of the attempts to extend the context
transformers may have run into the quadratic cost (complexity, memory) issue,
and that probably the majority of proposed solutions to this problem are de-
facto injecting recurrence, one way or another.
REPLY SHARE
I have a vague memory that there's a field called psycholinguistics that studies
things like...
although you might write down a grammar of English that sugges6s there are
arbitrarily long valid sentences, there is in practise a limit to what actual human
speakers will put up with, Experimentally ,measure how much stuff there can be on
:
the parser stack before a human decides the sentence is unreasonable.
REPLY SHARE
Omer Jul 20
I strongly disliked this review. It is incredibly condescending and equally wrong (and
often "not even wrong"). It honestly strikes me as odd that the reviewer, a self-
admitted layman, believe they can so easily identify the supposed absurd and
"insane" mistakes of thousands of brilliant researchers. The alternative hypothesis,
that the reviewer might be misunderstanding something, should've a at least
crossed their mind. This has all the red flags typical of a crank (possibly due to the
source material, but still...).
Like the reviewer, I'm not a linguist, but even I could spot such misunderstandings
in this review. These include conceptual misunderstandings (such as the
misconception about the meaning and implications of "recursion"), contextual
misunderstandings (e.g. regarding the long history of structuralism, from De
Saussure onward), and disciplinary misunderstandings (there are many counter-
Chomskyan linguists and even entire such linguistics departments; none subscribe
to Everett's views to the best of my knowledge, for what seem like good reasons).
Believing all the linguists are dense and mindless cult-followers lacking critical
thinking is just silly.
REPLY (2) SHARE
This review correctly mentions Foucault being extremely cited, and this very
blog has a post claiming that Foucault was basically a liar!
https://slatestarcodex.com/2018/01/04/book-review-madness-and-civilization/
REPLY SHARE
I feel these interested amateur reviews are high risk high return. Sometimes
you read something that shares a glimpse into another world and has an
infectious feeling of excitement - and sometimes you get a smart person totally
misunderstanding a really complex topic and leaving you feeling vaguely
annoyed.
REPLY SHARE
/\\//\\//\ Jul 20
Any idea what percentage of the Chomsky citations are to his political polemics?
Especially among grad students, his political writings are, because of their volume
and the predictability of his opinions, an ‘easy cite’ for students looking to frame a
debate.
REPLY (1) SHARE
I would expect that even his linguistic articles have _huge_ citation numbers,
because half the articles include "Following the general premise of Minimalism
(Chomsky, 2000)" or some such intro, which artificially inflates quotations
even if the rest of the article doesn't directly engage with Chomsky's work at
all.
REPLY SHARE
I thought it was pretty clear that the opening paragraphs of the review, which
purport to describe Chomsky's quasi-divine status, were tongue-in-cheek? -
anyway, I very much enjoyed this review - I have only read Wolfe's book, but now I
want to read Everett's own accounts.
REPLY SHARE
Turtle Jul 20
I thought this was a great review, and I wanted to share another interesting thing
about the Piraha. They are really happy. Relentlessly happy. In the 30 years Everett
studied them, there were zero suicides. The idea was a completely foreign concept.
Here’s Everett:
“I was still a very fervent Christian missionary, and I wanted to tell them how God
had changed my life. So, I told them a story about my stepmother and how she had
committed suicide because she was so depressed and so lost, for the word
depressed I used the word sad. So she was very sad. She was crying. She felt lost.
:
And she shot herself in the head, and she died. And this had a large spiritual impact
on me, and I later became a missionary and came to the Piraha because of all of this
experience triggered by her suicide. And I told this story as tenderly as I could, and
tried to communicate that it had a huge impact on me. And when I was finished
everyone burst out laughing.”
They burst out laughing! It just didn’t make any sense to them that someone would
even consider doing that.
Does this have to do with their language? Well, language and culture are deeply
intertwined, and we know that the Piraha only use language centred on the present.
They have very limited vocabulary to describe the past and the future. So it
suggests a culture that’s deeply focussed on the present. In Western culture we
sometimes say that depression is living in the past, anxiety is living in the future,
and getting someone out of their head and into the real world (the now) is the point
of therapy
Expand and mindfulness
full comment training.
REPLY (1) SHARE
drosophilist Jul 21
The Piraha may be very happy, but if their reaction to being told “this person
was very sad and then she killed herself” is to burst out laughing, then I would
describe them as somewhere between “zero emotional intelligence” and
“stone cold sociopaths.” Not great role models!
REPLY (1) SHARE
Turtle Jul 21
I picture it like - imagine someone told you “this person was so sad that
she ate a thousand bananas until her stomach exploded.” You might laugh,
right? Just cause the concept is so ridiculous and unheard of, not because
you don’t have empathy for people with exploding stomachs.
REPLY SHARE
I love these ACX book reviews, and will read this one in its entirety later, but is it too
much to ask them to be a little more brief? With the length of this one I feel it’s
probably easier just to read the book itself.
REPLY (2) SHARE
B_Epstein Jul 20
:
The fact is that many (most?) winners/ front-runners have been very long,
historically. So the revealed preference is for longer rather than shorter
reviews.
Of course we’re all selected for liking Scott’s writing here- and it leans towards
thoroughness at the expense of brevity.
REPLY SHARE
I was somewhat thinking this, but also thinking that it could have used a longer
section actually talking about the book being reviewed!
REPLY SHARE
Sniffnoy Jul 20
Have to point out with the other commenters that the idea that linguistics is
primarily Chomskyan is not true.
REPLY SHARE
static Jul 20
The thing that I find striking about Chomsky is the claim that nothing other than
grammatical analysis really matters. From a philosophy of language perspective, I
am much more sympathetic to the Wittgensteinian view, as expanded by Kripke.
Chomsky in the end seems quite reductionist, mistaking his understanding of how
there can be formal systems that describe how languages function via grammar to
mean that other aspects shouldn't be studied. His project reminds me of
philosophical work trying to derive the basics of arithmetic from the rules of logic,
and then declaring that nothing else you could do with that was worthwhile. Even
without that, there were plenty of things to do in mathematics- and different logical
foundations are possible for the empirical facts that can be observed. If the
Chomsky project ends up just describing a very simple rule for recursion, it may not
be all that useful as a foundation for study of more complex phenomena, as the
complexity that must be built on top of that foundation in order to describe common
uses of language makes those descriptions less useful.
REPLY (1) SHARE
static Jul 22
Garald Jul 20
" Perhaps the most interesting suffixes, however (though these are not unique to
Pirahã), are what linguists call evidentials, elements that represent the speaker’s
evaluation of his or her knowledge of what he or she is saying. There are three of
these in Pirahã: hearsay, observation, and deduction."
This sort of thing is not so unusual - for instance, Quechua has them. (A page in a
grammar book on my bookshelf lists the situations in which one particular evidential
particle can be used - some examples from memory: I was there but I was asleep, I
was there but I was very drunk, I was there but I was less than five years old...)
REPLY SHARE
This review is bizarre. Chomsky doesn't have a stranglehold on linguistics. It's like
writing a long essay criticizing Aristotle's stranglehold on physics.
REPLY (2) SHARE
:
Kenny Easwaran Jul 21
While it’s overstated, I think it’s right. There are chomskyan departments and
anti-chomskyan departments, but everyone has opinions on Chomsky. There is
no comparable figure in physics.
REPLY SHARE
I totally agree!! After the first few paragraphs I was seriously confused about
what this guy had actually read. It just seems the opposite of any recent “mass
appeal” linguistics book.
REPLY SHARE
> He does explicitly claim (in the aforementioned paper and elsewhere) that Pirahã
probably has no longest sentence, which is about the most generic anti-recursion
statement one can make.
Ppau Jul 20
Okay next time someone named Everett proposes new fundamental ideas about a
field I'll be listening from the beginning
REPLY SHARE
Did Chomsky really declare that the first language evolved abruptly a mere 50,000
years ago? If that is a prediction of his theories then its epic stupidity largely
discredits them on its own. Anatomically modern humans appeared over 300,000
years ago for a start, and it seems misguided, verging on cranky, to dispute that
throughout that time they were fully capable of speech with a proper language,
even if this may have lacked some modern refinements.
I mean one can teach chimps and gorillas to speak, at least simple sentences. They
may grunt and huff and puff in trying to imitate human vocalizations, but when they
hold up cards saying things like "Koco want banana" and suchlike, there is no doubt
they are meaningfully communicating, and is cast iron evidence they are thinking in
words up to a point. Even my cat clearly recognises a few words, such as her name
:
and "din dins"!
REPLY (3) SHARE
Omer Jul 20
If you assume that being anatomically modern means having a fully developed
language ability, then you're making a circular argument (and ironically, you're
kinda agreeing with Chomsky while doing it, though with respect to a slightly
different of definition of "anatomically modern").
Also, writing appeared abruptly about 5,000 years ago, even though humans
300,000 years ago were anatomically capable of writing just like us. This
shows that your reasoning is flawed, and it's reasonable to suggest a similar
trajectory for speech. In fact, this makes more sense for speech: there is
evidence (like types of aphasia) indicating that speech, unlike writing, isn't
genetically stable and can break down isolatedly. While not definitive proof, this
supports the hypothesis of an abrupt development of speech.
Finally, archaeological evidence (as far as I know) shows that about 50,000
years ago, there was a significant cultural leap in art and technology that
spread quickly worldwide. The fact that there were no major anatomical
changes during this time, as you mentioned, actually supports the idea that
language suddenly emerged rather than contradicting it.
REPLY SHARE
Not to pick on you but there is a lot of confident incorrectness about linguistic
theory because we are prone to assuming that being able to speak language is
a good qualification to start theorising about it but it isn't. Things like "word"
and "sentence" have intuitive meanings to lay people but they are pre-
theoretical terms. When you put in the work to analyse their role in a larger
explanatory model (such as those starting with Eric Lenneberg and Charles
Hockett, who made contributions to the biology of language long before
Chomsky), you realise that these things can be counter-intuitive and you can
reasonably come away with the conclusion that non-humans communicate
with things we might not want to call "words".
I mean, just open journals like Linguistic Inquiry, Syntax, Glossa or Natural
Language and Linguistic Theory etc. - that should have been foundational to a
piece which ultimately disparages the field alongside Chomsky but there is no
evidence in those pages of what the review describes because the Everett
drama is totally unrepresentative.
I think underlying all this, there's really a badly misunderstood model that
people have for how academia works. Perhaps biased by reasonable critiques
Expand full comment
of the postmodernised humanities but that brush can't be used to tarnish
REPLY SHARE
That's not what evidence from teaching apes language actually suggests, and
anatomically modern humans that seemingly didn't engage in modern
behaviour are an argument in the other direction, that is, that they were _not_
capable; I've already touched on both in other comments.
REPLY SHARE
avalancheGenesis Jul 20
Must admit to skimming the last two sections. Three parts preface, one part actual
review of the book in question, one part kludged-on red meat for rats (ratmeat?),
with the former roughly 65% of the entire review, and cornucopious blockquotes
throughout (so many blockquotes, mostly from...not-actually-the-book-being-
reviewed)...it's just A Lot. Yet despite all the references and scene-setting, I still find
myself with that nagging [Citation Needed] feeling of insufficiently-supported
statements. This was not helped by c.f.'s to other works that the author admits to
not having actually read/understood, nor ~all the quotes being from within the first
few dozen pages of each work. I want to assume the full texts were indeed read,
and their contexts explicated in prose, but...I dunno, it reminds me a lot of sundry
mandatory Works Cited bibliographies in school stuffed with trite surface-level
references. A palette of bon mots more than integral supporting research. One can
certainly stitch together an argument this way, but it shambles rather than running,
which is optimal for zombies but not rhetoric.
Ultimately I leave with no more knowledge than I started of what linguistics even is,
or why it's important outside departmental funding disputes, or why anyone takes
:
Chomsky seriously. "It's not my theory that's wrong, it's the world!" says the
dismissively towering Ivyist who's never done any fieldwork. (Also, I still don't
understand what an I-language is. The spherical cow of languages?) Despite
reading it,comment
Expand full it's Not For Me, sorry.
REPLY (1) SHARE
Without commenting on the rest, I-language is the mental module for language
in each individual's brain, which gives rise to the individual's observed
language, "idiolect".
REPLY (1) SHARE
avalancheGenesis Jul 21
This would have been helpful to know while reading, thank you. So
language is theorized to be necessarily individual and arises from a
particular <s>candidate gene</s> shared brain module, yet nevertheless
follows certain universal patterns, despite the trivially observable vast
differences between discrete..geocultural expressive groupings, or e-
languages. (And maybe - maybe - some languages that don't follow all the
axioms.) A meta-language, if you will, and an essential H. sapiens
evolutionary stepping stone leading to enhanced cognition overall. That's
at least the seed of a potentially interesting and falsifiable hypothesis. (I
wonder what Chomsky thinks about Neuralink and similar proposed
technologies...)
Should have amended that I...never actually learned proper grammar, and
wasn't even introduced to the concept of certain fundamentals like "parts
of speech" until late teens. Even today, the formalisms continue to
baffle...some language simply seems intuitively correct, and other
constructions not, without really being able to explicate why. Chomsky
obviously means more than *just* the formal Rules Of Proper Language
one learns in school when he refers to Grammarcy Hall, but that's
definitely a part of it...so not having that, uh, brain module is perhaps a
large part of why I bounce hard off linguistics. Not an impediment to
actual speech-as-communication-or-internal-thinking (I assume perfect
written/verbal SAT/WAIS measures *something* real), but probably a big
stumbling
Expand block for
full comment understanding the theoretical underpinnings.
REPLY (1) SHARE
:
Dmitrii Zelenskii Jul 21
Not really? If the rules are part of the brain, you don't need to teach
them. More generally, you can't explain how specialized modules of
the brain work by the mere fact you possess them, any more than
possessing a heart makes you a cardiologist.
REPLY (1) SHARE
avalancheGenesis Jul 21
By contrast, you could also argue for a biologically innate language ability,
hardcoded in, that nevertheless evolved gradually, through multiple levels of
selection, from a proto-language towards today through millions of years of
selection for better linguistic ability.
And, frankly, I feel that this would be a more logical alignment of the abrupt-gradual
and hardcoded-general questions than the one the review describes linguists as
going for. Hardcoded linguistic ability should emerge at the speed of evolution;
language as a teachable invention, after a "hardware overhang", can spread at the
speed full
Expand of culture.
comment
But not all evolution is gradual. If the intuition about the system is broadly
correct, the recursive property is as abrupt as switching from RNA to
RNA+DNA, and this lines up well with ontogenesis.
REPLY (1) SHARE
VNodosaurus Jul 21
That's a strange analogy, given that we don't know how the RNA-DNA
transition happened. But yes - evolutionary changes can happen quickly.
However, quick by evolutionary standards is still pretty slow by cultural
standards.
REPLY (2) SHARE
VNodosaurus Jul 21
(And as for "don't know how" — that makes the analogy more
illuminative for me: the nature of the nucleotides implies it should've
been rather abrupt, as ribonucleotides don't form stable chains (I
mean the "vertical" connection, not the complementarity-based
"horizontal" one) with disoxyribonucleotides, so it wasn't nucleotides
changing one by one in a hybrid molecule, so we can conclude
abruptness without seeing it.)
REPLY SHARE
:
luciaphile Jul 20 · edited Jul 20
Commenters have expressed on other reviews a desire for economy where book
reviews are concerned.
This was really economical to me. Kind of the difference between a NYT book
section review and a NY Review of Books review: something (the former) can be
short and yet in some cases a waste of time compared to the longer thing.
I found the dispute and more so the background info fascinating but I would never
want to read a whole book on it. This was “just right” in Goldilocks terms.
I loved reading about Chomsky and not having any of the obligatory stuff about his
politics.
And yet - by the end - it did seem to me that his undoubted brilliant scholarship -
displayed as time went on (if this review is anything to go by) a sort of unearned
totalizing tendency, rather in keeping with his natural radicalism.
REPLY SHARE
I like how the author presented acronyms in parenthesis before using them. The
Author of The Family That Could Not Sleep review did not do this, and I struggled
slightly because of it.
REPLY SHARE
Arby Jul 21
"I’m not going to rehash all of the details, but the conduct of many in the pro-
Chomsky faction is pretty shocking."
isn't that also exactly how pro-Chomsky faction behaves regarding his political
views? weird.
REPLY SHARE
vindication Jul 21
Both these people raise red flags for me. Chompsy is the archetype celebrity
scientist holding a field hostage to his ridged interpretation, and Everett is making
limited claims but writing as if he is overturning the entire field.
REPLY (2) SHARE
:
Dmitrii Zelenskii Jul 21
Everett's claims would plausibly overturn the entire generative linguistics field
if they were true, make no mistake. They suggest interdependence between
language and culture at a level incompatible with language being a module of
mind.
REPLY (1) SHARE
Arby Jul 21
I mean, both Hebrew and Arabic still use an abjad, which is literally
worse than hieroglyphics. Writing systems are sticky unless you have
Ataturk or Red Terror at hand.
I agree about the linearity issue. I was just thinking about how
easy it makes teaching kids how to read, which makes perfect
sense since it was made specifically to make literacy easy for
people without much formal education.
REPLY SHARE
:
MichaeL Roe Jul 21
This review should come with a strawman warning. I am not a linguistics academic
but enjoy books on the topic and this reviewer’s characterisation of linguistics as
just Chomsky seems completely at odds with recent books. There are multiple
established schools of thought and most recent books seem to have moved on -
but reading this you would think there was some David and Goliath struggle.
Also I felt the LLM conclusions were pretty weak - if you’re going to bring them in I’d
want to get more of a “So What?”. Concluding that LLMs show that Chomsky is
wrong seems a bit too obvious.
REPLY (2) SHARE
I was intrigued to read this book review and indeed, it seemed to be the old
perspective that is irrelevant today.
REPLY SHARE
As for Pirahã language and people, we have to take into account that this a very
small group of only 200-300 speakers who live in harsh primitive environment. It is
not unimaginable that they have experienced societal upheavals like violence from
neighbouring tribes, diseases, famine, loss of habitats etc.
Spoken languages change rather quickly and we cannot even be sure that what
Everest witnessed wasn't some kind of pidgin formed from merging several local
dialects and languages. Wikipedia gives an example that women sometimes
pronounce s as h which is an indication of a different linguistic background. You
could expect eventual development into full-fledged creole language. However, the
number of speakers are so low that now the school has been opened for them
where they learn Portugese instead of Pirahã.
REPLY (1) SHARE
mst Jul 24
- Pirahã does not typologically look like a creole at all: it has complex verbal
morphology for one thing, creoles usually lack verbal morphology altogether. It
has phonemic tone which I believe is rare in creoles. Etc.
You make good points that Pirahã language could not be pidgin or creole.
However, I think that simplified verb structure of creole is due to mixing of
dissimilar languages like English and Chinese. What if those two
languages are on dialectum continium like Russian and Ukrainian? It is
:
called Surzhyk (Ukrainian-Russian pidgin) and I don't believe it is
standardized and yet contains full complexity of verbal forms. Even the
most defining features such as a specific pronunciation of Ukrainian [g] is
undefined, some speakers might use it and other not.
In case of Pirahã it is not that women speak differently but that some
women speak differently. How is it possible for such a small group to
develop different phonetic features? We think of Pirahã people as a static
group but reality might be that they are mixing with neighbouring tribes,
possibly even with violence (like stealing wives from neighbour tribes?). I
don't know how much I can trust researchers who are religious
proselytizers. They always have their own agenda, even with Everest
losing his faith in Christianity, he still might be a person of a strong
ideology
Expand fullinstead
commentof science.
REPLY SHARE
VNodosaurus Jul 21
And so - the calm Chomskyan response to Everett saying that the Piraha language
doesn't have recursion would be "if this is true, then it's not a true language".
Eventually that's what Chomsky concluded. But if you equate language and mind,
then "the Piraha don't have a true language" is interpreted as "the Piraha are
literally subhuman". And so the response was as toxic as you'd expect from a
debate involving
Expand full commentthe word "subhuman".
REPLY SHARE
:
owlmadness Jul 21
‘John, the man that put people in the water in order to clean them for God, that lived
in a place like a beach with no trees and that yelled for people to obey God’.
Heck, I'd reject that too! Never mind the tortuous syntax, it's the esthetics that's
the deal-breaker here.
‘John cleaned people in the river. He lived in another jungle. The jungle was like a
beach. It had no trees. He yelled to people. You want God!’
I guess these short punchy statements might grow tiresome after a while. But then
again, I could also imagine getting used to them.
And meanwhile, doesn't the very fact that these sentences occur together and form
a coherent whole imply that parataxis is all you really need to coordinate a
collection of ideas? Surely no one is suggesting that Chomskyan grammar stands or
falls on the presence or absence of an explicit conjunction like 'and' as opposed to
-- in this case -- a period that's in essence fulfilling the same function?
REPLY SHARE
mbsq Jul 21
It sounds like Chomsky just did his own version of formal logic and declared it to be
empirically significant.
REPLY (1) SHARE
Thanks for your comments here, I think they perfectly complement the
impression I had from other conversations with Chomskian linguists. It's
typically toxic, in bad faith, overclaiming their (mostly theoretical)
achievements, and it lacks any sort of self-irony and applicability. Also
they constantly reference to their supreme leader more than anyone else,
who is the greatest example of this behavior. I mean, Darwin didn't refer to
his predeccessors like "mere collectors", neither did Linney, Mendeleev or
:
other global theorists. They typically expressed some gratitude and
admiration to those who 'collected stuff' for them. I think now that the
success of this approach was less 'grammarian generative' and more
'dissertation generative', because it opens an infinite field of theorizing. I
mean, you'd probably expect that good theory would somehow predict
how we learn languages, how we model them, and what languages can
and cannot exist, or how they evolve. In all of this Ch. didn't provide much.
I don't devaluate their work fully, that'd be a very low bar given probably
majority of linguists still following this paradigm. But now I think it did more
harm than
Expand good, suppressing
full comment less dogmatic research and thinking.
REPLY SHARE
"He mostly seems to take issue with the idea that some region of our brain is
specialized for language. Instead, he thinks that our ability to produce and
comprehend language is due to a mosaic of generally-useful cognitive capabilities,
like our ability to remember things for relatively long times, our ability to form and
modify habits, and our ability to reason under uncertainty."
This makes me think of Cecelia Heyes' Cognitive Gadgets: The Cultural Evolution of
Thinking (2018). She argues that many things we think of as innate, like "theory of
mind" are learned. "Theory of mind" is the idea that you think a lot like me: you have
emotions like I do, you have desires like I do, etc. On the other hand, our brain
makes it relatively easy to learn some "cognitive gadgets" and not others.
REPLY SHARE
foobar6754 Jul 22
"The intent here is not to pooh-pooh Chomsky" writes the author of a review that
transparently tries to do just that; and fails spectacularly.
The crux of the argument is that Piraha doesn't use recursion but as Chomsky has
explained many times (e.g. https://youtu.be/u6Lk79bnUbM?t=0h22m47s) what is of
interest is what human capacities are, not whether they are always expressed. In
:
the video above Chomsky makes the analogy to a tribe that wears a patch over one
eye, which doesn't refute humans having the capacity for binocular vision.
The reviewer goes on to imply that this means the theory is vacuous or unfalsifiable
("the statement that some languages in the world are arguably recursive is not a
prediction; it’s an observation, and we didn’t need the theory to make it... can we
test it by doing experiments involving real linguistic data...?") but that is not true:
linguists like Moro *have* done experiments that show that humans don't seem to
be able to process certain made-up linear-order (i.e. non-recursive) languages as
easily as made-up recursive ones and use non-language areas of the brain to do so
(see Moro, "Secrets of Words"), which suggests that humans cannot understand
arbitrary linguistic structures but can understand recursive ones, just as UG would
have predicted. Had the result been the reverse it would have falsified (or at least
severely weakened) the theory. (Note that Piraha is like a degenerate recursive
language [no center embedding], so it is still "natural" for humans). Also note that
the Moro experiment does falsify the "mosaic of generally-useful cognitive
capabilities" claim made later on (which is an actually vacuous theory as it doesn't
explain anything in any case).
Clive Jul 22
foobar6754 Jul 22
Yes (obviously?). At some point there was no light sensor, then there was.
It then evolved further.
REPLY (1) SHARE
It's not at all obvious, because I'm not sure others interpret "eye" as
inclusive of "light sensor." Can you clarify what you're claiming?
REPLY (1) SHARE
foobar6754 Jul 23
foobar6754 Jul 23
Continue Thread →
You said they can't process them "as easily," but then you say, "which
suggests that humans cannot understand arbitrary linguistic structures but can
understand recursive ones." Is it that they can't understand them, or can't
understand them as easily? Also, "people will understand languages with
recursion more readily than ones without it" strikes me as a fairly tepid
prediction that isn't exclusively consistent with UG or nativist views more
generally. I'd also have some questions about who the participants in these
studies were; if they'd already been enculturated into one set of languages
rather than another, those difficulties may have been acquired rather than
innate.
> Also note that the Moro experiment does falsify the "mosaic of generally-
useful cognitive capabilities" claim made later on (which is an actually vacuous
theory as it doesn't explain anything in any case).
Could you say a bit more about what it falsifies and how?
*Purposefully*
Expand conflating?
full comment How do you know what their intentions were?
REPLY (1) SHARE
foobar6754 Jul 23
> Is it that they can't understand them, or can't understand them as easily
> ... strikes me as a fairly tepid prediction that isn't exclusively consistent
with UG or nativist views more generally
It's not tepid at all, it provides reasonably strong evidence that language
processing requires hierarchical structures, not arbitrary ones, which is
what UG predicts and is not what "general processing" theories predict.
It's probably one of the most important recent experimental results in
linguistics.
> I'd also have some questions about who the participants in these studies
were; if they'd already been enculturated into one set of languages rather
than another, those difficulties may have been acquired rather than innate.
I have no idea what you mean. Obviously people in the experiment have
pre-existing language ability. (It's not ethical to raise a child from birth in a
fake-language experiment)
> Could you say a bit more about what it falsifies and how?
>I have no idea what you mean. Obviously people in the experiment
have pre-existing language ability. (It's not ethical to raise a child
from birth in a fake-language experiment)
But even the results as described don't seem to show that it's
*required.* Without more details I'd also have questions about why
people may have responded this way.
Okay, that makes sense, thank you. I'm still wondering why this is the
case, though.
>There are crucial instances of evolution that are not gradual but
sudden.
Expand full comment
foobar6754 Jul 23
>If they do, then could it be that it's easier for people to process
languages that share similar features to the languages they
already speak?
It wouldn't explain the fMRI data. You can certainly come up with
:
a hypothesis that is consistent with the experiments without UG
but they are just harder to believe than the "simple" explanation
that language is recursive and as far as I know don't have
corresponding supporting experiments. It's true that it isn't
irrefutable dispositive proof but that rarely exists in psychology
and psychology-adjacent fields.
> But even the results as described don't seem to show that it's
*required.*
Required for what? I can put a gun to your head and force you
talk without recursive sentence structure, what has that proven?
If you have a Turing machine and only feed it regular grammars
does that prove it is not a Turing machine? If you have a joint with
6 degrees of freedom but observe it only moving in 1 does that
prove that it doesn't have 6? If you observe a small tribe in the
Amazon that crawls on all fours throughout their lives does that
Expand full comment
prove that bipedalism is not a "required" property of humans?
REPLY (1) SHARE
Why not?
> It's true that it isn't irrefutable dispositive proof but that
rarely exists in psychology and psychology-adjacent fields.
foobar6754 Jul 23
> At least
Expand with respect
full comment to the last one, the answer
REPLY SHARE
https://www.kvetch.au/p/dont-sleep-there-are-snakes
REPLY SHARE
quiet_NaN Jul 22
>> Any expression larger than two words, then, requires recursion, regardless of
whether there is embedding in that expression. For instance the noun phrase “My
favourite book” requires two iterations of Merge, (Merge(favourite, book)=
[Favourite book], Merge(my, [favourite book])= [my [favourite book]]) and therefore
is an instance of recursion without embedding.
For crying out loud, the fact that I am able to chain instances of Merge (a glorified
concat) does not imply recursion any more than the fact that I can write (1+2)+3
means that addition is recursive.
Coming at this from computer science, that example would be possible with the
following formal grammar (in EBNF):
possessive="my"|"your"|...
adjective="favorite"|"green"|...
noun="book"|"duck"|...
By contrast, I would reserve "recursion" for cases where a production rule refers
directly or indirectly to itself:
The claim is as silly as the claim that a bike is self-similar because a bike is made-
up-of wheels and a frame, and the wheels are made-up-of a tires, spokes and a
hub. Hence to describe a bicycle, you have to use 'made-up-of' on things which are
Expand full comment
themselves 'made-up-of' things, hence self-similarity.
REPLY (1) SHARE
:
Omer Jul 22
Does it really make sense to you that an entire field is full of clueless nitwits
who’ve never heard of this exotic "formal grammar" of yours?
I honestly don't see why you (and apparently many other) are having an issue
with the term "recursion" here. As someone in computer science, you must
have heard terms like "recursive grammar" and "recursively enumerable sets"
before, right?
And by the way, also from wiki: "Addition is a function that maps two natural
numbers (two elements of N) to another one. It is defined recursively as: a+0=a
and a+S(b)=S(a+b)". I admit this is possibly a bit nitpicky, but it too suggests
that you might want to reconsider your understanding of the term "recursion"...
REPLY (1) SHARE
quiet_NaN Jul 31
> The term 'heart', when used by rhino researchers refers to the big blood
Expand full comment
vessels between the blood pumping organ (BPO) and the lungs.
REPLY (1) SHARE
Omer Jul 31
Alice K. Jul 23
Not even close to finished with the review and I am so intrigued I will absolutely be
buying this book and more. So that bodes well as a vote.
REPLY SHARE
SurvivalBias Jul 23
By far the best review I've read so far, thank you! Also finally learned something
about this Chomsky guy (seems like another case where great intelligence is used
in large part or even mostly to defeat itself).
REPLY SHARE
Fantastic post, also introduced me to a new rabbit hole on finitism and ultrafinitism.
:
REPLY SHARE
meeting with a Psychic medium and I have to say, it's really worth it. It had been
6months since my husband left me without saying anything.All thanks to Dr Adoda I
was able to connect with my husband again with his powerful reunion love spell and
now he loves me more than he used to. Eternally grateful with The Great Dr Adoda! I
highly recommend, His Website is http://dradodalovetemple.com
REPLY SHARE
Do any of you readers also start to feel that sentences such as,
> Large language models (LLMs) like ChatGPT are not perfect, but they’re getting
better all the time
are repeated too much? I mean 2 years ago, ChatGPT released i-forgot-what-
version and everyone was amazed, but from there, "the law of diminishing returns"
hit hard. To the educated eye, one can see that the model simply stopped
advancing "so greatly" and all the features they advertise nowadays are some extra
pre and postprocessing tasks rather than the capabilities of model itself along with
maybe extended memory or so.
REPLY SHARE
Chris K. N. Aug 3
Good review. I find it interesting that the Christian missionary is the one making
such a strong case for gradual evolution and natural selection, and the Marxist
professor is the one arguing some version of human exceptionalism and mysterious
endowment.
I also got confirmation of my priors, that Chomsky would be (would have been)
more helpful if he could just dial it down to about 65% and make room for a bit more
humility in most things.
REPLY SHARE
im very late; but does anyone have example of linguistic theories about non
recursive languages?
Im anti the hyper recursion of programming languages; for example, while you could
:
define the allocation of space using "binary space partition trees"(and I use
bspwm), you really really shouldnt, artists and designers subdivide 2d space with
grids for a reason.
Likewise instead of 2+3*4 instead of letting that statement equal 20 reducing the
complexity and let more parsing be left to right; everyone thinks you *need* to push
everything into a 10 step hyper abstract process; for what exactly?, you shouldnt
ever depend on ambiguous statement grammer.
REPLY SHARE