Dennett Upside Down Cake: Thinking About Language Evolution in the 21st Century

About two years ago Wintz placed a comment on Replicated Typo’s About page in which he lists several papers that make good background reading for someone new to the study of linguistic and cultural evolution. I’ve just blitzed my way through one of them, Language is a Complex Adaptive System (PDF) by Beckner et al (2009)*, and have selected some excerpts for comment.

The point of this exercise is to contrast the way things look to a young scholar starting out now with the way they would have looked to a scholar starting out back in the ancient days of the 1960s, which is when both Dennett and I started out (though he’s a few years older than I am). The obvious difference is that, for all practical purposes, there was no evolutionary study of language at the time. Historical linguistics, yes; evolutionary, no. So what I’m really contrasting is the way language looks now in view of evolutionary considerations and the way it looked back then in the wake of the so-called Chomsky revolution—which, of course, is still reverberating.**

Dennett’s thinking about cultural evolution, and memetics, is still grounded in the way things looked back then, the era of top-down, rule-based, hand-coded AI systems, also known as Good Old-Fashioned AI (GOFAI). In a recent interview he’s admitted that something was fundamentally wrong with that approach. He’s realized that individual neurons really cannot be treated as simple logical switches, but rather must be treated as quasi-autonomous sources of agency with some internal complexity. Alas, he doesn’t quite know what to do about it (I discuss this interview in Watch Out, Dan Dennett, Your Mind’s Changing Up on You!). I’m certainly not going to claim that I’ve got it figured out, I don’t. Nor am I aware of anyone that makes such a claim. But a number of us have been operating from assumptions quite different from those embodied in GOFAI and Language is a Complex Adaptive System gives a good précis of how the world looks from those different assumptions.

Usage Based Theories

p. 8:

Usage – based theories of language acquisition (Barlow & Kemmer, 2000) hold that we learn constructions while engaging in communication, through the “interpersonal communicative and cognitive processes that everywhere and always shape language” (Slobin, 1997). They have become increasingly influential in the study of child language acquisition (Goldberg, 2006; Tomasello, 2003)…. They have turned upside down the traditional generative assumptions of innate language acquisition devices, the continuity hypothesis, and top – down, rule – governed processing, replacing these with data – driven, emergent accounts of linguistic systematicities.

In this case, of course, generative alludes to Chomsky’s theorizing, which revolutionized linguistics starting in the late 1950s and 1960s. Chomsky took ideas from formal logic and applied them to the grammar of natural language. In this view the syntactic structure of a sentence is seen to be analogous to a proposition in a formal system. The derivation of that structure is analogous to a formal proof.

In Aspects of the Theory of Syntax (1965) Chomsky introduced the distinction between linguistic competence and performance. The distinction has proven to be interesting, influential, vague, and problematic. Roughly speaking, competence is about the idealized grammar as stated in a formal description. Performance is what happens in actual speech, with its imperfections, slips of the tongue, and finite memory. The last is very important as Chomsky’s formal model(s) found natural language syntax to be formally equivalent to a universal Turing machine. But that requires infinite memory, and no real computing device, nor even the human brain, has infinite memory available.

Continuing on p. 8:

Constructionist analyses chart the ways in which children’s creative linguistic ability, their language system, emerges from their analyses of the utterances in their usage history using general cognitive abilities, and from their abstraction of regularities within them. In this view, language acquisition is a sampling problem, involving the estimation of the population norms from the learner’s limited sample of experience as perceived through the constraints and affordances of their cognitive apparatus, their human embodiment, and the dynamics of social interaction.

Back in the 1960, 70, and into the 80s, computational approaches to language relied on hand-coded grammars and semantics, and prespecified acoustic analysis. Starting in the middle and late 1970s statistic techniques were developed for speech recognition where the computer learned how to map continuous speech into discrete phonemes, syllables and morphemes. Such techniques were then applied to other aspects of language processing, most notably machine translation.

Ideolects and Beyond

Skipping on to p. 12:

Language exists both in individuals (as idiolect) as well as in the community of users (as communal language). Language is emergent at these two distinctive but inter – dependent levels: an idiolect is emergent from an individual’s language use through social interactions with other individuals in the communal language, while a communal language is emergent as the result of the interaction of the idiolects. Distinction and connection between these two levels is a common feature in complex adaptive systems. Patterns at the collective level (such as bird flocks, fish schools, or economies) cannot be attributed to global coordination among individuals; the global pattern is emergent, resulting from long – term local interactions between individuals.

Now, it’s one thing to talk about flocking as an emergent behavior of individual birds. That’s relatively simple behavior compared to language behavior, which necessarily trails off into everything else through semantic linkage. Just what kind of thing is this “communal language”? Does it make any sense at all to characterize its structure and mechanisms by an idealized formal grammar of linguistic competence? Not much, I’d say, not much.

Continuing on p. 12:

Therefore, we need to identify the level of existence of a particular language phenomenon of interest. For example, language change is a phenomenon observable at the communal level; the mechanisms driving language change, such as production economy and frequency effects that result in phonetic reduction, may not be at work in every individual in the same way or at the same time. Moreover, functional or social mechanisms that lead to innovation in the early stages of language change need not be at work in later stages, as individuals later may acquire the innovation purely due to frequency when the innovation is established as the majority in the communal language.

In short, there is no such thing as THE English language, or THE Mandarin language. We’ve got individual ideolects on the one hand, and various convergences and divergences among them.

Now, on p. 13

Both communal language and idiolects are in constant change and reorganization. Languages are in constant flux, and language change is ubiquitous. At the individual level, every instance of language use changes an idiolect’s internal organization (Bybee, 2006).

Think about that for a minute: “every instance of language use changes an idiolect’s internal organization.” Really? Is language really that Heraclitean? I don’t know, but I’ve seen reports by neuroscientist Walter Freeman that are very suggestive in this regard.

Freeman investigates the olfactory cortex in rats (see, e.g. Walter J. Freeman, How Brains Make Up their Minds, Weidenfeld and Nicholson, 1999). He’s discovered that when a rat learns to identify a new odorant, two things happen. Of course, a new neural “signature” for that odorant is added to the rat’s repertoire of neural signatures. What’s more interesting, the existing signatures for the body of learned odorants all change. The whole system is reconfigured to in response to the new learning (p. 83):

A new odorant is learned by adding a new attractor with its basin, but, unlike a fixed computer memory, an attractor landscape is flexible. When a new class is learned, the synaptic modifications in the neuropil jostle the existing basins of the packed landscape, as the connections within the neuropil form a seamless tissue. This is known as attractor crowding. No basin is independent of the others.

It’s difficult to see how those attractors, and their basins, correspond to Dennett’s rigidly fixed memes that pass from one brain to another. Dennett’s memes, which he has compared to computer viruses, are like those hand-coded rules of Good Old Fashioned AI systems. They are “imposed” on the system “from above,” by a programmer operating outside of the system. Dennett’s mental memes are neurally implausible, a fossil left over from an older intellectual era.

I’ll let Bleckner et. al have the last word. Again from p. 13:

Language evolves far away equilibrium as other complex systems do. As we define language primarily through dynamical rules, rather than by forces designed to pull it to a static equilibrium, it shares, along with almost all complex systems, a fundamentally far – from – equilibrium nature. An open system continues to change and adapt as its dynamics are ‘fed” by energy coming into the system, whereas a closed system will reduce to a stable state or equilibrium (Larsen – Freeman & Cameron, 2008).

* * * * *

* The ‘Five Graces Group’: Clay Beckner, University of New Mexico; Richard Blythe, University of Edinburgh; Joan Bybee, University of New Mexico; Morten H. Christiansen, Cornell University; William Croft, University of New Mexico; Nick C. Ellis, University of Michigan; John Holland, Santa Fe Institute; Jinyun Ke, University of Michigan; Diane Larsen – Freeman, University of Michigan; Tom Schoenemann, James Madison University.

** There’s a discussion over at reddit concerning Chomsky’s current importance. H/t Dan Everett via Facebook.

6 thoughts on “Dennett Upside Down Cake: Thinking About Language Evolution in the 21st Century”

  1. Regarding: “It’s difficult to see how those attractors, and their basins, correspond to Dennett’s rigidly fixed memes that pass from one brain to another.”

    “Attractors” are not memes (despite the fact that Dan Sperber claims otherwise). In evolutionary contexts, the term “attractor” often refers to an adaptive peak on a fitness landscape. By contrast memes are things that move on fitness landscapes. Also, the context here is individual learning. Memes are conventionally to do with social learning – making any link even more tenuous. No wonder you don’t see much correspondence.

  2. Re: “Dennett’s mental memes are neurally implausible, a fossil left over from an older intellectual era.”

    To recap, the idea that Dennett supports is that memes are composed of cultural information. This exists inside brains in the same way that it exists inside books, DVDs, and computers and in light waves, and vibrations in the air. Information-carrying signals are substrate neutral, and can be represented in any medium.

    Despite your curious attempts to dismiss and ridicule this idea, there’s absolutely nothing wrong with it. Indeed, most modern students of cultural evolution (though not me) *define* culture as being composed of information. “Ideational culture” they call the concept. So, you’re not just arguing with Dennett and Blackmore, but also with Boyd, Richerson, Mesoudi, Durham – and many others.

    I find the level of confusion you attribute to your opponents on this whole issue to be rather surprising. What counts as phenotype and genotype in cultural evolution affects what counts as a developmental process – and it changes what we use the terms “phenotype” and “genotype” to mean. This isn’t just a matter of convention, and some conceptions of what counts as a cultural “phenotype” are more useful than others. However your resulting anti-Dennett tirade seems quite surprising. It is true that my views are more closely aligned with his, but my response to such a famous philosopher showing an interest in the exciting area of cultural evolution is to welcome him with open arms to the field – in the hope that others will follow in his footsteps.

  3. …my response to such a famous philosopher showing an interest in the
    exciting area of cultural evolution is to welcome him with open arms to
    the field – in the hope that others will follow in his footsteps.

    It seems to me that you’ve got that backwards. Dennett was working on memetics long before you were. You are in no position to welcome him into it. The field isn’t yours. You have no proprietary rights in it.

    It seems to me that, to use an analogy, you, and Dennett and many others, are in effect using a meter stick to measure phenomena where the action taking place on the centimeter and millimeter scale. Your conceptualization is too coarse to capture the action.

  4. In that passage I was specifically referring to Walter Freeman’s work in neuroscience. In Freeman’s neuroscientific model an attractor would correspond to a sensory perception (he investigates rat olfaction), or an idea, even an ideotype as I have defined it (I worked this out in some detail in my book on music, Beethoven’s Anvil). He is talking about something that exists in a mind.

    It’s not clear to me just how Sperber uses the idea. In Explaining Culture, Ch. 5, he has a footnote that suggests he has the idea from complex dynamics, but he doesn’t develop it very well in the body of the text. But it’s not Sperber’s conception that interests me, as should be obvious from the fact that I quoted a well-known neuroscientist. Now maybe you think that neuroscience is irrelevant to the study of what happens inside people’s heads. Maybe you think your informatic memetics allows you to skip over all that messy neuroscience and psychology. You are wrong.

  5. I’ve always hoped that some day I might run into Aristotle walking his woolly mammoth in the busy metropolis of Timaru. Then I could hug him, give him a dedicated copy of my newest book and say ‘Ta!’ in exchange for his contribution to my field: science.

  6. “… cultural information. This exists inside brains in the same way that it exists inside books, DVDs, and computers and in light waves, and vibrations in the air.”

    Ah, no. Books are dead matter, paper marked with ink. Brains are highly complex living matter and are the most energy-intensive kind organic tissue. Processing information is work and burns energy. Books can’t process anything.

    If this mysterious information stuff can exist in books and brains on equal terms then it is a conceptualization that’s too nebulous to be of any value in understanding, for example, how it is that brains read and write books. Yes, informatic processes can utilize a wide variety of substrates, but that does not relieve you, the thinker, of the responsibility of examining the physical mechanisms that are operating.

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.