Universal Grammar and Biological Variation: An EvoDevo Agenda for Comparative Biolinguistics
- First Online:
- 1.8k Downloads
Recent advances in genetics and neurobiology have greatly increased the degree of variation that one finds in what is taken to provide the biological foundations of our species-specific linguistic capacities. In particular, this variation seems to cast doubt on the purportedly homogeneous nature of the language faculty traditionally captured by the concept of “Universal Grammar.” In this article we discuss what this new source of diversity reveals about the biological reality underlying Universal Grammar. Our discussion leads us to support (1) certain hypotheses advanced in evolutionary developmental biology that argue for the existence of robust biological mechanisms capable of canalizing variation at different levels, and (2) a bottom-up perspective on comparative cognition. We conclude by sketching future directions for what we call “comparative biolinguistics,” specifying which experimental directions may help us succeed in this new research avenue.
KeywordsBiolinguistics Evolutionary developmental biology (EvoDevo) Genetics Language disorders Variation
The present article is concerned with the amount and kind of variation that we think linguists and researchers in allied disciplines should wrestle with if they are to contribute to a proper characterization of the biological foundations of language and, in so doing, to a rapprochement of the cognitive sciences with the biological sciences, an enterprise that we refer to as biolinguistics (Di Sciullo and Boeckx 2011; Boeckx 2013; Boeckx and Grohmann 2013).
Our main contention in this article is that although the comparative method has figured prominently in linguistics, the objects routinely compared (languages, dialects, sociolects) may not be the only, or indeed the most appropriate ones to shed light on the biological foundations of our species-specific linguistic capacity. There are, we claim, deeper layers of variation to explore and to understand. Indeed, as we intend to show here, these deeper layers of variation beg questions regarding the proper biological interpretation of standard concepts in the field of (bio)linguistics, such as Universal Grammar (Chomsky 1965). We believe that in part it is the failure to properly address these sources of variation that has rendered the adjective “universal” (and in fact the whole argument of language as a specific component of human biology) more controversial than it should be.
At the same time, we also think, and will argue in what follows, that linguists have asked questions regarding the locus of linguistic variation (traditionally construed) that would be useful to extend to the layers of variation we are about to highlight. But for this to be successful, it is important to carefully select, and properly conceptualize, the tools one uses to establish comparisons. Here, we think it is imperative to formulate these tools at the right level of granularity to allow for interdisciplinary exchange, in line with Poeppel and Embick (2005). It is our conviction that once this is done, these tools and questions about variation would enrich the range of studies that currently constitutes what we may want to call “comparative biolinguistics,” to be distinguished from the traditional label of comparative linguistics, for reasons that will be spelled out in this article.
Our reflections are structured as follows. First, we will clarify why we think this new kind of linguistic variation is real, and problematic for standard characterizations of certain central concepts in the field of (bio)linguistics. In a nutshell, linguists routinely acknowledge, and, we believe, have quite successfully dealt with, linguistic variation at the surface (languages, dialects, sociolects, and the like). At the same time, they have usually regarded the faculty of language giving rise to these variants as uniform across the species (pathologies aside), and genetically determined. However, the recent emphasis on the biological underpinnings of language—the return of biolinguistic concerns—has begun to reveal deeper layers of variation down to the genetic level that make these standard assumptions unreasonable, and beg the question of where the uniformity or universality of grammar comes from.
We will then proceed to argue that in fact this deeper variation is problematic mainly for naive approaches to the biology of language. However, we find that these approaches are still the majority within the field, and should therefore be corrected to accommodate the variation we find.
We will then suggest that new theoretical approaches coming from biology, paradigmatically, evolutionary developmental biology (EvoDevo), could help us deal with, and account for, all the observed variation, while offering sources of uniformity to reliably give rise to species-typical linguistic capacities. As a matter of fact, we claim that certain EvoDevo concepts could contribute significantly to our understanding of the nature of language disorders.
Finally, we will sketch some concrete and novel ways in which linguists and other scientists in adjacent fields could contribute to a comparative biolinguistics. In this final section, we are led to point out some important limitations of recent tools used in this domain.
Layers of Variation
Variation thoroughly pervades language. The human language faculty manifests itself in the form of many different languages, which are in turn (slightly) diverse across social groups, interactional contexts, geographical areas, and so on. Ultimately, differences can be found from one person to another, and even regarding the same person, for instance, when placed in different scenes. All of this is very familiar, and, we feel, linguists have fairly successfully coped with all this variation (which, of course, is not to say that everything is understood at this level). It is now clear that linguistic phenomena vary in systematic and constrained ways, and can be accounted for by the right mixture of general principles governing cognition and statistical biases (see Baker 2001; Labov 2001; Yang 2006; Pearl 2007; Biberauer 2008; Culbertson 2010, for accessible overviews and concrete proposals to capture this variation).
At the same time, the human faculty for language has routinely been assumed to be uniform within the species (Chomsky 1965, 1980), an assumption captured by the term “Universal Grammar.” For many people, this assumption is a central tenet of the Chomskyan revolution in the language sciences. As is well known, this revolution brought about a radical shift of focus in language studies, a shift oriented towards biology, and away from behaviorism; indeed, a shift that provides the foundations of modern biolinguistics.
The way in which language is acquired by the child, which suggests that language “learning” mechanisms are biased or constrained in certain ways.
The fact that specific language deficits recurrently appear whenever certain brain areas are impaired (either developmentally or after a stroke, a trauma, or a tumoral process).
Ultimately, the existence of language-related components in other extant or extinct species.
These arguments in favor of Universal Grammar continue to be made even by those who otherwise distance their positions from Chomsky’s in a number of ways (see, e.g., Jackendoff 2002, Chap. 4). Recent advances in neuroscience and molecular biology have allowed us to gain a better understanding of the biological underpinnings of language. For instance, we are now in a position to accurately know which brain areas and circuits are active during language processing (see Stemmer and Whitaker 2008; Friederici 2011; or Friederici and Gierhan 2013 for overviews). Similarly, we have identified many of the genes that contribute to the development and the initial wiring of these areas and circuits during growth (Benítez-Burraco 2009; Graham and Fisher 2012). Ultimately, the role of these brain areas and circuits, and of these genes in other extant (and even extinct) species are being studied, revealing (deep) homologues of (aspects of) the language faculty (see the contributions in Di Sciullo and Boeckx 2011; Boeckx and Grohmann 2013).
However, this revival of biolinguistic concerns has not yet substantially changed some of the concepts that remain at the center of the field, such as Universal Grammar. In the vast majorities of studies in biolinguistics (we will return to exceptions), what is called the “linguistic genotype,” i.e., the whole set of genes involved in ultimately allowing for language growth in the child, or “that part of our genetic endowment that is relevant to our linguistic development” (Anderson and Lightfoot 1999, p. 702), is assumed to be uniform across the species, pathologies aside. Moreover, we continue to come across many studies in which the “linguistic brain”—i.e., the whole set of brain areas and circuits involved in language processing—is expected to be equally organized and sharply defined in all individuals. For many, these “language areas” only process specific linguistic features and operations. Ultimately, the faculty of language is supposed to be equally implemented in all subjects as one of the modules encompassing their minds/brains. This module is further thought to be present ab initio, part of a genetic program of sorts (see, e.g., Wexler 1999 and Guasti 2002; and see Longa and Lorenzo 2008 for a critical overview of this “genocentric” literature).
Our reading of the growing literature on the biological foundations of language suggests to us that the state of affairs described in the previous paragraph is quite an erroneous, and in fact simplistic, view of reality. In particular, a fair amount of evidence exists that suggests that the human faculty for language is not really uniform. To be fair, some of this evidence is not new, but we feel that its significance has not yet been properly appreciated. Doing this is the primary purpose of our article.
When one thinks about the possibility of a variable implementation of the language faculty, one is likely to think about facts like: (1) different linguistic modalities can coexist in the same subject, as people bilingual in oral and sign languages nicely exemplify (Emmorey and McCullough 2009); (2) psycholinguistic measures are varied across the normal population (Fenson et al. 2000); (3) language disorders, which plausibly represent different breakdowns of the faculty, are very diverse by nature, and, as the relevant literature has revealed, sometimes difficult to distinguish from stages of normal language development.
All of this is appropriate for the lessons we want to draw in this article. But we think that the current revival of the biolinguistic approach has substantially expanded the realm of variation regarding language and linguistic phenomena. For example, it is now clear that it is not a handful but hundreds of genes that contribute to regulating the development and the functioning of the neural substrate of language (Benítez-Burraco 2009). Importantly, these “language genes” are polymorphic, with some variants giving rise to pathological conditions, but with others being present as well within the normal population. In fact, pathological alleles can be only regarded as such for certain populations and/or environmental conditions. Lastly, the same pathogenic allele can give rise to different language and/or cognitive disorders in different subjects. The celebrated “language gene” FOXP2 and its interactome nicely exemplify this complex state of affairs (see Watkins 2011; Rodenas-Cuadrado et al. 2013 for reviews). Additionally, it is quite difficult to draw a precise map of the neural substrate of language, since the limits of the brain areas involved are rather changeable from one subject to another, and of course, in different clinical conditions (Fedorenko and Kanwisher 2009; Prat and Just 2011)—not to mention the additional problem that at the end of the day mapping is not explaining (Poeppel 2012). Ultimately, developmental trajectories followed by language acquisition, while displaying similar milestones, can be quite diverse, particularly at the cognitive/neurobiological levels (Bates et al. 1988; Dehaene et al. 1997). Language ontogeny in pathological populations is even more diverse, yet equally non-random (Thomas et al. 2009). It is now evident that similar cognitive profiles, in the normal population but also across pathologies, can rely on different brain architectures (Karmiloff-Smith 2010). “Modules are not born; they are made” (Bates et al. 1988, p. 284), although their basic wiring is achieved before birth, plausibly, under genetic guidance. This means that, eventually, we will have to address the diverse sources of variation (genetic, neurobiological, etc.) just mentioned in the context of a developmental perspective, allowing for different trajectories that eventually converge phenotypically.
Unfortunately, for linguists who confess a biological orientation, even those directly concerned with language development, this kind of variation “at the bottom” is ignored. They tend to idealize away from it, at their own peril. We say this because we think that the layers of variation just mentioned and the developmental dynamics that they involve lead to an important conclusion: it seems that there can be different ways of implementing a (more or less) functional faculty of language (see also Hancock and Bever 2013), and that talk of a “linguistic genotype” is fraught with difficulties.
Our main point is that the naive depictions of the biology of language that continue to dominate the literature must be improved. In our opinion, all this variation “at the bottom” can be reconciled with a certain notion of universality, but only if biolinguists are willing to engage seriously, and comprehensively, with the biology literature; i.e., only by pursuing a program that has been called biolinguistics in the strong sense of the term in Boeckx and Grohmann (2007). We wish to stress that this is not a message exclusively directed towards linguists. As we show below, a productive comparative biolinguistics also needs to take into account lessons from linguistics about the nature of language in order to develop appropriate tools for comparison.
Which Biology Does Universal Grammar Require?
In our opinion, the root of the problem discussed in the previous section lies in the assumption that language features are directly rooted in the genome. As the literature (reviewed in Longa and Lorenzo 2008, 2012; Lorenzo and Longa 2009) reveals, a “linguistic genotype,” uniform across the species, is explicitly postulated in numerous publications. This linguistic genotype is further equated to a Universal Grammar. Ultimately, nativism is conflated with geneticism.
However, as we have already pointed out, most (if not all) “language genes” are polymorphic, with some alleles affecting language development also in the “normal” population. In addition, genes do not code for cognitive properties. A direct link between the genotype and the phenotype is not only simplistic, but biologically untenable, given the way in which genes contribute to developmental processes, and how development actually takes place. Genes are not blueprints. Developmental processes also depend on non-genetic factors (Oyama et al. 2001; Newman et al. 2006; Bateson and Mamelli 2007).
Concerning the neural substrate of language, it seems that brain areas actually perform basic kinds of computations that are recruited for different, high-level cognitive functions. As Poeppel and Embick (2005, p. 112) state, “differently structured cortical areas are specialized for performing different types of computations, and… some of these computations are necessary for language but also for other cognitive functions.” Consequently, cognitive capacities such as language are very probably cross-modular by nature. They result from the interplay of these diverse brain areas performing basic, low-level activities (Griffiths 2007). At the same time, it is only these structures that are the final output of genetically driven developmental processes. In fact, it seems that it is only their basic architecture that is genetically encoded, while their functional specificities are environmentally driven in various ways that remain to be elucidated. This is what ultimately supports the claim that modules are not born, but made, and that there is not just one way of implementing a functional language.
Our main point here is that we must seriously study how developmental dynamics, of the sort that is at the heart of EvoDevo approaches in biology (Oyama et al. 2001; West-Eberhard 2003; Callebaut and Rasskin-Gutman 2005; Carroll 2005; Müller 2007), takes place if we really want to adequately deal with this issue of variation in language, and eventually, to achieve a real biological depiction of the language faculty. In a real sense it is the complex and changing interaction between the organism and its environment that shapes the final cognitive architecture of the brain. Different factors, both internal and external, affect language development, to the extent that the different cognitive phenotypes can emerge from the same genotype (the reverse is also true: the same phenotypes can emerge from distinct genotypes). Put differently, it is clear that static depictions of language at all levels of analysis are inadequate, particularly at the biological level. In all fairness, we should point out that linguists are quite familiar with the idea that variation and change are tightly interwoven. For instance, they are well aware of the fact that language change is always preceded by a phase of variation, with different linguistic variants coexisting within the same community of speech (Weinreich et al. 1968 and much subsequent work). However, we do not find this lesson always reflected in the literature on language acquisition and developmental disorders. For us, the right position to adopt is this one: “to understand developmental outcomes, it is vital to identify full developmental trajectories, to assess how progressive change occurs from infancy onwards, and how parts of the developing system may interact with other parts differently at different times across ontogenesis” (Karmiloff-Smith 2009, p. 58).
In order to fully exploit the resources offered to biolinguistics by EvoDevo, we think that it is just as important to stress that although we have insisted on variability so far, there are, of course, many sources of universality—over and above the genes. For example, at the neurobiological level we observe that anatomical variability is quite constrained. In this way, myelinization patterns, receptor maps, cytoarchitectonic probability maps, and other structural features can be confidently established (Zilles and Amunts 2009). Similarly, functional variability seems to be constrained as well, to the extent that regions of interest can be identified (Fedorenko et al. 2010). In sum, although variation is omnipresent, the brain still exhibits a robust structure when processing language (Grodzinsky 2010). At the molecular level, we observe that the initial wiring of the linguistic brain is similarly achieved in all subjects under the guidance of a core set of genetic cues (Benítez-Burraco 2009). When it comes to language growth in the child, we find as well that developmental itineraries are also constrained although not fully predetermined, as Lenneberg (1967) already noted.
Diffuse effects on the brain and on cognitive capacities/abilities are the norm. In fact, developmental disorders are better characterized by associations across domains than by dissociations between them (Bishop 2002).
Deficits in low-level, more generalized processes usually manifest as disturbances of upper, more specialized processes, which ultimately give rise to shortcomings in even higher-level, more specific cognitive capacities (Karmiloff-Smith 2009).
Importantly, impaired, delayed, or deviant systems are still adaptive. It is indeed worth bearing in mind that substantially preserved linguistic abilities can be achieved in spite of deeper cognitive impairments (Sirois et al. 2008; Parisse and Maillart 2009).
At the same time, our reading of the literature suggests to us that breakdowns and compensations, whenever they occur, do not proceed randomly. In reality, some aspects of language processing seem to be particularly vulnerable in all pathological conditions, while others seem to be preserved in all of them. For instance, inflectional morphology is problematic not just for people with specific language impairment (Marchman et al. 1999), but also for those suffering from speech-sound disorder (Mortimer and Rvachew 2010), Down’s syndrome (Eadie et al. 2002), or (a subtype of) autism (Roberts et al. 2004). Ultimately, only some pathological phenotypes have been described, while others have not been observed, a situation that we think could benefit from being modeled in terms of morphospaces or adaptative landscapes (Svensson and Calsbeek 2012). It seems, then, that although there is not just one way of implementing a linguistic brain, it is also true that there are not so many ways of implementing a functional faculty of language.
We believe that key EvoDevo concepts like canalization, developmental plasticity, robustness, evolvability, or adaptative landscapes will greatly help in clarifying, understanding, and eventually explaining the problem, and the full scope of variation in language. In all situations language development turns out to be sensitive to environmental changes, to the extent that different cognitive architectures may result from different linguistic input, as we observe in bilingual people (developmental plasticity). At the same time the language faculty has been shown to be remarkably resistant to (some sort of) environmental perturbations (robustness) to the extent that it recurrently emerges in all individuals, even in some pathological conditions (canalization). Moreover, some components of the language faculty seem to be very resistant to damage and/or to evolutionary change (again, robustness); at the same time linguistic systems seem prone to change (evolvability). All these properties result from the modular organization of the biological substrate of the faculty at all levels of analysis (Wagner and Altenberg 1996; Bergman and Siegal 2002; Kitano 2004). In particular, we want to suggest that developmental dynamics strongly canalizes the existing variation, to the extent that the same phenotype—i.e., a language faculty—can robustly emerge at the term of growth from diverse genotypes and brain architectures. Phenotypic uniformity, then, may be achieved in spite of, and along with neurobiological and genetic diversity.
Conceptually speaking, the state of affairs we have described in this section reminds us of the “embryonic hourglass” situation, discussed from an EvoDevo perspective in Newman (2011). As Newman observes in the context of a discussion on the evolution of animal eggs (2011, p. 467), “why can taxa within a given phylum exhibit very different egg types, pass through a common intermediate morphology (the so-called ‘phylotypic stage’), only to diverge again” (hence the metaphor of the “hourglass”)? Could the logic of the “self-organizing physical processes” that Newman relies on to answer this question also apply to situations like the one we have discussed here in the context of language? It is too early to know, but we think it is worth beginning to think about language development in this way.
We think that this EvoDevo-inspired approach to variation in language is bound to be of great interest for clinical linguistics. In particular, this distancing between the genotype and the phenotype has the potential to explain why in some people affected by a particular language disorder the sequence of the candidate genes is normal (phenocopy), but also why language can be preserved in individuals who are endowed with a pathogenic copy of one of these “language genes” (null penetrance). From a broader perspective, we are in fact tempted to argue that language disorders can be construed as conditions in which that process of canalization has failed to cope with the underlying variation (thus preventing reaching particular degrees of development). Similarly, they can be construed as decanalized states, following the model by Gibson (2009). According to this view, the pervasiveness and the high prevalence of complex genetic diseases among modern populations is a consequence of the uncovering of cryptic genetic variation resulting from the evolution of the human genome, and environmental and cultural perturbations (see Benítez-Burraco and Boeckx 2013 for additional remarks concerning this possibility and also for its implications regarding the evolution of language).
Nonetheless, for us, the crucial point is the recurrent outcome of research that suggests that breakdowns and compensations in language disorders do not occur randomly, and ultimately, that it is only certain normal, impaired, delayed, or deviant faculties of language that emerge in the course of development. We think that this situation is similar to the one that linguists often stress in their studies on language comparison: variation, though large and substantial, is not random, and appears to be confined to only certain components of grammar (Berwick and Chomsky 2011; Boeckx 2011; Boeckx and Leivada 2013a). But as we have already argued, we think that linguists err in taking the genotype to be the source of universality. And so we would like to encourage linguists to abandon this genocentric assumption and embrace the variation we find “at the bottom” by developing adequate tools to characterize it, an issue we return to in the next section.
In closing this section, we would like to ask why it should be the case that adaptability itself is limited or constrained in specific ways, to the extent that some perturbations cannot be eventually compensated by developmental dynamics. We’d like to suggest that certain cognitive processes are more vulnerable per se than others to damage or to developmental disturbances because they rely on less resilient neural networks and thus have less robust compensatory mechanisms. This would be due, we think, to their evolutionary novelty (Toro et al. 2010; Mantini et al. 2013). In fact, the most noticeable outcome of the biological study of language is that the genetic, physiological, and even cognitive mechanisms underlying language are actually robust after thousands of years of stabilizing selection (in other words, because they have a long evolutionary history), while language itself is very delicate. Probably, as suggested by Gibson (2009), the stable equilibrium observed in primates was disrupted by our evolutionary history as a species (in particular, by population bottlenecks and migratory movements; see Mellars 2006), by novel mutations, and by cultural changes. These changes brought about cognitive systems known as modern languages. But at the same time these changes may well have uncovered all that cryptic variation, decanalized the whole system, and ultimately, made language so sensitive to damage (but, we want to stress again, only to some kinds of damage). In other words, the human language faculty is easy to disturb because it is an evolutionary novelty, but at the same time it relies on robust biological mechanisms that are able to compensate many kinds of damage because they are considerably older. Plausibly, this may shed light on why our reading of the literature suggests to us that the same components of language tend to be affected in many language disorders, and why many other aspects of the linguistic phenotype are always quite preserved, and eventually, why these conditions are so prevalent among humans.
We think that this picture properly adjusts to current views of language evolution that embrace continuity and view novelty as the result of a reorganizational process rather than a product of innovative genes (West-Eberhard 2003; Müller 2010), but also with the view of language as a cognitive faculty resulting from the interface of components (cognitive, neural, genetic) otherwise not specifically linguistic (see Boeckx 2013 for review and discussion). The specificity of language would thus rely on the pervasive tendency of the components of the language faculty to interface whenever growth takes place in the presence of a suitable amount of linguistic stimuli.
Tools for Comparative Biolinguistics
Nothing in biology makes sense except in the light of comparison. As the opening passages of Darwin’s two most famous books make clear, the business of biology is variation, variation, variation. Without variation, there can’t be any meaningful selection, any descent with modification, any origin of species. It is natural, then, to expect biolinguists to be fans of the comparative method as well. But as we have been at pains to point out above, a significant aspect of variation has been ignored by many biolinguists. To make matters worse, for much of its (recent) history, biolinguistics has been contrastive, not comparative, in the following sense: biolinguists have emphasized that language is the exclusivity of humans (and that among humans, the language faculty is uniform).
Recently, however, as De Waal and Ferrari (2010) have noted, a significant shift of perspective seems to be under way in cognitive science: the sharp contrastive character of top-down approaches is progressively being replaced by an “increased appreciation that the basic building blocks of cognition might be shared across a wide range of species” (p. 201). This bottom-up approach, seeking to establish “cognitive phylogenies” (Fitch et al. 2010), focuses on the fundamental capacities underlying larger cognitive phenomena and is more in line with the Darwinian logic of descent (Hauser et al. 2002). We think that this shift of perspective, along with the appreciation of variation and non-genetic sources of universality we have urged biolinguists to develop, provides the basis for a genuine, productive, comparative biolinguistic agenda.
To advance this new comparative research program, it is, of course, crucially important to pay attention to the tools one uses to compare. For obvious reasons, not all the tools developed by linguists are equally useful in this respect. In fact, given the modular proclivities of classical cognitive science (Piattelli-Palmarini 2001), the difficulties in exporting linguistic technology outside of the comfort zone of comparative linguists were to be expected. Not surprisingly, progress on the genetic basis of language capacities has been called “a linguist’s nightmare” (Piattelli-Palmarini and Uriagereka 2011), and although scholars have long been captivated by the parallels between birdsong and human speech and language, concrete, theoretically-informed proposals capturing the differences and the similarities across vocal learning capacities are hard to come by.
Understanding developmental and evolutionary aspects of the language faculty requires comparing adult languages users’ abilities with those of non-verbal subjects, such as babies and non-human animals. Classically, comparative work in this area has relied on the rich theoretical frameworks developed by linguists in the generative grammar tradition. However, the great variety of generative theories and the fact that they are models of language specifically makes it difficult to know what to test in animals and children lacking the expressive abilities of normal, mature adults. We suggest that this problem can be mitigated by tapping equally rich, but more formal mathematical approaches to language.
By resorting to the Chomsky hierarchy, Fitch and Hauser hoped to avoid “theory-internal” debates corresponding to the choice of a theoretical idiom (Head-driven Phrase Structure Grammar, Government-and-Binding, Lexical-Functional Grammar, etc.) and use computational primitives that were not so language-specific as to vitiate any cross-species, or cross-domain comparison. As Heinz (2011, p. 146; emphasis in original) writes, “since any pattern is a language [in the sense of the Chomsky hierarchy], a distinct advantage of the Chomsky Hierarchy is [that] it allows for the comparison of patterns from different domains.”
Building on Chomsky’s (1956, 1957) foundational results concerning the limitation of finite-state machines to capture natural language generalizations, Fitch and Hauser (2004) claimed that cotton-top tamarins could not detect structures in stimuli that went beyond the computational capacity of finite-state automata. The results proved controversial in more than one way (Goudarzi 2006; Liberman 2006; Anderson 2008), but so have the results (Gentner et al. 2006; Abe and Watanabe 2011) suggesting that some songbirds outperformed cotton-top tamarins, achieving learning results beyond the finite-state boundaries (Van Heijningen et al. 2009; Berwick et al. 2011, 2012; Ten Cate and Okanoya 2012). Apart from issues of experimental design, we believe that the overarching problem lies in the adequacy of the Chomsky hierarchy in assessing cognitive profiles. Although formal language theory can certainly help in designing experiments shedding light on mental abilities (see the papers collected in Fitch and Friederici 2012), it suffers from a major problem in the context of biolinguistics. It is indeed well known that the Chomsky hierarchy is of limited use in characterizing human linguistic competence. As Berwick et al. (2012) correctly observe, the hierarchy is both “too weak and too strong,” failing as it does to cut natural language at its joints.
Heinz and Idsardi (2011, 2013) and Heinz (2014) usefully summarize important lessons that linguists have derived from applying the Chomsky hierarchy to the study of natural languages. They highlight the fact that not all natural language patterns fall exactly in the same range within the Chomsky hierarchy. In fact, as Fig. 1 reveals, patterns are quite scattered. For example, phonological patterns do not appear to require grammars that distinguish infinitely many states, unlike some syntactic patterns, which appear to require formal grammars that do. This distinction between these two domains of the language faculty highlights the fact that the human language faculty is not monolithic. It is more like a mosaic, with all the implications this has for evolutionary studies. Heinz and Idsardi also comment on the hypothesis that natural language patterns are at most mildly context sensitive (Joshi 1985) and stress that we should not conclude from this that were it true, any mildly context-sensitive pattern is automatically a possible natural language one.
There are three lessons that we would like to draw from the use of the Chomsky hierarchy in comparative biolinguistics. First, choosing to use this tool amounts to ignoring most of the research done on the nature of human languages over the past 50 years. Such research has moved away from the Chomsky hierarchy, in large part because it became clear very quickly that it does not “uniquely” characterize human language in the sense that it does not identify any (sub)region of the hierarchy as the exclusive property of natural language. As such, it does not characterize precisely enough the capacity we as humans have. Consider the fact that, as numerous linguists have observed, no natural language has rules that require counting past two. But it is a logically possible language pattern, one that the Chomsky hierarchy can capture. In fact, as Heinz (2014) notes, it is a regular pattern, falling well below the attested power range of natural language syntax. But it is a constraint that significantly shapes the human language capacity, one that we would want to ask comparative questions about. The Chomsky hierarchy does not allow us to do this. Only detailed theoretical linguistics work does.
Second, comparative biolinguistics experiments that rely on the Chomsky hierarchy tend to ignore the divide between phonology and syntax stressed by Heinz and Idsardi (2011, 2013), or rather, experimenters tend to take the notion of “syntax” too literally. The divide emphasized by Heinz and Idsardi (2011, 2013) is actually one between phonology on the one hand and syntax-semantics on the other. To the best of our knowledge, none of the evidence for syntactic patterns falling outside the regular language class comes from “pure” syntactic patterns; rather, all of them involve patterns with (structural) semantic consequences. This is an important consideration for experiments because virtually all comparative biolinguistic experiments using the Chomsky hierarchy are artificial language experiments that seek to target the learning of pure syntactic patterns (these languages don’t mean anything). But there are no such patterns in natural languages. Not surprisingly, when semantic cues are added to the experiment, as in Fedor et al. (2012), these were found to boost the learning of more complex formal grammars in humans. By ignoring semantics, artificial language experiments may well be removing the component that gives natural language syntax its distinctive computational signature.
The third, and perhaps most important, lesson that we’d like to draw is that it is not at all clear what the expectations of the relevant experiments are, due to the inherent limitations of the Chomsky hierarchy in capturing the true nature (i.e., the constraints) of natural languages. This is not to say, of course, that such experiments are pointless. They can tell us many things, but it is not clear that they allow us to draw solid conclusions concerning the system linguists call natural languages. To make ourselves clear, suppose we found out that a non-human species were capable of mastering a mildly-context-sensitive language in the context of an artificial language experiment. What would we be able to conclude from this? Our answer is, not much. In fact, we would only be able to conclude that they were able to learn this pattern. But it would not immediately tell us the underlying algorithm used. As reviewed in Ojima and Okanoya (2013), all the artificial grammar experiments to date suffer from this problem, as there is more than one way to acquire a particular pattern. (Remember that one of the strengths of the Chomsky hierarchy is that it allows for the comparison of patterns from different domains. But this is also its weakness—it is too nonspecific to exclude alternative cognitive ways of capturing a given pattern.)
As should be clear by now, even if a non-human species were capable of mastering a mildly-context-sensitive language in the context of an artificial language experiment, we would not be able to conclude anything regarding the human language faculty. Because there is no (sub)region of the Chomsky hierarchy that is exclusively occupied by natural languages, we would not be able to conclude that the success of non-human primates shows that they have a component of the human language faculty. It is for this reason that we are skeptical about the use of the Chomsky hierarchy to identify language regions in the brain (Fitch and Friederici 2012; Moro and Chesi in 2014) or to draw linguistic inferences from artifacts in the fossil record (Camps and Uriagereka 2006; Balari and Lorenzo 2013; Longa 2013).
As Boeckx (2013) observed, finding a substitute, or complement to the Chomsky hierarchy to construct cognitive phylogenies will be a serious challenge for the years to come. In the remainder of this section, we would like to sketch a possible research avenue that seems to us to have the right properties, and that connects to some EvoDevo concerns discussed above, as it draws on aspects of our biology that are conserved across species, and at the same time that are known to vary across human populations (language disorders).
To begin with, we’d like to step back and consider what we believe was the main factor behind the renewed interest in the comparative method in linguistics at the end of the 1970s. The “new comparative syntax,” as it has been called, grew out of proposals articulated in Chomsky (1981). These proposals, as Chomsky acknowledged on numerous occasions (e.g., Chomsky 2009), grew out of reflections inspired by the work of Jacob and Monod (1961) on gene regulation. In the fullness of time the Jacob-Monod model developed into EvoDevo genomics (Carroll 2005). The most iconic finding of this field is the hox gene set, which confirmed Monod’s prediction that “what is true for E. coli is also true for the elephant.” As is now well established, the set of genes regulating development across a wide range of species is shared, which greatly enhances the possibilities of comparison. Against the background of this deep conservation, species differences (variations) can be thought of as little tweaks and nudges, like the 30 variations of the aria that Bach offered us in what is now known as the Goldberg variations. Arthur (2004), for instance, suggests that all variations reduce to instances of heterochrony (different timing of gene expression), heterotopy (different location of gene expression), heterometry (more of the gene product being made), and heterotypy (change in the nature of the gene product; e.g., switch on different target genes).
This model to understand variation was borrowed into linguistics, where it came to be known as the Principles-and-Parameters approach, with the principles providing the underlying uniformity and the parameters the sources of the surface variations (Chomsky 1981; Baker 2001). The analogy worked well for two decades, allowing for considerable empirical progress, but in recent years, the foundational assumptions of the Principles-and-Parameters model have been questioned (Newmeyer 2004, 2005; Boeckx 2011, 2014; Boeckx and Leivada 2013a). It appears that in order to capture the variation that comparative linguists focus on, something else is needed, perhaps something along the lines of Boeckx and Leivada (2013b). However, the logic of Principles and Parameters may be just what is needed in the context of comparative biolinguistics. Such a model need not require genes to provide the relevant parameters (even for biology, genes may be followers, not just leaders; Newman and Bhat 2009; Schwander and Leimar 2011), but its logic demands that one find an aspect of deep conservation on which variants could be grafted.
Buzsáki et al. (2013) may provide just what is needed in this context. They observe that despite the several-thousandfold increase of brain volume during the course of mammalian evolution, the hierarchy of brain oscillations (brain rhythms) remains remarkably preserved. This conserved aspect of our biology is directly relevant for comparative biolinguistics: it offers the possibility of conceiving of cross-species differences, or, as Buzsáki et al. (2013) discuss, of cognitive diseases, as slight variations (disruptions) within the preserved network constellation that would constitute a universal brain syntax (Buzsáki 2010)—dysrhythmias and oscillopathies, as Buzsáki et al. (2013) call them. Put differently, the preservation of brain rhythms in mammals would be the cognitive scientist’s hox genes.
Obviously, to put this hypothesis to the test in the context of language, it is necessary for linguists to translate their findings concerning the properties of the human language faculty in terms of brain rhythms, to offer a mind/brain model on which to formulate parameters giving rise to distinct cognitive profiles. This translation step may, in fact, be independently necessary to bridge the gap between mind and brain. David Poeppel has written eloquently and accessibly about the challenges neurolinguistics faces (Poeppel 2005, 2011, 2012; Poeppel and Embick 2005). The heart of the matter, according to Poeppel, is the “granularity mismatch” (or “mapping”) problem: the objects of study in theoretical linguistics and in neuroscience don’t match. As a result, mapping one onto the other has proven impossible. Accordingly, Marr’s (1982) vision of cognitive neuroscience based on linking levels of analyses, to which biolinguistics should aspire, remains distant.
we need linguistic models that are explicit about the computational primitives (structures and operations) they require, and that attempt to define linguistic problems at a fine enough grain that one can discuss algorithmic and implementational approaches to their solution. We need a list of computations that linguistic theorists deem indispensable to solve their particular problem (e.g., in phonology, syntax, or semantics). (Fitch 2009, p. 298)
[l]inguists and psycholinguists owe a decomposition (or fractionation) of the particular linguistic domain in question (e.g., syntax) into formal operations that are, ideally, elemental and generic. Generic formal operations at this level of abstraction can form the basis for more complex linguistic representation and computation. (Poeppel 2005, p. 11)
A rhythm-based model may have just the right kind of characteristic envisaged by Poeppel. In fact, direct evidence of the fruitfulness of this approach in the language domain comes from Poeppel’s own work, beginning with Poeppel (2003) and culminating with Giraud and Poeppel (2012).
What Poeppel and colleagues have shown is that by focusing on the endogenous rhythms generated by the cortex, it is possible to understand (as opposed to merely localizing) the cerebral specialization for speech perception and production, and to shed light on the nature of phrasal phonology. The main thesis is that neuronal oscillations contribute to cognition in several ways: for example, by segregating information and organizing spike timing. Specifically, a series of oscillations (in the delta, theta, and gamma ranges) appear to be able to track the dynamics of speech. In doing so, they “chunk” or “package” incoming information into units of the appropriate temporal granularity. This packaging corresponds to units of phrasal phonology (linking the algorithmic and computational levels). Taking linguistic processes as involving multiple subprocesses with different characteristics, as opposed to being monolithic, each one associated with different frequencies of neural oscillations, may offer us the right component parts to identify what goes wrong in disorders, or how different orchestrations may give rise to cross-species cognitive differences, and endophenotypes.
The translation work needed will be slow, of course, but the fact that we already have working candidate models of the right format for some aspects of language like phonology (Giraud and Poeppel 2012), for which we have good animal models (vocal learners) as well as a growing amount of genetic information (the FOXP2 interactome), suggests that the first fruits of a comparative biolinguistics may not be too distant.
It is true that one finds less genetic variation in our species than in our cousins (a conclusion reinforced by the recent sequencing of the genomes of a large number of great apes from across Africa and Southeast Asia; Prado-Martínez et al. 2013), but we should not idealize away from the variation that nonetheless exists. That variation is real, and contains important lessons concerning the biological foundations of the human language faculty. In fact, we have argued here that this variation is only problematic for the naive—and unfortunately, still dominant—conceptions of the biological underpinnings of language in the language sciences. While there are many more sources of variation than those linguists tend to study, there are also many more sources of universality that canalize this variation and allow for languages to reliably develop in the individual than linguists tend to assume. We have furthermore suggested that lessons from comparative and theoretical linguistics may profitably be extended to deeper layers of variation, and could offer a starting point for a new branch of comparative linguistics, one we have called comparative biolinguistics. This new subfield may help correct the exuberant (genocentric) nativism that appears to be so problematic (Boeckx and Leivada 2013a), particularly for achieving a fruitful assimilation of linguistics to biology, allowing for a real, biologically grounded biolinguistics to emerge.
- Abe K, Watanabe D (2011) Songbirds possess the spontaneous ability to discriminate syntactic rules. Nat Neurosci 14:1067–1074Google Scholar
- Anderson SR (2008) The logical structure of linguistic theory. Language 84:795–814Google Scholar
- Anderson SR, Lightfoot DW (1999) The human language faculty as an organ. Annu Rev Physiol 62:697–722Google Scholar
- Arthur W (2004) Biased embryos and evolution. Cambridge University, CambridgeGoogle Scholar
- Baker M (2001) The atoms of language. Basic Books, New YorkGoogle Scholar
- Balari S, Lorenzo G (2013) Computational phenotypes: towards an evolutionary developmental biolinguistics. Oxford University Press, OxfordGoogle Scholar
- Bates E, Bretherton I, Snyder L (1988) From first words to grammar: individual differences and dissociable mechanisms. Cambridge University Press, CambridgeGoogle Scholar
- Bateson P, Mamelli M (2007) The innate and the acquired: useful clusters or a residual distinction from folk biology? Dev Psychobiol 49:818–831Google Scholar
- Benítez-Burraco A (2009) Genes y lenguaje. Reverté, BarcelonaGoogle Scholar
- Bergman A, Siegal M (2002) Evolutionary capacitance as a general feature of complex gene networks. Nature 424:549–552Google Scholar
- Berwick RC, Chomsky N (2011) The biolinguistic program: the current state of its development. In: Di Sciullo AM, Boeckx C (eds) The biolinguistic enterprise. Oxford University Press, Oxford, pp 19–41Google Scholar
- Berwick RC, Okanoya K, Beckers GJ et al (2011) Songs to syntax: the linguistics of birdsong. Trends Cogn Sci 15:113–121Google Scholar
- Biberauer T (ed) (2008) The limits of syntactic variation. John Benjamins, AmsterdamGoogle Scholar
- Bishop DVM (2002) Motor immaturity and specific speech and language impairment: evidence for a common genetic basis. Am J Med Genet 114:56–63Google Scholar
- Boeckx C (2011) Approaching parameters from below. In: Di Sciullo AM, Boeckx C (eds) The biolinguistic enterprise. Oxford University Press, Oxford, pp 205–221Google Scholar
- Boeckx C (2013) Biolinguistics: forays into human cognitive biology. J Anthropol Sci 91:1–28Google Scholar
- Boeckx C (2014) What principles & parameters got wrong. In: Picallo C (ed) Linguistic variation and the minimalist program. Oxford University Press, Oxford. pp 155–178Google Scholar
- Boeckx C, Grohmann KK (2007) The biolinguistics manifesto. Biolinguistics 1:1–8Google Scholar
- Boeckx C, Grohmann KK (eds) (2013) The Cambridge handbook of biolinguistics. Cambridge University Press, CambridgeGoogle Scholar
- Boeckx C, Leivada E (2013a) Entangled parametric hierarchies: problems for an overspecified universal grammar. PLoS One 8:e72357Google Scholar
- Boeckx C, Leivada E (2013b) On the particulars of universal grammar: implications for acquisition. Presented at the workshop, “Towards a (non-syntactic) theory of variation.” Bilbao, Spain, June 2013Google Scholar
- Buzsáki G (2010) Neural syntax: cell assemblies, synapsembles, and readers. Neuron 68:362–385Google Scholar
- Buzsáki G, Logothetis N, Singer W (2013) Scaling brain size, keeping timing: evolutionary preservation of brain rhythms. Neuron 80:751–764Google Scholar
- Callebaut W, Rasskin-Gutman D (eds) (2005) Modularity: understanding the development and evolution of natural complex systems. MIT Press, Cambridge, MAGoogle Scholar
- Camps M, Uriagereka J (2006) The gordian knot of linguistic fossils. In: Rossello J, Martin J (eds) The biolinguistic turn. PPU, Barcelona, pp 34–65Google Scholar
- Carroll SB (2005) Endless forms most beautiful: the new science of EvoDevo and the making of the animal kingdom. Norton, New YorkGoogle Scholar
- Chomsky N (1956) Three models for the description of language. IEEE T Inform Theory 2:113–124. Reprinted, with corrections. In: Luce RD, Bush R, Galanter E (eds) Readings in mathematical psychology. Wiley, New York, pp 125–155Google Scholar
- Chomsky N (1957) Syntactic structures. Mouton, The HagueGoogle Scholar
- Chomsky N (1965) Aspects of the theory of syntax. MIT Press, CambridgeGoogle Scholar
- Chomsky N (1980) Rules and representations. Columbia University Press, New YorkGoogle Scholar
- Chomsky N (1981) Lectures on government and binding. Foris, DordrechtGoogle Scholar
- Chomsky N (2009) Opening remarks and conclusion. In: Piattelli-Palmarini M, Salaburu P, Uriagereka J (eds) Of minds and language: a basque encounter with Noam Chomsky. Oxford University Press, Oxford, pp 13–43Google Scholar
- Culbertson J (2010) Learning biases, regularization, and the emergence of typological universals in syntax. PhD Thesis, Johns Hopkins UniversityGoogle Scholar
- De Waal F, Ferrari PF (2010) Towards a bottom-up perspective on animal and human cognition. Trends Cogn Sci 14:201–207Google Scholar
- Dehaene S, Dupoux E, Mehler J et al (1997) Anatomical variability in the cortical representation of first and second language. NeuroReport 8:3809–3815Google Scholar
- Di Sciullo AM, Boeckx C (eds) (2011) The biolinguistic enterprise: new perspectives on the evolution and nature of the human language faculty. Oxford University Press, OxfordGoogle Scholar
- Eadie PA, Fey ME, Douglas JM et al (2002) Profiles of grammatical morphology and sentence imitation in children with specific language impairment and DS. J Speech Lang Hear Res 45:720–732Google Scholar
- Emmorey K, McCullough S (2009) The bimodal bilingual brain: effects of sign language experience. Brain Lang 109:124–132Google Scholar
- Fedor A, Varga M, Szathmáry E (2012) Semantics boosts syntax in artificial grammar learning tasks with recursion. J Exp Psychol Learn Mem Cogn 38:776–782Google Scholar
- Fedorenko E, Kanwisher N (2009) Neuroimaging of language: why hasn’t a clearer picture emerged? Lang Linguist Compass 3:839–865Google Scholar
- Fedorenko E, Hsieh PJ, Nieto-Castañón A et al (2010) New method for fMRI investigations of language: defining ROIs functionally in individual subjects. J Neurophysiol 104:1177–1194Google Scholar
- Fenson L, Bates E, Dale P et al (2000) Measuring variability in early child language: don’t shoot the messenger. Child Dev 71:323–328Google Scholar
- Fitch WT (2009) Prolegomena to a future science of biolinguistics. Biolinguistics 3:283–320Google Scholar
- Fitch W, Friederici AD (2012) Artificial grammar learning meets formal language theory: an overview. Philos Trans R Soc Lond B Biol Sci 367:1933–1955Google Scholar
- Fitch WT, Hauser MD (2004) Computational constraints on syntactic processing in a nonhuman primate. Science 303:377–380Google Scholar
- Fitch W, Huber L, Bugnyar T (2010) Social cognition and the evolution of language: constructing cognitive phylogenies. Neuron 65:795–814Google Scholar
- Friederici AD (2011) The brain basis of language processing: from structure to function. Physiol Rev 91:1357–1392Google Scholar
- Friederici AD, Gierhan SM (2013) The language network. Curr Opin Neurobiol 23:250–254Google Scholar
- Gentner TQ, Fenn KM, Margoliash D et al (2006) Recursive syntactic pattern learning by songbirds. Nature 440:1204–1207Google Scholar
- Gibson G (2009) Decanalization and the origin of complex disease. Nat Rev Genet 10:134–140Google Scholar
- Giraud A-L, Poeppel D (2012) Cortical oscillations and speech processing: emerging computational principles and operations. Nat Neurosci 15:511–517Google Scholar
- Goudarzi S (2006) Chomsky challenge: bird said to grasp key aspect of human language. Livescience.com, 26 AprilGoogle Scholar
- Graham SA, Fisher SE (2012) Decoding the genetics of speech and language. Curr Opin Neurobiol 23:43–51Google Scholar
- Griffiths PE (2007) Evo-devo meets the mind: towards a developmental evolutionary psychology. In: Brandon R, Sansom R (eds) Integrating evolution and development: from theory to practice. MIT Press, Cambridge, MA, pp 195–225Google Scholar
- Grodzinsky Y (2010) The picture of the linguistic brain: how sharp can it be? Reply to Fedorenko & Kanwisher. Lang Linguist Compass 4:605–622Google Scholar
- Guasti MT (2002) Language acquisition: the growth of grammar. MIT Press, Cambridge, MAGoogle Scholar
- Hancock R, Bever TG (2013) Genetic factors and normal variation in the organization of language. Biolinguistics 7:75–95Google Scholar
- Hauser MD, Chomsky N, Fitch WT (2002) The faculty of language: what is it, who has it, and how did it evolve? Science 298:1569–1579Google Scholar
- Heinz J (2011) Computational phonology. Part I: foundations. Lang Linguist Compass 5:140–152Google Scholar
- Heinz J (2014) Computational theories of learning and developmental psycholinguistics. In: Lidz J, Snyder W, Pater J (eds) The Cambridge handbook of developmental linguistics. Cambridge University Press, CambridgeGoogle Scholar
- Heinz J, Idsardi W (2011) Sentence and word complexity. Science 333:295–297Google Scholar
- Heinz J, Idsardi W (2013) What complexity differences reveal about domains in language. Top Cogn Sci 5:111–131Google Scholar
- Jackendoff R (2002) Foundations of language: brain, meaning, grammar, evolution. Oxford University Press, OxfordGoogle Scholar
- Jacob F, Monod J (1961) On the regulation of gene activity. Cold Spring Harbor Symp Quant Biol 26:193–211Google Scholar
- Joshi AK (1985) Tree adjoining grammars: how much context-sensitivity is required to provide reasonable structural descriptions? In: Dowty D, Karttunen L, Zwicky A (eds) Natural language parsing. Cambridge University Press, Cambridge, pp 206–250Google Scholar
- Karmiloff-Smith A (2009) Nativism versus neuroconstructivism: rethinking the study of developmental disorders. Dev Psychol 45:56–63Google Scholar
- Karmiloff-Smith A (2010) A developmental perspective on modularity. In: Glatzeder BM, Goel V, von Müller A (eds) Towards a theory of thinking. Springer, Berlin, pp 179–187Google Scholar
- Kitano H (2004) Biological robustness. Nat Rev Genet 5:826–837Google Scholar
- Labov W (2001) Principles of linguistic change. Blackwell, OxfordGoogle Scholar
- Lenneberg H (1967) Biological foundations of language. Wiley, New YorkGoogle Scholar
- Liberman M (2006) Starlings. http://itre.cis.upenn.edu/myl/languagelog/archives/003076.h
- Longa VM (2013) The evolution of the faculty of language from a Chomskyan perspective: bridging linguistics and biology. J Anthropol Sci 91:1–48Google Scholar
- Longa VM, Lorenzo G (2008) What about a (really) minimalist theory of language acquisition? Linguistics 46:541–570Google Scholar
- Longa VM, Lorenzo G (2012) Theoretical linguistics meets development: explaining FL from a epigeneticist point of view. In: Boeckx C, Horno MC, Mendívil JL (eds) Language, from a biological point of view: current issues in biolinguistics. Cambridge Scholars Publishing, Cambridge, pp 52–84Google Scholar
- Lorenzo G, Longa VM (2009) Beyond generative geneticism: rethinking language acquisition from a developmentalist point of view. Lingua 119:1300–1315Google Scholar
- Mantini D, Corbetta M, Romani GL, et al (2013) Evolutionarily novel functional networks in the human brain? J Neurosci 33:3259-3275. Erratum: J Neurosci 2013, 33:10934Google Scholar
- Marchman VA, Wulfeck B, Ellis Weismer S (1999) Morphological productivity in children with normal language and SLI: a study of the English past tense. J Speech Lang Hear Res 42:206–219Google Scholar
- Marr D (1982) Vision. Freeman, San FranciscoGoogle Scholar
- Mellars P (2006) Why did modern human populations disperse from Africa ca. 60,000 years ago? A new model. Proc Natl Acad Sci USA 103:9381–9386Google Scholar
- Moro A, Chesi C (2014) Computational complexity in the brain. In: Newmeyer F, Preston LB (eds) Measuring linguistic complexity. Oxford University Press, OxfordGoogle Scholar
- Mortimer J, Rvachew S (2010) A longitudinal investigation of morpho-syntax in children with speech sound disorders. J Commun Disord 43:61–76Google Scholar
- Müller GB (2007) Evodevo: extending the evolutionary synthesis. Nat Rev Genet 8:943–949Google Scholar
- Müller GB (2010) Epigenetic innovation. In: Pigluicci M, Müller G (eds) Evolution: the extended synthesis. MIT Press, Cambridge, MA, pp 307–333Google Scholar
- Newman SA (2011) Animal egg as evolutionary innovation: a solution to the “embryonic hourglass” puzzle. J Exp Zool B Mol Dev Evol 316:467–483Google Scholar
- Newman SA, Bhat R (2009) Dynamical patterning modules: a “pattern language” for development and evolution of multicellular form. Int J Dev Biol 53:693–705Google Scholar
- Newman SA, Forgacs G, Müller GB (2006) Before programs: the physical origination of multicellular forms. Int J Dev Biol 50:289–299Google Scholar
- Newmeyer FJ (2004) Against a parameter-setting approach to language variation. Linguist Var Yearb 4:181–234Google Scholar
- Newmeyer FJ (2005) Possible and probable languages: a generative perspective on linguistic typology. Oxford University Press, OxfordGoogle Scholar
- O’Donnell T, Hauser MD, Fitch WT (2005) Using mathematical models of language experimentally. Trends Cogn Sci 9:284–289Google Scholar
- Ojima S, Okanoya K (2013) Non-hierarchical nature of the Chomsky hierarchy-driven artificial-grammar learning. Presented at the Annual Meeting of the English Linguistic Society of Japan, Fukuoka, November 2013Google Scholar
- Oyama S, Griffiths PE, Gray RD (eds) (2001) Cycles of contingency: developmental systems and evolution. MIT Press, Cambridge, MAGoogle Scholar
- Parisse C, Maillart C (2009) Specific language impairment as systemic developmental disorders. J Neurolinguist 22:109–122Google Scholar
- Pearl L (2007) Necessary bias in natural language learning. PhD Thesis, University of MarylandGoogle Scholar
- Piattelli-Palmarini M (2001) Portrait of a “classical” cognitive scientist: what I have learned from Jacques Mehler. In: Dupoux E (ed) Language, brain and cognitive development: essays in Honor of Jacques Mehler. MIT Press, Cambridge, MA, pp 3–21Google Scholar
- Piattelli-Palmarini M, Uriagereka J (2011) FoxP2: a geneticist’s dream; a linguist’s nightmare. In: Di Sciullo AM, Boeckx C (eds) The biolinguistic enterprise. Oxford University Press, Oxford, pp 100–125Google Scholar
- Poeppel D (2003) The analysis of speech in different temporal integration windows: cerebral lateralization as ‘asymmetric sampling in time.’ Speech Commun 41:245–255Google Scholar
- Poeppel D (2005) The interdisciplinary study of language and its challenges. Technical report, Jahrbuch des Wissenschaftskollegs zu BerlinGoogle Scholar
- Poeppel D (2011) Genetics and language: a neurobiological perspective on the missing link (-ing hypotheses). J Neurodev Disord 3:381–387Google Scholar
- Poeppel D (2012) The maps problem and the mapping problem: two challenges for a cognitive neuroscience of speech and language. Cogn Neuropsychol 29:34–55Google Scholar
- Poeppel D, Embick D (2005) Defining the relation between linguistics and neuroscience. In: Cutler A (ed) Twenty-first century psycholinguistics: four cornerstones. Erlbaum, Hillsdale, NJ, pp 103–120Google Scholar
- Prado-Martínez J, Sudmant PH, Kidd JM et al (2013) Great ape genetic diversity and population history. Nature 499:471–475Google Scholar
- Prat CS, Just MA (2011) Exploring the neural dynamics underpinning individual differences in sentence comprehension. Cereb Cortex 21:1747–1760Google Scholar
- Roberts JA, Rice ML, Tager-Flusberg H (2004) Tense marking in children with autism. Appl Psycholinguist 25:429–448Google Scholar
- Schwander T, Leimar O (2011) Genes as leaders and followers in evolution. Trends Ecol Evol 26:143–151Google Scholar
- Sirois S, Spratling M, Thomas MS et al (2008) Précis of neuroconstructivism: how the brain constructs cognition. Behav Brain Sci 31:321–331Google Scholar
- Stemmer B, Whitaker HA (eds) (2008) Handbook of the neuroscience of language. Elsevier, AmsterdamGoogle Scholar
- Svensson EI, Calsbeek R (eds) (2012) The adaptative landscape in evolutionary biology. Oxford University Press, OxfordGoogle Scholar
- Ten Cate C, Okanoya K (2012) Revisiting the syntactic abilities of non-human animals: natural vocalizations and artificial grammar learning. Philos Trans R Soc Lond B 367:1984–1994Google Scholar
- Thomas MS, Annaz D, Ansari D et al (2009) Using developmental trajectories to understand developmental disorders. J Speech Lang Hear Res 52:336–358Google Scholar
- Toro R, Konyukh M, Delorme R et al (2010) Key role for gene dosage and synaptic homeostasis in autism spectrum disorders. Trends Genet 26:363–372Google Scholar
- Van Heijningen CAA, De Visser J, Zuidema W et al (2009) Simple rules can explain discrimination of putative recursive syntactic structures by a songbird species. Proc Natl Acad Sci USA 106:20538–20543Google Scholar
- Wagner GP, Altenberg L (1996) Complex adaptations and the evolution of evolvability. Evolution 50:967–976Google Scholar
- Watkins K (2011) Developmental disorders of speech and language: from genes to brain structure and function. Prog Brain Res 189:225–238Google Scholar
- Weinreich U, Labov W, Herzog M (1968) Empirical foundations for a theory of language change. In: Lehmann WP (ed) Directions for historical linguistics: a symposium. University of Texas Press, Austin, pp 95–195Google Scholar
- West-Eberhard MJ (2003) Developmental plasticity and evolution. Oxford University Press, OxfordGoogle Scholar
- Wexler K (1999) Maturation and growth of grammar. In: Ritchie WC, Bhatia TK (eds) Handbook of language acquisition. Academic Press, San Diego, pp 55–109Google Scholar
- Yang C (2006) The infinite gift. Schuster, New YorkGoogle Scholar
- Zilles K, Amunts K (2009) Receptor mapping: architecture of the human cerebral cortex. Curr Opin Neurol 22:331–339Google Scholar
Open AccessThis article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.