Deriving the Constant Rate Effect

The Constant Rate Hypothesis (Kroch 1989) states that when grammar competition leads to language change, the rate of replacement is the same in all contexts affected by the change (the Constant Rate Effect, or CRE). Despite nearly three decades of empirical work into this hypothesis, the theoretical foundations of the CRE remain problematic: it can be shown that the standard way of operationalizing the CRE via sets of independent logistic curves is neither sufficient nor necessary for assuming that a single change has occurred. To address this problem, we introduce a mathematical model of the CRE by augmenting Yang’s (2000) variational learner with production biases over an arbitrary number of linguistic contexts. We show that this model naturally gives rise to the CRE and prove that under our model the time separation possible between any two reflexes of a single underlying change necessarily has a finite upper bound, inversely proportional to the rate of the underlying change. Testing the predictions of this time separation theorem against three case studies, we find that our model gives fits which are no worse than regressions conducted using the standard operationalization of CREs. However, unlike the standard operationalization, our more constrained model can correctly differentiate between actual CREs and pseudo-CREs—patterns in usage data which are superficially connected by similar rates of change yet clearly not unified by a single underlying cause. More generally, we probe the effects of introducing context-specific production biases by conducting a full bifurcation analysis of the proposed model. In particular, this analysis implies that a difference in the weak generative capacity of two competing grammars is neither a sufficient nor a necessary condition of language change when contextual effects are present.


The Constant Rate Effect
In a seminal paper in historical syntax, Kroch (1989) proposed the Constant Rate Hypothesis: [W]hen one grammatical option replaces another with which it is in competition across a set of linguistic contexts, the rate of replacement, properly measured, is the same in all of them.(Kroch 1989:200) Initially (and still logically) a hypothesis, the notion of a constant rate has accumulated enough support over the last three decades for this to be referred to as the Constant Rate Effect, or CRE (see e.g.Pintzuk 2003:511).
The logic behind CREs is as follows: if a variant replaces another variant in two or more different contexts and the rate of change is the same in each of these contexts, then we should assume that only a single change has occurred.CREs have therefore been deployed to argue that two or more apparently unrelated surface changes are in fact manifestations of a single underlying change (Fig. 1).Unifying changes in this way provides strong support for approaches to language in which syntactic variation consists not primarily in lexical or contextual idiosyncrasies but in the values of a finite number of universal parameters, as in the classical Principles & Parameters approach (Chomsky 1981;Chomsky and Lasnik 1993).There is no necessary link between Principles & Parameters and CREs, as Pintzuk (2003:511) emphasizes; the variationist approach within which the Constant Rate Hypothesis is couched is theory-neutral.However, CREs are a useful tool in the armoury of diachronic syntacticians who wish to argue for "the controlling effect of abstract grammatical analyses on patterns in usage data" (Kroch 1989:239).
CREs offer a fresh perspective on the causation of changes.Kroch (1989:238) criticizes the approach to causation in which "the finding that a given context is most favourable to the use of an innovation is taken to show that the innovation is an accommodation to the linguistic functionality of that context."Where there is a disparity between contexts that share the same rate of change, this "reflects functional effects, discourse and processing, on the choices speakers make among the alternatives available to them in the language as they know it; and the strength of these effects remains constant as the change proceeds" (Kroch 1989:238).In other words: surface changes are to be thought of as reflexes of underlying grammatical changes; the discrepancies in frequencies seen at the surface level are due to extra-grammatical factors, or contextual effects, which are independent of the underlying change itself and constant across time.
Fig. 1 A classical example of a CRE: the emergence of do-support in Early Modern English in negative declarative and four types of interrogative sentences; data from Kroch (1989:224, Table 3).Periphrastic do is adopted at slightly different times in the different contexts, but the rate of adoption appears to be similar across contexts.Kroch (1989) identified loss of V-to-T movement as the underlying parametric change responsible for this constant rate The usual procedure for detecting a CRE in some diachronic data is to fit a logistic curve (1) to each of the contexts separately and then to compare the growth rates of these curves against each other.
Here, p t is the frequency of either the innovatory or the receding variant (or parameter value) in a given context at time t, and s is the (time-independent) rate of change in that context.The k parameter serves to translate the curve along the time axis, indicating the point of greatest growth, or the tipping point, of p t (Fig. 2).With this operationalization, we have the following procedure for establishing a CRE: a logistic curve of the form (1) is first fit to each of the contexts of interest separately and independently.Then, if variation among the s or 'slope' parameters for these curves is found to fall within a reasonable confidence interval, the change is said to proceed at the same ('constant') rate in all contexts.Variation among the k or 'intercept' parameters, on the other hand, is allowed and is where the contextual effects, independent of the underlying grammatical change, are thought to manifest themselves.This is the procedure used in a number of studies that have sought to establish CREs in various processes of change across a number of languages (e.g.Kroch 1989;Santorini 1993;Pintzuk 1995;Kallel 2005;Pintzuk and Taylor 2006;Kallel 2007;Fruehwald et al. 2009;Postma 2010;Durham et al. 2012;Wallage 2013;Gardiner 2015).Henceforth, we shall refer to it as the standard operationalization.1 Fig. 2 Three logistic curves (1) with identical s ('slope') parameters but differing k ('intercept') parameters

The non-linking problem
Initially, the logistic function (1) was adopted because of its practicability and its success in other disciplines such as population genetics, not because it followed from any established first principles: [G]iven the mathematical simplicity and widespread use of the logistic, its use in the study of language change seems justified, even though, unlike in the population genetic case, no mechanism of change has yet been proposed from which the logistic form can be deduced.(Kroch 1989:204) The logistic has since been derived from mathematical models of language acquisition independently by Niyogi and Berwick (1997) and Yang (2000); Ingason et al. (2013) provide a particularly clear illustration of how syntactic acquisition in successive generations can give rise to logistic change at population level.What has never been explicated in detail, however, is why different contextual reflexes of a single underlying change should be governed by logistics agreeing in their s parameters but freely varying in their k parameters: even though this operationalization has proved useful in gathering empirical support for the Constant Rate Hypothesis, it is not a model of the CRE itself. 2 In short, while the standard operationalization may adequately describe historical data, it fails to explain it, suggesting no mechanism for how contextual reflexes spring from underlying changes.The fact that under the standard formulation the independent contextual reflexes are not linked to each other, or to anything else, in this stronger sense we call the non-linking problem, and there are a number of reasons to believe that the problem is serious enough to warrant that the standard operationalization of CREs should be rejected.Firstly, note that fitting a number of independent logistics to a number of contexts in some data leaves variation among the k parameters entirely unexplained, even if we assume that the logistics agree in their s parameters as required by the standard operationalization.In principle, it is possible for this variation in k to be arbitrarily large, and it is therefore in principle possible to 'connect' two clearly unrelated changes-possibly separated by millennia on the time axis-as long as they happen to share the same growth rate.In principle, then, it is possible to be led to the absurd conclusion that a single change runs to completion in one context before it even takes off in another (Fig. 2).
Secondly, there are reasons to think that not all instances of logistics agreeing in their s parameters are in fact CREs in the sense that a single underlying grammatical change is being modulated in the usage of a speaker or group of speakers by constant contextual (functional, discourse-related, etc.) effects.The relevant evidence comes from studies in which the 'contextual effects' are not within-speaker but between-speaker effects or even outright contingencies.Wallenberg (2016) shows that relative clause extraposition is a gradually declining option across the histories of both English and Icelandic, and that the s parameters of the two curves do not differ significantly.Similarly, Willis (2017), in his study of the spread of the innovative second-person pronoun chdi in the recent history of Welsh, finds that in different regions of Wales the change is more or less advanced (i.e.different intercepts) but that the slopes of the changes are not significantly different.Corley (2014) tests for a CRE in the usage of negative concord between female and male speakers of Early Modern English, using data from Nevalainen and Raumolin-Brunberg (2003), and again finds no significant difference in slopes.What these case studies show is that the s parameters of two different changes may be similar for reasons other than being reflexes of a single abstract grammatical pattern, and thus that identity of slope parameters is not a sufficient condition for the assumption of a single underlying change.Wallenberg (2016:e244), for instance, notes explicitly that these are different populations, and suggests that the similarity of slopes may indicate that "the same forces are underlying the change" in both the English and the Icelandic populations-but however these forces are to be understood, we cannot be dealing with a CRE in the traditional sense, as all these authors recognize. 3hese two problems are in most cases only technical in the sense that a researcher will usually have independent reasons for ruling out such fantastical hypotheses: in particular, the inference that two apparently separate changes are reflexes of the same underlying phenomenon is usually motivated by a particular structural analysis which is arrived at on independent grounds.However, the theoretical importance of these problems is great: they demonstrate that the standard independent logistics formulation of the CRE can serve at most as a proxy to CREs, not as a model of them.If the CRE is a phenomenon-and the empirical support gathered for it over the last three decades suggests it is-this means we have so far failed to model one of the more well-established facts about language diachrony.Consequently, we have only a very approximate understanding of the dynamics of language change in the presence of contextual factors, and a number of questions remain wide open: if underlying changes are to be thought of as competition between two or more parametric options or grammars, and if CREs are thought to appear because of some sort of performance effects operating over that process of competition, how, exactly, do the two processes interact?What role does the magnitude of the performance effects play in the overall change?Could contexts that favour the innovatory variant be so favouring as to accelerate the change, and if so, can this accelerating effect be quantified and measured?Similarly, could disfavouring contexts slow the change down?Could they even block change in certain cases?These questions can only be answered with the help of mathematical models of change that accommodate mechanisms for both grammatical competition and contextual effects, and also define, without equivocation, the possible interactions between these two mechanisms over time.
The non-linking problem has, of course, not gone unnoticed in the literature.As Roberts (2007) puts it: One might wonder why [the CRE] should hold.It is unlikely to be a fact about the grammars themselves.Instead, it is plausible that it may be a fact either about speech communities or about the ways in which individuals choose among grammars available to them.As such, it may be attributable to sociolinguistic factors or to the dynamics of populations, or both factors acting in tandem.(Roberts 2007:313) Our aim in this paper is to propose a solution to the non-linking problem, and our concrete proposal is that the CRE occurs because of context-specific production biases which serve either to promote or to hinder an underlying change in progress.That is to say, we will argue that the CRE is indeed a fact about the ways in which individuals choose among grammars available to them, and propose a rigorous mathematical model of this kind of speaker behaviour.The result is a first step towards a mechanistic model of the CRE that not only describes the diachronic phenomenon but explains it by deriving it from independently plausible first principles of language acquisition and use.

Plan
The paper is structured as follows.In Sect.2, we augment Yang's (2000) mathematical model of grammar competition with production biases to account for variability across contexts.This results in a dynamical system in which the evolution of the underlying change-a parameter switch for us-and the evolution of the usage frequencies in a number of linguistic contexts feed into each other iteratively.We then derive analytical expressions for the time evolution of this system and show in Sect. 3 that in most cases it can be approximated by a constrained set of equations based on one logistic.In Sect.4, this approximation is used to derive a theorem concerning the possible temporal separation between two reflexes of one underlying change: we show analytically that under our proposed model the time separation between contexts always has a finite upper bound which is inversely proportional to the rate of the underlying change.This solves the problem of unconstrained variation in the k parameters: under our model, it is no longer possible for the curves of different contexts to be radically distant from each other in time.In the remainder of Sect.4, we proceed to test the model empirically from two complementary angles: (1) by investigating whether time separations observed in a number of previously established CREs agree with the predictions of our time separation theorem, and (2) by testing whether our model is able to distinguish actual CREs from pseudo-CREs, that is, surface changes that proceed at similar rates accidentally but that are clearly not reflexes of one and the same underlying change.
A side product of this investigation is an extension of some of the analytical results in Yang (2000).In Sect.3, we uncover all possible outcomes of the dynamical interplay between grammatical competition and production biases.A full bifurcation analysis of the two-grammar case shows that production biases can both induce change in settings where Yang's (2000) model outlaws change, and block change in settings where Yang's (2000) model predicts change.On the assumption that a model which incorporates the possibility of production biases is more realistic than one that does not, then, the assumption that language change is driven (solely) by distributional differences in the proportion of sentences parsed by different competing grammars is shown to be too simple.A theorem resulting from our analysis of the extended model shows that, when production biases are in operation, such differences are neither necessary nor sufficient for change, though they continue to play an important role in any given change process in a way that can be quantified exactly.In Sect.5, we offer a brief account of the nature of production biases; Sect.6 concludes.

Learning competing grammars
Empirical work on language variation and change has demonstrated the limitations of the traditional view of parameter setting as a once-and-for-all process which leaves the learner with a unique grammar at the point of maturation: speakers have, at least during periods of change, access to more than one grammar (Kroch 1989(Kroch , 1994(Kroch , 2000;;Santorini 1992;Pintzuk 2003).As pointed out by Santorini (1992:619), this intra-individual co-existence of multiple grammatical systems is "an ability for which the phenomena of multilingualism, diglossia and intrasentential code-switching provide independent and incontrovertible evidence"; see also the discussion in Roberts (2007:319-331).This notion has been formalized by Yang (2000Yang ( , 2002) ) in his mathematical model of competition-driven change, on which our model of the CRE is based.We therefore begin by reviewing the operating principles behind this model, focussing on the presentation in Yang (2000).
This model construes language change as a learning process in a homogeneous, well-mixing population with non-overlapping generations.At each iteration, we can therefore think of the population as a single individual who sets parameters based on the linguistic output of the previous generation, abstracting away entirely from the social and geographical structure of that population.In the competing grammars framework, each biologically possible grammar of human language G i is associated with a weight which gives the probability of an individual using that grammar.The Fig. 3 Venn diagram of the sentences generated by G 1 (the set L 1 ∪ L X ) and by G 2 (the set L 2 ∪ L X ).Here, L 2 represents a greater proportion of sentences than L 1 , and so the advantage of G 2 is greater than that of G 1 framework allows any number of those grammars to compete; however, during wellstudied and relatively well-understood periods of language change, it usually seems to be the case that two grammars are in competition.Since, additionally, this renders the mathematics of the model particularly tractable, we focus on the two-grammar case in all that follows.
Let G 1 and G 2 be these two grammars, and denote their weights with p t and q t , respectively, indexed for generational time t. 4 The basic insight behind Yang's (2000) model is that each grammar has its time-independent (parsing) advantage, which is simply the proportion of sentences the other grammar cannot parse (out of all sentences generated, in abstracto, by either grammar).There are then fundamentally three kinds of sentence: sentences of type L 1 , which G 1 but not G 2 parses; sentences of type L 2 , which G 2 but not G 1 parses; and sentences of type L X , which both grammars parse (Fig. 3).The language learner receives primary linguistic data (PLD)-the linguistic output of the generation at time step t-and his task is to arrive at weights p t+1 and q t+1 for the two competing grammars in his own generation.Letting α denote the advantage of G 1 and β that of G 2 , then (assuming he samples his environment uniformly) the learner is confronted with a number of sentences drawn from the following distribution: Based on this input, the learner is assumed to set parameters in accordance with linear reward-penalty learning, an off-the-shelf learning algorithm from mathematical psychology (Bush andMosteller 1951, 1958;Narendra and Thathachar 1989).Of crucial importance here are the two quantities c t = βq t and d t = αp t = α(1 − q t ), known as the penalty probabilities of the two grammars: c t is the probability of the learner encountering a sentence which G 1 cannot parse and d t the probability of a sentence which G 2 cannot parse.It can be shown (Narendra and Thathachar 1989:162-163) that, if the learner's training sample is large enough, eventually he ends up with a weight q t+1 which is well approximated by Assuming c t = 0 and d t = 0 without loss of generality, this equation may be reduced to the more useful form where we write ρ = α/β for the ratio of the parsing advantages.Equation (4), then, relates the grammar weights of the (t + 1)th generation to those of the tth generation, thereby defining the inter-generational or diachronic dynamics of a sequence of (reliable) linear reward-penalty learners. 5 It follows that α < β, or ρ < 1, is a sufficient condition for grammar G 2 to overtake grammar G 1 : Theorem 1 (The Fundamental Theorem of Language Change; Yang 2000:239) Assume reliable learners, so that (4) holds.Then In other words, the grammar with the greater parsing advantage will necessarily win out in the long term.The difference equation ( 4) may in fact be solved for t to yield where q 0 is the weight of G 2 at the point of actuation of the change (Appendix A.1, Corollary 4): hence as soon as the value of q 0 is known, the entire change trajectory can be predicted.Furthermore, it is not difficult to show that this solution is equivalent to with s = − log(ρ) and k = − log(ρ) −1 log(q −1 0 − 1).Thus, assuming that learners receive representative samples of their linguistic environments, a diachronic sequence of such learners exhibits logistic evolution.In particular, the slope of the trajectory is directly dependent on the advantage ratio ρ such that the smaller ρ (the more advantageous G 2 is), the faster the change from G 1 to G 2 , and vice versa. 5For two competing grammars, the linear reward-penalty learning algorithm assumes the following form for learning rate 0 < γ < 1 (see Yang 2000 for more details).Assuming that the learner's initial guess for the weight of grammar G 2 is Q 0 = 0.5 (no a priori bias), then, for input sentence s = 1, . . ., N, the learner picks G 2 with probability Q s−1 (and G 1 with probability 1 − Q s−1 ), attempts to parse the sentence, and sets Thus, Q is increased with successful parsing events and decreased with unsuccessful parsing events.Finally, we set q t+1 = Q N .Under the simplifying assumption that N → ∞, the learner does not have to contend with a finite dataset or a critical period.It is of course false, but like much work in learnability and modelling we adopt it here in order to derive analytical approximations such as (4) which would otherwise be difficult, if not impossible, to derive.This approximation holds in the following sense: q t+1 converges to a normal distribution with mean q t+1 = c t /(c t + d t ) and a variance which tends to 0 as γ → 0 and Nγ → ∞ (Narendra and Thathachar 1989:162-163).Assuming a finite learning sample would introduce a stochastic component (noise) to the system, and exploring the consequences of this falls beyond the scope of the present paper.This is the gist of the competing grammars model of language change; for more details, see Kroch (1994), Yang (2002), Pintzuk (2003) and especially Heycock and Wallenberg (2013), who apply the model to a concrete case study involving the loss of verb movement in Scandinavian.

Competing grammars and contextual biases
To account for contextual effects and the CRE, we now assume the existence of K linguistic contexts 1, . . ., K, with each sentence generated by G 1 or G 2 belonging to one and only one of these contexts. 6Each context i is equipped with a context weight λ i that gives the proportion of sentences that fall in that context (out of all sentences generated by either G 1 or G 2 ); clearly, since we are dealing with proportions, we require λ 1 + • • • + λ K = 1 (Fig. 4).In addition to these weights, each context is associated with a fixed (constant over time) production bias b i which can be positive, negative or zero.In the first case, the context favours G 2 ; in the second, it favours G 1 ; and in the third case, the context is neutral with respect to the two grammars. 7 Now consider a language learner acquiring his grammar weights based on the output of generation t of speakers.With Yang (2000), we assume that the tth generation has internalized grammar weights p t and q t .Where our treatment diverges is the effect these weights have on the language acquisition process of the (t + 1)th generation.Rather than assuming that p t and q t feed directly into the acquisition process in the following generation, we assume that speakers of the tth generation may promote or demote the two weights p t and q t in different linguistic contexts in different ways, subject to the context-specific production biases b i .It is then on the basis of this usage, modulated by the contextual biases b i and the context weights λ i , that the next generation of learners must infer their grammar weights.
Letting q (i) t denote the probability with which a speaker of the tth generation uses grammar G 2 in context i, and similarly for p 6 Formally, this means that the contexts constitute a partition of the set L 1 ∪ L X ∪ L 2 in the usual settheoretic sense: the contexts are pairwise disjoint subsets of L 1 ∪ L X ∪ L 2 and their union equals the whole of L 1 ∪ L X ∪ L 2 .
7 Associating positive production biases with a favour for G 2 over G 1 (rather than G 1 over G 2 ) is but a convention and does not affect the dynamics of our model: reverting the biases would merely swap the labels of the two grammars.
Fig. 5 The biasing functions F and G have to satisfy the requirement |F |, |G| ≤ min{p t , q t } = min{q t , 1 − q t } (see Appendix A.2 for a proof) and thus land in the shaded region of this plot.The parabolic curve shown here gives the most parsimonious such upper bound, the product where F and G are some (yet undetermined) functions which modulate the effect of the bias b i on production.These functions must satisfy two requirements: and the following is a theorem.
Theorem 2 Functions F = F (b i , p t ) and G = G(b i , q t ) satisfy the conditions ( 7) and (8) if, and only if, they satisfy Theorem 2 thus implies that the functions F and G are necessarily the additive inverse of each other, and that their absolute value is necessarily bounded from above by the minimum of p t and q t .Technically an infinite number of functions satisfy this pair of conditions, so we need to ask what these functions actually are (Fig. 5).The simplest, most parsimonious choice is to consider the product p t q t , which is guaranteed to be bounded from above by both p t and q t whenever 0 ≤ p t , q t ≤ 1.In other words, we suggest setting where the contextual biases b i range from −1 (maximally G 1 -favouring) through 0 (neutral) to 1 (maximally G 2 -favouring).This choice for the functions F and G has a number of intuitively satisfying features.For example, (10) implies that if either p t = 1 or q t = 1, then F = G = 0 (since in the first case q t = 0 and in the second case p t = 0 and consequently p t q t = 0) and no biasing will apply.Empirically, this means that if a grammar has been acquired categorically, no contextual biases will be able to skew usage in the direction of the other grammar.This is intuitively right: if a grammatical option has been acquired categorically, then by definition the competing option does not exist for the speaker and no grammar-external biasing ought to be able to apply.This behaviour of our biasing mechanism in the limits q t → 1 and q t → 0 is just one manifestation of a more general feature of the model: that while the biases b i themselves are constant and do not change over time, the magnitude of the effect of these biases on usage does depend on the state of the underlying change: the effect is the strongest midway through the change (from Fig. 5, we see that the effect is the strongest when q t = 0.5) and tails off to zero in the limits q t → 1 (completion) and q t → 0 (actuation).As we will see in Sect.4, this is what the empirical data also show. 8ith ( 11) in place, it is possible to work out the diachronic, inter-generational dynamics of our model (assuming, again, that learners receive large input samples).What generation t outputs in this extended model is not the distribution given in ( 2), but a combination of grammar advantages (α and β), grammar weights (p t and q t ), context weights (λ i ) and context biases (b i ).The penalty probability for grammar G 1 now becomes λ i b i p t q t = β(q t + Bp t q t ), (12) where the index i runs through the contexts i = 1, . . ., K and where we write B = K i=1 λ i b i for convenience. 9The quantity B, which may be regarded as the net bias operating on the language acquisition process weighted by the context proportions λ i , turns out to be a decisive quantity in our model: from (12), we immediately see that if B = 0, the penalty c t reduces to the Yangian penalty c t = βq t .Our model, then, generalizes Yang's (2000) model and reduces to the latter in the special case that the contextual biases are 'in balance'-if either all the biases are zero or if G 2 -favouring (positive) biases cancel out the effect of G 1 -favouring (negative) biases.
Entirely symmetrically, the penalty for grammar G 2 reads Assuming reliable learners, we may now use these penalty probabilities to write down the inter-generational difference equation that relates q t+1 to q t for the extended Fig. 6 Inter-generational change in Yang's (2000) model (top) and our model (bottom).After parameter setting, the learner ends up with a weight q t for grammar G 2 (and p t for grammar G 1 ).In our model, this weight is then attenuated in production by the context-specific production biases b i so that the actual probability of using G 2 in the ith context is q (i) t = q i +b i p t q t (see text for details).This biased probability, together with the advantage ratio ρ = α/β and the context weight λ i , then determines the PLD for the following generation model: equation ( 4) becomes Recalling that p t = 1 − q t , this may be written as where ρ = α/β as before and

The Constant Rate Effect
To summarize, we propose to augment Yang's (2000) model of grammar competition with a set of production biases b i which modulate the grammar weights p t and q t in actual linguistic production.This modulation is implemented by a mechanism which, we have shown, has to operate within certain analytical bounds.Within those bounds, we have suggested that the most parsimonious mechanism be adopted, corresponding to our particular choice of the bias-modulating functions F and G, as explained above.The diachronic behaviour of this extended model is characterized by equations ( 11) and ( 15): the difference equation ( 15) gives the evolution of the underlying grammar weight q t , whilst equation ( 11) supplies the context-specific value of this probability, modulated by the contextual production biases.The flowchart in Fig. 6 illustrates the inter-generational dynamics that result from this mechanism, comparing our extension of Yang's (2000) model to the original.Before moving on to an empirical evaluation of our proposed model, we first ask whether it produces, in broad qualitative terms, the right kind of behaviour.To this Fig.7 The behaviour of our model with two different sets of production biases, for advantage ratio ρ = 0.5 and initial value q 0 = 0.01 (1% usage of G 2 at the point of actuation): the evolution of both the underlying probability q t (•) as well as that of the contextual usage probabilities q (1) t (5), q (2) t (×) and q (3) t ( ) is shown up to q t = 1 − q 0 = 0.99.In each case, the context weights are set at λ 1 = 0.
The two negatively-biased contexts ( and ×) outweigh the one positively-biased context (5), and as a consequence, the change from G 1 to G 2 takes much longer than in (a).The trajectory of q t is also not strictly logistic in this case, as is evident from the fact that it is not symmetric about the midpoint q t = 0.5: passage from q 0 = 0.01 to q t = 0.5 takes longer than passage from q t = 0.5 to the final value q t = 0.99 end, Fig. 7 shows the behaviour of our model in two different situations involving three arbitrarily chosen contexts: in a situation in which the contextual biases are in balance and cancel each other out (B = 0; Fig. 7a), and in a situation in which the net effect of biases in favour of the conventional variant G 1 conspire against the propagation of the innovative variant G 2 (B < 0; Fig. 7b).Impressionistically, our model produces a CRE in both cases: the probability of use of G 2 increases roughly at the same rate in each context, with a characteristic temporal shift between the propagation curves of the individual contexts.This suggests that our model is able to replicate the central intuition of Kroch (1989) that different reflexes of one underlying change ought to proceed at similar rates, and that the output of our model can, in principle, approximate the empirical situations that have been suggested as CREs in the literature.
In these two cases, the evolution of the underlying probability q t is different, however, because of the different biasing that applies in each case.In Fig. 7a the contexts are 'in balance' (B = 0), which by the preceding analysis implies that the evolution of q t itself is logistic.In Fig. 7b, on the other hand, G 1 -favouring biases outweigh G 2 -favouring biases (B < 0), hindering the propagation of G 2 .This is reflected in the fact that the evolution of the underlying q t is slowed down.Even though G 2 still overtakes G 1 in the limit, the trajectory of q t is no longer strictly logistic (careful examination shows that it is not symmetric about the midpoint q t = 0.5, but rather exhibits slower change for q t < 0.5 and faster change for q t > 0.5).This motivates us to consider extreme model parameter regimes, particularly the subspace where B is negative, in more detail.
In equation ( 15), the factor t depends on q t whenever B = 0.This complicates the analysis of the extended model significantly: while the Yangian equation ( 4) can be solved for t to yield the logistic function, we are not aware of a closed-form solution to the more complex nonlinear difference equation ( 15) except in the singular case B = 0, where the equation reduces to (4).This has the undesirable practical consequence that there is no trivial way of fitting our model to data-lacking a closed-form curve for the underlying probability q t from which to derive curves for the contextual reflexes q (i) t , there simply are no closed-form contextual curves to fit.The best one can do is to iterate the model for various choices of model parameter values and initial conditions and compare the resulting trajectories against empirical data, an approach which soon becomes computationally prohibitive as the number of logically possible model parameter combinations grows as a superlinear function of the number of model parameters.To tackle this problem, we will in the next section conduct a full analysis of the behaviour of our model in the limit t → ∞ and show that, under most empirically meaningful combinations of model parameter values, the underlying trajectory q t is well approximated by a logistic curve.Thus, even though we cannot write down the solution of q t for arbitrary times t, and even though we know that for some parameter values (such as when B < 0) the evolution of q t is not logistic, we can use logistic functions to approximate the true value of q t .This will form the basis of our curve-fitting procedure in Sect. 4. A reader who is willing to skip the technicalities of the logistic approximation may advance straight to Sect. 4.

Advantage versus bias
As we have noted above, the Fundamental Theorem of Yang's (2000) model is that a more advantageous grammar will necessarily overtake a less advantageous one: if ρ < 1 (α < β) and learners are reliable, then q t → 1 as t → ∞, and thus grammar G 2 overtakes G 1 (Theorem 1).A nontrivial consequence of extending the model with production biases is that this theorem no longer holds: a difference in the proportion of input parsed by the two competing grammars is neither sufficient nor necessary for language change.While this is a minor observation from the point of view of the CRE, which is the main focus of the present paper, the failure of the Fundamental Theorem under suitable combinations of grammar advantages and production biases is an interesting finding from the vantage point of the theory of language change in general, and we will therefore pursue it briefly in this section.The bifurcation scenario here outlined will also play a role in the logistic approximation that we develop in the following subsection for model evaluation purposes.
The production biases b i can be positive, negative or zero.In the first case, the context in question favours G 2 over G 1 ; in the second case, G 1 is favoured; and in the third case, the context is neutral.The scalar product B = K i=1 λ i b i of context weights and production biases turns out to play a critical role in determining how, and if, change from G 1 to G 2 happens.If there are negatively biased (G 2 -disfavouring) contexts, and if their share of all sentences in the language learner's PLD is large enough, change from G 1 to G 2 can be blocked even if the advantage of G 2 is greater than the advantage of G 1 .On the other hand, if there are sufficiently strong positively biased (G 2 -favouring) contexts, G 2 may overtake G 1 even if the latter's advantage exceeds that of the former.A critical value B c of the net bias B in fact exists such that change from G 1 to G 2 is guaranteed whenever B > B c but is blocked whenever B ≤ B c : Theorem 3 (The Extended Fundamental Theorem of Language Change) Assume reliable learners, so that (15) holds.Let q 0 be the weight of grammar G 2 at the point of actuation, let B = K i=1 λ i b i , and let Then In other words, G 2 overtakes G 1 if, and only if, B > B c .
In dynamical-systems terminology, the production bias mechanism induces a bifurcation in the parameter space of the extended model: small tweaks made to either the biases (b i ) or to the proportion of input falling in each context (λ i ) can alter the trajectory of language change entirely by determining which of the two grammars will win out (Fig. 8).An immediate consequence of Theorem 3 is that if no context is negatively biased, G 2 will overtake G 1 whenever ρ < 1: Proof Since B c < 0 for any choice of q 0 , if ρ < 1.
Even though production biases, then, can induce or block change in parameter regimes where such behaviour is impossible in Yang's (2000) original model, there are limits to how much of an effect the biases can have over grammar advantages.Briefly put, if G 2 is much more advantageous than G 1 (0 < ρ 1), then no amount of negative bias can block change, and, on the other hand, if G 1 is much more advantageous than G 2 (ρ 1), no amount of positive bias can make G 2 overtake G 1 .How much is much depends on the boundary condition q 0 : Corollary 2 If ρ < q 0 /(1 + q 0 ), then q t → 1 as t → ∞, regardless of the value of B. If ρ > (2 − q 0 )/(1 − q 0 ), then q t → 0 as t → ∞, regardless of the value of B.
Fig. 8 The outcome of language change for different combinations of advantage ratio ρ and net bias B, for two different initial weights for G 2 : q 0 = 0.001 (left) and q 0 = 0.9 (right).The thick curve corresponds to the subset of this parameter space where B = B c , the critical bifurcation value.If B > B c , grammar G 2 overtakes; if B < B c , grammar G 1 prevails; and if B = B c , the system falls in an equilibrium where q t = q 0 for all t (Theorem 3).Yang's (2000) model corresponds to the dashed line running at B = 0 (no contextual effects, or contexts wholly in balance) and thus predicts that G 2 wins for any 0 < ρ < 1 and loses for any ρ > 1.Since B has both a lower and an upper limit (−1 ≤ B ≤ 1), the advantage ratio ρ has critical values, dependent on the boundary condition q 0 , such that if ρ lands beyond one of these values, no amount of out-of-balance bias can overthrow the advantage-induced dynamical outcome (Corollary 2).In the figure on the left, for instance, any ratio ρ greater than about 2 guarantees G 2 to fail.In the figure on the right, any ratio ρ smaller than about 0.5 guarantees G 2 to succeed, no matter what the combination and magnitude of contextual biases Figure 8 illustrates.

Logistic approximation
The above results show that the outcome of grammar competition in the presence of context-specific production biases is determined by a complicated interaction between these biases (b i ), the proportion of input that falls in each context (λ i ) and the ratio of the parsing advantages of the two competing grammars (ρ).This is because in our model the language acquisition mechanism and the production bias mechanism constitute a feedback loop across iterated applications over multiple generations of language learners, the production biases modulating the acquisition of the grammatical weights p t and q t .As we noted in Sect.2.3 (Fig. 7), this feature of the model also implies that when the production biases are particularly strong, they will cause the evolution of the underlying grammar weights to be non-logistic.The feedback loop gives rise to a nonlinear difference equation for which we have no solution in the general case, and the following problem immediately arises: how can the predictions of our model be tested against empirical data if there is no closed-form curve which to fit?
Even though the evolution of q t is, strictly speaking, logistic only when B = 0, eyeballing trajectories such as the one in Fig. 7b suggests that these trajectories are Fig. 9 (a) Error of fit (sum of squared residuals; nonlinear least squares regression) of a logistic function to trajectories of the underlying probability q t generated by our model for various combinations of advantage ratio ρ and net bias B, for initial condition q 0 = 0.01 (1% usage of G 2 at the point of actuation).The dashed vertical lines give the critical value B c of the bifurcation parameter for each selection of ρ.We find that trajectories of q t are closely approximated by logistics except in the immediate vicinity of the bifurcation threshold B c at which change from G 1 to G 2 is blocked.(b-c) Best-fitting slope (s) and intercept (k) values found by these regressions still S-shaped and perhaps well approximated by logistics.To explore this possibility, we performed a sweep across the model parameter space, generating trajectories of q t from the initial condition q 0 = 0.01 (1% usage of G 2 at the point of actuation) in the regime B > B c (i.e. in the parameter regime where G 2 is guaranteed to oust G 1 by Theorem 3), until q t had reached the value q t = 1 − q 0 = 0.99.We then proceeded to fit a logistic curve to each of these trajectories in order to investigate how well the trajectory may be approximated by a logistic.Figure 9a gives the errors of these fits, showing that the trajectories are closely approximated by logistics whenever ρ is not too large and B is not too close to the critical bifurcation threshold B c .Figures 9b-c supply the best slope (s) and intercept (k) coefficients found by these regressions.We find that s is a decreasing function of ρ and an increasing function of B: the more advantageous G 2 is, and the more G 2 is favoured by the production biases, the steeper the underlying change, as one would expect.The intercept coefficient k, in turn, is an increasing function of ρ and a decreasing function of B: the less advantage G 2 has and the more the production biases tend to disfavour G 2 , the more the curve of the underlying change is shifted towards positive time.

The Time Separation Theorem
Under the logistic approximation from Sect. 3, the usage of grammar G 2 in the contexts i = 1, . . ., K is described by a set of K equations where qt is a logistic function approximating the true underlying probability q t .What historical language corpora give us are usage frequencies in various contexts, and we therefore wish to fit curves of the form ( 18) to such data.The fact that under the logistic approximation all such curves are tied to qt , which itself has a closed-form solution, now facilitates this empirical evaluation: even though the individual context curves q (i) t themselves are not logistic (unless b i = 0), they are easily derived from one that is.In what follows, we will take a look at a number of case studies, fitting context curves with the help of a nonlinear least squares optimization algorithm.
Estimating the goodness of fit of these regressions is one important goal of this exercise.However, our main aim is to solve the non-linking problem identified in Sect.1.2.Specifically, we wish to demonstrate that our model does not allow arbitrarily large time separations between contexts, operationalized as the difference between the points in time at which different context curves reach their tipping point, or the point in time at which the context frequency of the overtaking grammar equals 0.5 when ( 18) is generalized for real-valued t.The logistic approximation gives us a straightforward proof of this.
Theorem 4 (The Time Separation Theorem) For any two contextual reflexes of an underlying change from G 1 to G 2 approximated by a logistic qt with slope s, the maximal time separation at tipping points is It is to be noted that (s) is inversely proportional to s-the slower the rate of change, the more time separation is allowed between any two contexts and vice versa.
To fit the system (18) to a set of data points, we first define reasonable ranges of variation for the s and k parameters of the logistic qt that we wish to probe.We then loop through the values contained in these ranges, finding the best fitting bias parameters b i for each pair (s, k) using a nonlinear least squares optimization algorithm such as the Gauss-Newton procedure (Bates and Watts 1988), bearing in mind the bounds −1 ≤ b i ≤ 1.Finally, out of all these regressions, we pick the combination of s, k and b i that provides the best fit to the data in question.The whole procedure is detailed in pseudocode in Appendix A.5. 10We now proceed to an evaluation of the model by comparing its predictions against three historical changes for which a CRE has been reported in the literature: the emergence of periphrastic do in the history of English (Kroch 1989), the earliest stages of the English Jespersen Cycle (Wallage 2013), and, to take a phonological example to illustrate the generality of the procedure, the loss of final fortition in Early New High German (Fruehwald et al. 2009).In Sects.4.2-4.4we first briefly summarize the linguistics of each change, reproduce the relevant empirical data, and visualize the fit of our model to the data when the regression is conducted using the procedure outlined above.In Sect.4.5, we take a more quantitative angle and report the numerical errors of these fits, comparing them to the errors that an application of the standard procedure based on individual logistics (cf.Sect.1.1) would produce.Finally, in Sect.4.6, we take a look at a pseudo-CRE-a case where the standard independent logistics operationalization reports a CRE but where this conclusion is patently absurd from other considerations (cf.Sect.1.2)-in order to investigate whether or not our model, too, is prone to report false positives in such cases.

Periphrastic do in English
The first case study we will consider is perhaps the best known instance of a CRE: Kroch's (1989) interpretation of Ellegård's (1953) data on the rise of periphrastic do in Early Modern English.The variable in question is whether a form of do is used in certain contexts, as in (20a), or not, as in (20b) (examples from Kroch 1989:216).
'Where does the Green Knight hold him?' b.How great and greuous tribulations suffered the Holy Appostyls . . .?
'How great and grievous tribulations did the Holy Apostles suffer?' In modern standard English, a form of do is required in a number of contexts, including all interrogatives as well as negative declaratives.What Ellegård's (1953) data show is that, on the surface, the use of do appears to 'take off' in the different contexts at different rates: for instance, between around 1500 and 1650, negative questions exhibit a much higher proportion of do than affirmative wh-object questions or negative declaratives, though the latter contexts eventually catch up (Table 1).Kroch (1989) conducted a regression on these contexts and showed that logistic curves fitted to Table 1 Proportion of do in five different contexts: negative declaratives, negative questions, affirmative transitive questions, affirmative intransitive questions, affirmative wh-object questions.From Kroch (1989:224,  them individually did not differ much in their slope (s) parameters, although they did differ in terms of the intercept (k) parameter.This particular example, while perhaps the most celebrated instance of a CRE, is in fact not the most straightforward instance found in the literature, primarily due to a 'dip' in the later portion of the change in some contexts, which makes the progression of do non-monotonic; see Warner (2005) and Ecay (2015) for detailed discussion, concluding that other factors (and possibly another grammar) are at play.Kroch (1989) also identifies the dip and consequently focusses on the first seven data points of Ellegård's (1953) data only; we follow his practice here.
When the algorithm from Appendix A.5 is used to fit a model of the form (18) to these data, the picture in Fig. 10 emerges.On visual inspection, the fit is a good one.Crucially, our model is constrained to allow only so much time separation between any two contexts of one change (Theorem 4), illustrated in Fig. 10 as the horizontal bar extending both ways from the tipping point of the curve of the underlying grammar probability.We find that fitting the model to Ellegård's (1953) data drives two of the context curves (negative questions and affirmative object questions) to the very extremes of the range licensed by the model-in other words, for these contexts, the production biases need to be maximal in order for the model to fit the data.Crucially, however, the data are described well by the regression curves so obtained, an observation which we back up quantitatively in Sect.4.5.

English Jespersen Cycle
Our second case study, also from the history of English, involves the replacement of preverbal ne/ni by postverbal not during the Middle English period.This change Fig. 10 Fit of our model to the data on English periphrastic do (curves: model; points: data from Table 1, first seven periods).On visual inspection, the fit to each context is a good one.Theorem 4 implies a maximal time separation, illustrated here as the horizontal bar extending both ways from the tipping point of the theoretical curve for the underlying grammatical change (no production biases).The best-fitting parameters found by the regression are s = 0.031, k = 1547.677,with bias sizes b i as follows: −0.885 for negative declaratives, 1.000 for negative questions, 0.656 for affirmative transitive questions, −0.647 for affirmative intransitive questions, and −1.000 for affirmative wh-object questions.With s = 0.031, the maximal time separation licensed by the model is roughly 56.(21) a.We ne moten halden Moses e lichamliche.
we NEG need observe Moses' law bodily 'We need not observe Moses' law in body.' b. ac of hem ne speke ic noht but of them NEG spoke I not 'but I did not speak of them' c.I know nat the cause.
I know not the cause 'I do not know the cause.' This replacement of negators, a cross-linguistically common diachronic pathway, is referred to as Jespersen's Cycle; see Wallage (2008) and Ingham (2013) for detailed discussion of the English development.For our purposes, the change that is important is the replacement of Stage 1 of Jespersen's Cycle-negation by ne alone, as in (21a)-with Stage 2, bipartite negation, as exemplified by (21b).Wallage (2013) shows that Stage 2 is favoured with discourse-old propositions during the middle of the change, but that a CRE obtains (Table 2).Again, on purely visual inspection, our model fits the data well, and the variation observed between discourse-old and discourse-new propositions falls, roughly, within the time bounds prescribed by the Time Separation Theorem (Fig. 11).2).On visual inspection, the fit to each context is a good one, though the poor time resolution of the data is a problem.Theorem 4 implies a maximal time separation, illustrated here as the horizontal bar extending both ways from the tipping point of the theoretical curve for the underlying grammatical change (no production biases).The best-fitting parameters found by the regression are s = −0.016,k = 1203.788,with bias sizes b i as follows: −1.000 for discourse-old propositions and 1.000 for discourse-new propositions.(Note that in a case like this where the slope s is negative, a negative context bias means a preference for the overtaking grammar, whereas a positive bias indicates preference for the receding one.)With s = −0.016, the maximal time separation licensed by the model is roughly 110 years

Loss of final fortition in Early New High German
CREs are not found only with syntactic variables.Fruehwald et al. (2009) reanalyse data from Glaser (1985) on the loss of final fortition in (Bavarian) Early New High German, which is observable in orthographic variation of the period, e.g.tak vs. tag 'day (acc.sg.)', rat vs. rad 'counsel (acc.sg.)'.They argue that the orthographic variation clearly represents a phonological change in progress rather than shifting scribal tradition, and that fortition is the result of a single phonological rule whose loss is visible to different degrees in different contexts during the period of the change: /d/ exhibits fortition the most and /g/ the least, with /b/ showing an intermediate pattern (Table 3).Our model describes the data well, with the observed CRE again falling within the time bounds implied by the model (Fig. 12).This example illustrates that even though our model is based on the variational learner in Yang (2000), which is essentially a hypothesis about parameter setting in syntax, the logistic approximation (18) which underlies the curve-fitting procedure   3).On visual inspection, the fit to each context is a good one.Theorem 4 implies a maximal time separation, illustrated here as the horizontal bar extending both ways from the tipping point of the theoretical curve for the underlying grammatical change (no production biases).The best-fitting parameters found by the regression are s = −0.019,k = 1374.747,with bias sizes b i as follows: 0.353 for /b/, 1.000 for /d/, and −1.000 for /g/.With s = −0.019, the maximal time separation licensed by the model is roughly 93 years can legitimately be used to model CREs in any domain as long as the assumption of an underlying logistic change is justifiable.As Fruehwald et al. (2009:9) point out, "[t]he discovery of the Constant Rate Effect in phonological change is perfectly expected under normal generative theories of phonology when the mechanism of change is grammar or rule competition"-the learning algorithm a language learner uses in this case may (or may not) be different from the one assumed in Yang's (2000) model, but this notwithstanding, as long as some sort of underlying representation similar to the Yangian weights p t and q t can be assumed to exist, our production bias mechanism may be applied.

Comparison with standard procedure
Sections 4.2-4.4have adduced evidence to the effect that the model introduced in this paper can account for the CRE: the model gives fair fits to historical data even though it is constrained by the chronological bounds set by the Time Separation Theorem (Theorem 4).To make this argument more quantitatively, in this section we compare German.Generally speaking, the more constrained model defined in this paper does not fare worse than the less constrained, theoretically unmotivated standard operationalization.We suspect that the exceptionally good fit of the standard operationalization for the Jespersen Cycle is accounted for by sparsity of data (6 data points only), which means that any model that is little constrained will be favoured disproportionately these three fits to the standard operationalization of the CRE: that is, logistic curves agreeing in the s (slope) parameter but differing in their k (intercept) parameters.For each of the three case studies considered in Sects.4.2-4.4,then, we carry out two fits: one for our model, and another one for a model consisting of a set of logistic curves (one per context) where the s parameter is not allowed to vary between contexts but where such variation is allowed for the k parameter.We quantify the goodness of fit of these regressions in the usual way, by the normalized sum of squared residuals: in other words, for each context of a given change, for each time period, we calculate the displacement between the empirically attested frequency and the value predicted by the model, square this displacement, sum over all time periods and over all contexts, and divide by the number of data points.Thus, the better the fit, the closer the sum of squared residuals is to zero.Some deviance from zero is always to be expected because of the noisy nature of historical language data.However, this measure is still able to capture the difference between models which are good fits, but subject to noise, from models which are simply bad fits to the data in question.
Figure 13 shows a comparison of the goodness of fit of the two models for each of the three case studies, operationalized using the sum of squared residuals.The crucial finding is that our model, which places more constraints on the shape and placement of the regression curves, fares no worse than the standard procedure in two out of three cases: in other words, a more constrained, theoretically motivated model which generates empirical predictions (in the form of the Time Separation Theorem) performs no worse than a less constrained, theoretically unmotivated model.The exception to this are the data on the English Jespersen Cycle, where the less constrained standard formulation reports a very low error.This appears to be a result of the very small number of data points-just three time periods and two contexts-for this particular case study.Low data resolution necessarily gives a disproportionate advantage to the less constrained model over any model that incorporates more assumptions.2 and 3.The different contexts exhibit similar rates of change across the two historical changes by accident: the slope of the underlying change is −0.016 for the English Jespersen Cycle and −0.019 for Early New High German fortition (see captions to Figs. 11 and 12).This means that the standard 'same slope, different intercepts' procedure for detecting CREs in historical data is liable to produce a false positive in this case.Our model, which implies an upper bound on the time separation possible between any two contexts of one underlying change (Theorem 4), can help to diagnose a 'change' such as this as a pseudo-change

A pseudo-CRE
Above, we have shown that the proposed model can account for the CRE, in the sense that it gives good fits to three historical changes-fits which are, in two of these cases, no worse than fits conducted using the standard operationalization of CREs.It remains to be shown that introducing this more constrained model can actually solve some of the underspecification issues the standard operationalization suffers from.As discussed in Sect.1.2, the method of 'same slopes, different intercepts' is susceptible to false positives: the fact that a number of logistics agree in their s or slope parameters is insufficient evidence that a single underlying change is at hand (see Corley 2014;Wallenberg 2016;Willis 2017 for examples where the 'contexts' cannot possibly be assumed to be evidence of underlying grammatical unity).
Here, we construct a pseudo-CRE by combining the two changes investigated in Sections 4.3 and 4.4: the early stages of the English Jespersen Cycle and loss of fortition in Early New High German.As it turns out, these two changes happen to propagate at very similar paces by accident (Fig. 14).The standard operationalization of CREs is, then, expected to report a CRE between changes to English sentential negation and Bavarian phonology, a conclusion which is clearly absurd.
The fact that the two changes are separated in time, however, means that the more constrained model introduced in this paper can correctly diagnose the pseudo-CRE.Figure 15 gives the residual errors of both the present model and the standard one for this pseudo-CRE, along with the errors for the actual CREs investigated above.The pattern that emerges is striking: for each of the actual CREs, our model reports an error on the order of 0.005, whereas for the Anglo-Bavarian pseudo-CRE the model generates an error that surpasses 0.03.The standard operationalization, by contrast, reports similar errors for all changes, failing to distinguish between the pseudo-CRE and the actual CREs.
The pseudo-CRE thus sheds light on the manner in which our model constrains variation in the time dimension-a constraint that is not built into the model as a premise but that follows from first principles in the form of the Time Separation Theorem.Ultimately, the amount of time separation allowed between any two contextual reflexes of a single underlying change depends on s, the slope of the underlying logistic qt .To obtain a more intuitive interpretation of the relationship between contextual time separations and the rate of the underlying change, it is useful to convert the slope parameter into a quantity that measures the time the change needs to go from actuation to completion, using the time-to-completion calculations proposed by Ingason et al. (2013:96-97).Namely, it can be shown that for slope s, gives the time it takes for a change to proceed from initial frequency q0 to final frequency 1 − q0 , for any (small) q0 with 0 < q0 ≤ 0.5.Choosing q0 = 0.01, a reasonable choice corresponding to 1% usage of the new variant at the point of actuation, this 'inverse slope' then gives a time-to-completion of time units (e.g.years) for any slope s.Theorem 4, on the other hand, implies a maximal time separation of units between any two contexts of a change proceeding at rate s.Since 1.8/9.2≈ 0.2, this means that the maximal time separation between any two contexts of a single underlying change is roughly a fifth of the time it takes for the change to go to completion in any context individually.
This fact can be used as a heuristic to evaluate purported CREs.For the Anglo-Bavarian pseudo-CRE, for instance, the time-to-completion for q0 = 0.01 is T 0.01 (−0.0175) = 2 0.0175 log(99) ≈ 525 (25) years when calculated for slope s = −0.0175,which is the arithmetic mean of the slopes found by our regressions for the two changes previously (see captions to Figs. 11 and 12).This implies that the time separation between any two contexts should be no more than 0.2 • 525 = 105 years.On visual inspection, however, the empirical time separation between the discourse-old and /d/ 'contexts' must be at least 300 years (Fig. 14).This, essentially, is why the model is able to diagnose the pseudo-CRE.

Discussion
In this paper, we have augmented Yang's (2000Yang's ( , 2002) ) variational learner with production biases that vary by context-the first time, to our knowledge, that this has been done. 11We have used this model to make precise the important intuition of Kroch (1989) that variation between contexts in the increasing use of a new variant may, under certain circumstances, be due to the interaction of a single underlying change with fixed contextual biases.Two important issues remain to be discussed: the diachronic implications of the interaction between language acquisition and production biases, and the nature of the production biases themselves.

Which grammar wins?
Yang's (2000) Fundamental Theorem of Language Change, given earlier as Theorem 1, can be paraphrased follows: when grammars compete, the one with the greater parsing advantage will win.In Sect. 3 we have shown that this result does not hold in our model.Instead, bias and advantage together determine which grammar will the precise way in which this works is given in our Extended Fundamental Theorem (Theorem 3).In one sense this result is unsurprising: one of the great virtues of Yang's (2000Yang's ( , 2002) ) model of the learner and of diachronic change is its simplicity, and our model introduces additional complexity.It is therefore not a particular surprise that our more complex model does not yield the same intuitive generalization.On the other hand, it is not a necessary consequence of this complication that the Fundamental Theorem fails to hold.As Fig. 8 shows, if we add to our model the stipulation that contextual weights must always be precisely in balance (B = 0), then Yang's Fundamental Theorem does hold.Such a stipulation would be wholly unmotivated, as far as we are aware, and represents a more complex model than ours.Moreover, we have not, of course, shown that the Fundamental Theorem of Language Change is false-merely that it is false under the assumptions we make.Whether or not it is false, empirically speaking, depends on how well our model, and Yang's (2000Yang's ( , 2002) ) model, correspond to reality: specifically, whether a model that incorporates the effect of contextual biases as ours does is more realistic than one that does not, and more realistic than one that constrains the net bias.We think that is right, but it is likely that a full consideration of the facts of real-life acquisition and change will require a model that is substantially more complex than any that has been proposed thus far.One feature of Yang's model, with or without our extension, is that it is completely impossible for a grammar G 2 to overtake and defeat another grammar G 1 if the weak generative capacity of G 2 is a proper subset of that of G 1 ; yet a preference for exactly this kind of subset is often invoked in the context of acquisition in the form of the Subset Principle (Berwick 1985;Manzini and Wexler 1987), and, in the domain of phonology at least, retreat to the subset is a frequentlyattested diachronic pathway, since unconditioned mergers are well-attested and have precisely the effect of reducing the number of forms generated by the grammar (see e.g.Labov 1994:551).Future work will need to address these questions of realism, as well as pursuing further analytical consequences of simpler (and thus more tractable) models like this one.

The nature of production biases
Up to now we have remained mute with respect to the ontology of production biases, beyond stating that they are biases that affect production.In principle, such biases could assume a number of forms.In a word order change such as OV to VO, for instance, one possibility for interpreting the fixed biases we have proposed is as a reflection of performance pressures in the sense of Hawkins (1994Hawkins ( , 2004)).Hawkins's (2004:38) principle of Minimize Domains states that "[t]he human processor prefers to minimise the connected sequences of linguistic forms and their conventionally associated syntactic and semantic properties in which relations of combination and/or dependency are processed."This general principle is made concrete using a metric of Early Immediate Constituents (EIC), which serves to favour syntactic structures with a uniform directionality of branching.Importantly, EIC does not penalize right-branching (e.g.VO) or left-branching (e.g.OV) grammars directly, instead disfavouring individual structures with a disharmonic directionality of branching, for instance when a head-final VP is embedded under a head-initial TP.This is equivalent to a context-specific production bias in our sense.Hawkins conceptualizes Minimize Domains and EIC as principles of parsing rather than of production, but notes that there is evidence that EIC might be involved in production too (Hawkins 2004:106), and states that "if EIC can be systematically generalized from a model of comprehension to a model of production [. . .] then so much the better" (Hawkins 1994:427).Hawkins's principles have also been reformulated as principles of derivational/computational complexity by Mobbs (2008) and Walkden (2009).
In phonological change, meanwhile, the biases can be interpreted as well established articulatory phonetic effects.Final fortition, for example, is known to be more likely to apply to velar consonants than to coronal consonants and more likely to apply to coronal consonants than labial consonants (Ohala 1983), and this order of preference seems to be observed diachronically as final fortition emerged in the history of Frisian (Tiersma 1985).12These two phonological and syntactic examples are intended to give a flavour of how contextual production biases can be interpreted, not to exhaust the range of possibilities.For other changes, other biases might be necessary: for instance, in Wallage's (2013) data, discourse-old propositions favour Stage 2 of the Jespersen Cycle during the change, and the biases here could plausibly reflect Gricean maxims of cooperative communication.
The above-mentioned biases-constraints on syntactic processing, articulatory pressures, pragmatic principles-are plausibly innate in the sense that they are shared by all speakers across all languages and are not subject to change.This is why in our model definition we maintained that the biases b i be diachronically constant.Note that the logic here is not just that constant biases imply Constant Rate Effects-we are actually defending the stronger claim that Constant Rate Effects occur if, and only if, diachronically constant biases impinge on an underlying change.Time-dependent biases, or random biases, would result in change processes in which the trajectories of different linguistic contexts are not parallel to each other-something we might refer to as an 'Inconstant Rate Effect.' Having said that, it is not inconceivable that some non-innate biases are constant on sufficiently long timescales so that they may give rise to Constant Rate Effects: this will be the case when the underlying change itself is fast enough to be carried to completion within the timeframe in which the biases stay fixed.This could be true of certain sociolinguistic biases, and here the biasing mechanism of our model is in agreement with sociolinguistic work (e.g.Labov 2001:ch.9) which has found some types of sociolinguistic bias modulation to be strongest midway through the change, just as in our model (cf.Fig. 5).13

Conclusion
Building on earlier work that derives logistic evolution as a population-level property of language change (Niyogi and Berwick 1997;Yang 2000Yang , 2002)), we have provided a mechanism for the Constant Rate Effect proposed by Kroch (1989).We have done this by enriching Yang's (2000Yang's ( , 2002) ) model of acquisition and change with a contextual bias mechanism that links different context curves to a single underlying change.The work also provides a method of testing for CREs that is demonstrably superior to the traditional method of 'same slope, different intercepts,' since it is a consequence of the model that there is a fixed upper bound on the time separation of contextual curves.We have shown that this enables us to distinguish certain types of pseudo-CREs from instances in which a single underlying grammatical change is actually plausible.We have also shown that advantage, in the Yangian sense, is not the only factor at play in determining the ultimate outcome of a situation of grammatical competition, if the basic assumptions of our model hold true.
The upshot of all this is that it is now possible to test whether divergent usage frequencies in corpora across different contexts during the course of a change in fact mask a deeper underlying grammatical homogeneity, and to do so in a more restricted and principled way than has been possible to date.Crucially, the method we propose is not only methodologically superior to the standard operationalization of CRE testing: our model in fact derives the possibility of CREs, and sets tight bounds on the kind of empirically observed situation that can be said to constitute a CRE.assumed to return the bias size b i found by this regression, and the third to give the error (in terms of the normalized sum of squared residuals) of the fit.

Fig. 4
Fig. 4 Venn diagram of the sentences generated by G 1 and G 2 , partitioned by four contexts 2, λ 2 = 0.4 and λ 3 = 0.4.(a) Here the biases are b 1 = 1, b 2 = −1 and b 3 = 0.5.With these choices, B = λ 1 b 1 + λ 2 b 2 + λ 3 b 3 = 0, and consequently the positively-biased contexts (5 and ) cancel out the effect of the negatively-biased context (×), resulting in logistic evolution of the underlying probability q t (•).(b) Here the biases are b 1 = 1, b 2 = −1 and b Fig. 10Fit of our model to the data on English periphrastic do (curves: model; points: data from Table1, first seven periods).On visual inspection, the fit to each context is a good one.Theorem 4 implies a maximal time separation, illustrated here as the horizontal bar extending both ways from the tipping point of the theoretical curve for the underlying grammatical change (no production biases).The best-fitting parameters found by the regression are s = 0.031, k = 1547.677,with bias sizes b i as follows: −0.885 for negative declaratives, 1.000 for negative questions, 0.656 for affirmative transitive questions, −0.647 for affirmative intransitive questions, and −1.000 for affirmative wh-object questions.With s = 0.031, the maximal time separation licensed by the model is roughly 56.5 years

Fig. 12
Fig. 12Fit of our model to the data on loss of final fortition in Early New High German (curves: model; points: data from Table3).On visual inspection, the fit to each context is a good one.Theorem 4 implies a maximal time separation, illustrated here as the horizontal bar extending both ways from the tipping point of the theoretical curve for the underlying grammatical change (no production biases).The best-fitting parameters found by the regression are s = −0.019,k = 1374.747,with bias sizes b i as follows: 0.353 for /b/, 1.000 for /d/, and −1.000 for /g/.With s = −0.019, the maximal time separation licensed by the model is roughly 93 years

Fig. 13
Fig. 13Error (sum of squared residuals normalized by number of data points) of the fit of our model (grey) and the standard procedure (black) for the three changes examined in Sects.4.2-4.4:periphrastic do in Early Modern English, Jespersen Cycle in Middle English, and loss of final fortition in Early New High German.Generally speaking, the more constrained model defined in this paper does not fare worse than the less constrained, theoretically unmotivated standard operationalization.We suspect that the exceptionally good fit of the standard operationalization for the Jespersen Cycle is accounted for by sparsity of data (6 data points only), which means that any model that is little constrained will be favoured disproportionately

Fig. 14
Fig.14An 'Anglo-Bavarian pseudo-CRE' that attempts to combine Jespersen's Cycle in Middle English with loss of final fortition in Early New High German: data from Tables2 and 3.The different contexts exhibit similar rates of change across the two historical changes by accident: the slope of the underlying change is −0.016 for the English Jespersen Cycle and −0.019 for Early New High German fortition (see captions toFigs.11 and 12).This means that the standard 'same slope, different intercepts' procedure for detecting CREs in historical data is liable to produce a false positive in this case.Our model, which implies an upper bound on the time separation possible between any two contexts of one underlying change (Theorem 4), can help to diagnose a 'change' such as this as a pseudo-change

Fig. 15
Fig. 15 Error (sum of squared residuals normalized by number of data points) of the fit of our model (grey) and the standard procedure (black) for the three changes examined in Sects.4.2-4.4,plus the pseudo-CRE of Fig. 14.Our model correctly distinguishes the pseudo-CRE from the actual CREs vector of length K to hold best-fitting bias sizes 8: E * ← error of fit, initialized to a b 1 , . . ., b K ) ← vector of length K to hold bias sizes 13: for i ∈ {1, . . ., K} do 14: F ← REGRESS(s, k, i) s * , k * , b * , E * 27: end procedure

Table 2
Proportion of Stage 1 negation (preverbal ne) in the English Jespersen Cycle in discourse-old and discourse-new contexts.From Wallage (2013:12, Table 1) Fig. 11 Fit of our model to the data on the first two stages of the English Jespersen Cycle (curves: model; points: data from Table

Table 3
Proportion of fortition of the three plosives /b/, /d/ and /g/ in Early New High German.From Fruehwald et al.