2,052
Views
1
CrossRef citations to date
0
Altmetric
Research Article

Markov Models for Multi-state Language Change

ORCID Icon & ORCID Icon

ABSTRACT

Historical linguistics has witnessed an upsurge in quantitative corpus studies. The bulk of these studies involve the use of regression modelling. We point out a number of potential problems with this approach, and offer an alternative. For a multi-state language change, we propose a Markov model in continuous time. The major advantage of this technique, which has been used in medical contexts, is that it is especially geared towards dealing with time as a variable of interest, while it still allows one to look at the effect of several covariates. In this proof-of-concept article, we look at morphological shifts in preterites in Dutch, from 800ad to 2000ad (n = 14,314). This is a well-researched field, allowing us to investigate the performance of the multi-state Markov model.

1 Introduction

1.1 Inferential Statistics with the Variable Time

Recent decades have seen a drastic increase in the use of quantitative techniques in linguistics, to the extent that Janda (Citation2013) speaks of ‘the quantitative turn’. Indeed, textbooks on quantitative methodology such as Baayen (Citation2008a), Gries (Citation2013), Levshina (Citation2015), and Winter (Citation2019) are among the best-selling in the field, and a bibliometric investigation of linguistic journals for the period 2005–2014 (Mohsen et al., Citation2017) revealed that the two most cited articles were decidedly quantitative in their approach and dealt with the application of mixed-models to the investigation of language (Baayen et al., Citation2008b; Jaeger, Citation2008). Though some subfields have been more reluctant in embracing this trend, hardly any subfield has remained completely untouched, and historical linguistics is no exception. Though historical linguistics has always relied on archival data (absent introspection as an easy access to facts), and early efforts to digitize such data can be counted as pioneering work (Hockey, Citation2004), full-blown integration of inferential statistic methods has advanced somewhat slower than in other subbranches of linguistics, if only because textual attestations are notoriously refractory and patchy. Labov famously quipped that “historical (…) can be thought of as the art of making the best use of bad data” (Labov, Citation1994, p. 11). Today, this has changed, and articles that make use of state-of-the-art methods are not hard to come by, see Hilpert and Gries (Citation2016), Jenset and McGillivray (Citation2017) and Van de Velde and Petré (Citation2020) for overviews.

But what exactly are the state-of-the-art methods? While many techniques are in use, from simple significance tests (chi square, Fisher exact tests …) all the way to simulation studies (Baxter et al., Citation2009; Beuls & Steels, Citation2013; Blythe & Croft, Citation2012; Landsbergen et al., Citation2010; Lestrade, Citation2015; Nowak et al., Citation2001; Pijpops et al., Citation2015; Steels, Citation2016; Van Trijp, Citation2013), the main workhorse today is generalized linear (mixed)models, with time as an explanatory variable, often under multivariate control or in interaction with other covariates (see, a.o., De Cuypere, Citation2015; Fonteyn & Van de Pol, Citation2016; Gries & Hilpert, Citation2010; Petré & Van de Velde, Citation2018; Rosemeyer, Citation2016, Citation2014; Wolk et al., Citation2013; Zimmermann, Citation2017). This may seem reasonable, as regressing competing variants or frequency metrics on time can accommodate the typical S-curve that is observed in many cases of diachronic change (Blythe & Croft, Citation2012; Denison, Citation2003; Feltgen et al., Citation2017; Kroch, Citation1989; Pintzuk, Citation2003; Weinreich et al., Citation1968, p. 113).

While this approach appears to perform well, and can actually predict how the trajectory of a change proceeds, validated on a blinded part of the corpus (Van de Velde, Citation2017), they are, strictly speaking, in danger of giving too optimistic an image of the robustness of the model. The reason is that time is a peculiar variable, which does not exactly behave like other covariates entered in a multivariate model: ‘[T]he fact that time is a dynamic process provides challenges in formulating a model that are not present in settings where a typical linear or logistic regression model might be applied.’ (Hosmer et al., Citation2008, p. 2). One of the problems is that there is autocorrelation in the data, where consecutive time periods are not completely independent from one another, which can increase what in statistics is called Type I errors, by underestimating the standard error of the time variable (Cowpertwait & Metcalfe, Citation2009, p. 96). This can be solved in various ways. Baayen and Milin (Citation2010) and De Smet and Van de Velde (Citation2020a), for instance, apply a low-tech solution by adding the state of the outcome variable at a preceding time value as a covariate in the regression model. There is a burgeoning field of using methods which are specifically tailored to dealing with the temporal nature of the data, like time series analysis (Koplenig, Citation2017; Koplenig et al., Citation2016; Van de Velde & Petré, Citation2020), including Granger causality (Moscoso del Prado Martín, Citation2014; Rosemeyer & Van de Velde, Citation2020), and survival analysis (Van de Velde & Keersmaekers, Citation2020), a method widely applied in the medical sciences, engineering, and economics/sociology, where it is known as reliability analysis, duration analysis and event history analysis, respectively.

The purpose of the latter technique is to predict or model the time to an ‘event’, i.e. death of a patient, failure of a device etc. In linguistics, it can be used to assess lexical longevity, for instance, or to assess the time it takes for an incoming mutant to take over an old construction. In the latter case, the method is a viable competitor to generalized linear mixed models. By way of example, suppose we have two competing verbal suffixes (e.g. -μι vs. for the 1sg.pres in Classical Greek, or -th vs. -s for the 3sg.pres in Early Modern English), and we want to model the increase of the second variant of such a pair, we could either regress the variants as a binary outcome variable on time and some other covariates (frequency of the verb, e.g.) and maybe add a random effect for verb, or we could use survival analysis and model the time to the event, in this case the switch to the new variant, depending on some covariate, using a typical method like a Cox Proportional Hazard model (see Hosmer et al., Citation2008, pp. 67–90; Harrell, Citation2015, pp. 427–436, 475–533, for mathematical and conceptual underpinnings). In the latter case, each verb is treated as a ‘patient’, if we look at it from a medical perspective.

1.2 Real-life Complications

As often in life, matters are usually more complicated. Several problems are likely to occur: (i) some verbs may disappear from the language before they had a change to trade in the old suffix for the new suffix, by a process of lexical replacement, or by just falling out of use. (ii) Maybe there is no real transition from an old variant to a new variant, but rather a ‘hesitant’ back-and-forth trajectory, depending on the verb. Even if the majority of changes show directionality, not all of them are on a conclusive path from an old A to a new B. Variation may be much more pervasive and functional than we might be inclined to assume (De Smet et al., Citation2018; Van de Velde, Citation2014; De Smet & Van de Velde, Citation2020b). (iii) Some verbs may be scantily attested and only appear at random times, so that we do not know exactly when it has changed its inflection in the meantime. We can only observe that when the verb surfaces in the textual record, it has either changed or not changed since the last time it was seen, and we are unsure about its behaviour when it dove under the waterline. (iv) Verbs may vary at a certain point in time, depending on the author, genre, register etc. This is especially likely in times of change.

Problem (i), lexical death, is not unknown in medical science, of course, where time to an event, say developing a pathology, can be blocked by a patient’s death. A naïve solution would be to treat these patients as ‘censored’, the technical term in survival analysis for patients that are lost to follow-up, because they fail to show up at consultation or because the observational study following patients stopped at a fixed point in time. This is standardly taken into account in survival analysis. But death of a patient is different. Unlike the censored patients, a deceased patient cannot develop the pathology any more. The solution is to build a model with what is technically called ‘competing risks’ or a ‘multi-state model’. Healthy patients can either become ill or die. Dealing with competing risks is a mathematical complication survival analysis is well-equipped to cope with (Putter et al., Citation2007).

Problem (ii), back-and-forth verb trajectories, is harder, both for a generalized linear (mixed) model and for simple survival analysis. In generalized linear (mixed) models, we expect a monotonous increase or decrease over time. Solutions have been offered, e.g. in the form of adding a quadratic term to deal with U-shaped change – the idea also underlies the approach known as ‘growth curve analysis’ – or with generalized additive (mixed) models (GAMMs), for even ‘wigglier’ curves (see Winter & Wieling, Citation2016). If the change is vacillating, with a lot of back-and-forth, the models will be quickly overparametrized. In survival analysis, the solution is extending the multi-state model such that recovery is possible from one state, but not from another.

Problem (iii), data scarcity for intermediate periods, is often glossed over in general linear (mixed) models, where it is customary to smoothen the curve between the observed datapoints. This involves reducing the residuals as much as possible, and designing a continuous function over time. The problem of potential change in the unseen time slices is hardly ever addressed. In survival analysis, the problem is acknowledged. In medical or sociological settings, the situation of problem (iii) is known as ‘panel data’ or ‘interval censoring’, and has an impact on the underlying analytical methods in survival analysis (Therneau et al., Citation2020a, p. 13, pp. 28–29).

Problem (iv), synchronic variation, is solved in general linear (mixed) models by allowing multiple datapoints at the same moment in time. In logistic regression with a binary outcome, the proportion of the two variants at a certain point in time can be accommodated by a binomial factor, with a weight for the number of observations. It is less straightforward to cope with this problem in survival analysis. A particular patient either has the disease or does not at a certain observation point. A patient is either dead or alive. A mechanical device either functions or fails. This is an underlying assumption of the models. This necessitates an idealization of discreteness of different stages in (continuous) time, see also Krylov (Citation1995, pp. 19–20).

In short, general linear (mixed) models and survival analysis each have their advantages and drawbacks. The major drawback of the former is the overconfidence in the smoothness and monotonicity of the change, especially when there are gaps in the data, the major drawback of the latter is that it forces one to make a categorical decision about the state of the subject under investigation at a certain point in time. No single method is ideal. This echoes the words of the British statistician George Box: ‘All models are wrong; the practical question is how wrong do they have to be to not be useful.’

In this article, we will look at a well-studied linguistic change, which displays all problems (i)-(ii)-(iii)-(iv). We propose to use a multi-state Markov model. Because a lot is known about this particular change from extensive prior studies, we can validate the Markov model by confronting it with these earlier results. The change at issue is the preterite formation in Germanic. Our focus will be on Dutch.

2 Preterite Formation in Germanic

Ignoring the analytic perfect (ex. 1), which in some Germanic languages such as German, Dutch, and especially Afrikaans has been steadily gaining ground on the preterite for the past time reference (ex. 2 and 3), Germanic languages dispose of two ‘synthetic’ strategies to form the past. One is by changing the stem vowel (ex. 2), and the other is by adding a dental suffix (ex. 3). In historical linguistics, the former is called the ‘strong’ inflection, and is an Indo-European heirloom, while the latter is known as the ‘weak’ inflection, and is a Germanic innovation.Footnote1

  1. I have worked (analytic perfect)

  2. drink ~ drank (strong preterite)

  3. work ~ worked (weak preterite)

The distinction between the two inflection types is largely lexically specific: some verbs inflect strongly, other verbs inflect weakly. This distinction is, however, not diachronically stable. Through time, some weak verbs have become strong, but the more common ‘drift’ is for strong verbs to have become weak – understandably, as the weak inflection was originally the more innovative one, and is the default inflection for new verbs entering the system as loanwords. The gradual expansion of the weak inflection is a long-term change, spanning several hundred years. Historical linguistics have intensively studied these two competing morphological strategies for the Germanic preterite, both to reconstruct the earliest stages (i.a. Bailey, Citation1997; Hill, Citation2010; Lass, Citation1990; Mailhammer, Citation2007; Tops, Citation1974; Van Coetsem, Citation1990), and to trace the diachrony of the distribution (e.g. Anderwald, Citation2012; Cuskley et al., Citation2014 on 19th-century English; Lieberman et al., Citation2007 on Old English to Present-day English, Carroll et al., Citation2012 on Old High German to Present-day German, De Vriendt, Citation1965 on 16th-century Dutch; De Smet & Van de Velde, Citation2019, Citation2020a on 9th-century to 20th-century Dutch).

Of the quantitatively oriented studies, those which take a long term-perspective (> 1000 years, c.q. Carroll et al., Citation2012; De Smet & Van de Velde, Citation2019, Citation2020a; Lieberman et al., Citation2007), are invariably concerned with the major drift from strong to weak (or from irregular to regular), and do not take into account the reverse trajectory of verbs shifting (back) to the strong inflection, either temporarily or permanently. The number of (back)shifting verbs is, however, by no means marginal, at least not if we go by the figures in Cuskley et al. (Citation2014) on Late Modern English.

Methodologically, this poses a challenge. If we would set up a binary logistic regression, with strong inflection vs. weak inflection as the outcome variable, the time coefficient might be zero when the two trends (strong-to-weak and weak-to-strong) cancel each other out, even if for some types of verbs, there may be a clear diachronic trend. Adding a quadratic term (e.g. logitPweak=β0+β1time+β2time2), might alleviate the problem if the trend would turn out to be a U-shape, but it would underfit a more wavering trend. This is, in essence, problem (ii) referred to above. Problems (i) and (iii) are at play as well: verbs fall out of use in the course of time, and may hence cease to partake in the tug-of-war between the two inflectional types, and given the vast differences in token frequency, some verbs will regularly crop up in corpus studies, whereas others will only occasionally be witnessed. Problem (iv) manifests itself as well: for various synchronic slices, even if we take high-resolution snapshots per year, one and the same verb can occur both in the strong and in the weak inflection. Some linguistic studies amend this by setting a fixed value at a given time period (Carroll et al., Citation2012; De Smet & Van de Velde, Citation2019; Lieberman et al., Citation2007), which are derived by consulting reference grammars and dictionaries. Long-term large-scale corpus studies with higher resolution for variation (e.g. De Smet & Van de Velde, Citation2020a) are few and far between.

3 Methods and Data

3.1 Description of the Data

In this article, we will build a Markov model where, through time, verbs can either remain in the inflectional state they are in, or transition to an alternate state. One of the alternate states is lexical death. Though in principle, one can imagine a situation in which obsolete verbs are resurrected, we model this as a so-called ‘absorbing state’: once a verb crosses the Styx, there is no way back.

The first question is what the general lay-out of the Markov model should be here. Focusing on problem (ii), a simple transition diagram like , which one would use in cases of binomial regression, is not suited. We need a more complicated diagram, like , or, if we want to accommodate problem (i) as well, a diagram like .

Figure 1. Transition diagram used in most binomial regression models.

Figure 1. Transition diagram used in most binomial regression models.

Figure 2. Bidirectional transition diagram.

Figure 2. Bidirectional transition diagram.

Figure 3. Bidirectional transitional diagram with lexical death (competing risks illness-death model with recovery).

Figure 3. Bidirectional transitional diagram with lexical death (competing risks illness-death model with recovery).

The diagram in is well-known from the competing risks literature and is often called the illness-death model (Putter et al., Citation2007, p. 2414). The strong inflection can be compared to health, the weak inflection to illness, and lexical death to death. ‘Recovery’ from the illness state is possible, hence the double arrow between the strong and weak inflection.

We will feed this model with data for individual verbs, and see how likely it is that they transition to an alternate state over time. We took a large dataset, compiled for another study, consisting of all preterite and participle tokens of verbs which have been seen in the strong form in a corpus from the 9th to the 20th century, except for extremely frequent irregulars (zijn ‘be’, doen ‘do’, see ‘zien’, staan ‘stand’, gaan ‘go’, and worden ‘become’). We discarded the observations of the participles, retaining 189,841 observations of the preterites. These tokens represent 1630 different verb types, and 295 different verb stem types. With verb stems, we make abstraction of derivational-morphological operations: aanbieden (‘offer’), ontbieden (‘summon’) and bieden (‘offer’) are different verb types, but belong to the same stem. The observations were enriched for a large number of covariates. To keep things feasible, we will focus on two well-known covariates in this study, namely token frequency and the kind of vowel alternation pattern.

Frequency can be operationalized in a large number of ways (De Smet & Van de Velde, Citation2019), but we will enter it in the model as a ternary factor (‘low’, ‘medium’, and ‘high’), depending on whether the rounded base10 logarithm of the token frequency of the preterite of the stem was in the range 0 to 3, 4 to 6, or 6 to 10. Earlier studies have shown that the number of types is highest in the mid-range (a.o. Lieberman et al., Citation2007): low-frequency verbs tend to be weak, whereas high-frequency verbs do not count many members, due to the Zipfian distribution of words (Zipf, Citation1949). It is well-known that higher frequency verbs are less likely to change in inflection type.

Vowel pattern is also a ternary factor. Ablaut patterns come in different shapes, which are historically derived from 7 ablaut classes, but a major distinction is between so-called ABB, ABC and ABA verbs. ABB verbs have a distinct vowel in the present and the same vowel for the preterite and the participle (e.g. bijten /εɪ/ ~ beet /e:/ ~ gebeten /e:/, ‘bite’), ABC verbs have a distinct vowel in the present, past and participle (e.g. nemen /e:/ ~ nam /ɑ/ ~genomen /o:/, ‘take’), ABA verbs have the same vowel in the present and the participle, and a distinct vowel in the preterite (e.g. slapen /a:/ ~ sliep /i:/ ~ geslapen /a:/ ‘sleep’). Previous studies have shown that ABB preterites are best protected against weakening, because the distinct vowel in the present and the other two categories makes the ablaut more semiotically transparent, whereas ABA preterites are highly prone to change, because the vowel in the preterite works in isolation (De Smet & Van de Velde, Citation2019).

Not all of the observations in the dataset were useful for the present analysis. First, we discarded verbs that over time shifted from one of the 7 ablaut classes to another, because the vowel pattern may not be constant over time as well.Footnote2 Survival analysis can accommodate working with time-dependent covariates (Therneau et al., Citation2020b), but this adds another layer of complexity, which we will refrain from in this study. Furthermore, we kept derivationally complex verbs in the dataset, unless they were compounds (bezighouden ‘keep busy’, flauwvallen ‘faint’, raadplegen ‘consult’ etc.). This left us with 184,761 preterite observations of 286 stems.

Then we had to make three additional assumptions; first, related to problem (iv) above, we retained just one instance per year per verb stem. This resulted in a serious reduction of the data, as a given text, from a given year, may contain numerous instances of the same verb. This is fairly straightforward if the verb is categorically weak or strong in a given year, but when there were both strong and weak forms in a given year, we had to decide. In such cases, we gave priority to the weak form, in line with earlier research (De Smet & Van de Velde, Citation2019, p. 156) (unless the minority form represented 10% or less of the extant forms in that particular year, in which case we assigned the majority form to the verb in that year), leaving us with 14,314 observations of 285 verbs stems.Footnote3 This is the second assumption. The idea is that weak forms will, on average be underestimated in the textual record, because they often represent the prescriptively dispreferred form. The obvious problem is that a single weak observation in a given year inbetween a sea of strong observations in adjacent time periods, will lead to a rapid transition in the Markov model. On the other hand, we do want to capture the erratic behaviour of preterite inflection. Note that most other studies also drastically reduce the variation to build statistical models (Carroll et al., Citation2012; De Smet & Van de Velde, Citation2019; Lieberman et al., Citation2007). Third assumption: we assumed a verb was dead if its stem was not observed in the last century anymore.Footnote4 In this case, we assumed the verb lived on for another generation after its last observation. A generation was taken to be 30 years. We used our native speaker knowledge to assess the quality of this filtering mechanism to check whether the verbs that experience lexical dead are indeed obsolete. This turned out to be the case. Verbs experiencing lexical death include dwaan (‘wash’), kweden (‘say’), zwelten (‘succumb’), a.o. (see Appendix B).

3.2 Under the Hood of the Multi-state Markov Model

The competing risk survival analysis technique used in this paper is known as a ‘multi-state Markov model in continuous time’ (Jackson, Citation2019). Like any inferential statistical technique it tries to infer an underlying, idealized picture (hence: ‘model’) from a set of discrete observations. This model can be used to extrapolate or predict reality beyond the observed data sample.

In Markov processes, the state of an individual or an object relies solely on its prior state, not on its history. A crucial notion in multi-state Markov models is the transitional intensity. As shown in , various transitions are allowed, but not all transitions are equally likely. The intensity is the instantaneous risk that a verb will change from one state to the other, and can be captured by the following equation (Jackson, Citation2019, p. 2):

(4) qrst,Zt=limΔt0P(St+Δt=s|St=r)Δt(4)

The transition intensity q for a pair of states r and s is a function of time t or a set of variables Z(t). These variables can be fixed at the level of the individual (in our model: the verb stem), or can be time-varying. As said above, we will limit ourselves to variables that remain constant for the individual and are known at the outset.

This transition intensity of state r to state s is defined as the limit of a time increment Δt as it goes to 0, of the probability of the assumed state S for an individual at time t+ Δt (which equals the probability of state s given that the assumed state at time t was state r) divided by Δt.

For each transition, a q can be calculated, which can be represented in the format of a matrix Q, in which the rows give the ‘from’ states and the columns the ‘to’ states. If 3 is the absorbing state (death), then we can substitute the last row with zeros. The diagonals are the negative of the sum of the other values at that row (qrr=rsqrs).

(5) Q=q11q21q31q12q22q32q13q23q33=q12+q13q210q12q21+q230q13q23q31+q32=0(5)

The transitional probabilities at time t involve exponentiating the scaled matrix Q (6), which involves solving a set of linear differential equations (see Cox & Miller, Citation1965 for details).

(6) Pt=etQ(6)

(Time-independent) covariates are entered by multiplying a baseline qrs0 with an exponentiated covariate vector Z(t) multiplied by a coefficient βrs.

(7) qrsZt=qrsbaselineeβrsZt(7)

The underlying assumption is that the effect of the covariate is proportional on the basis transition intensity. In that sense, it is reminiscent of a proportional hazard model, commonly used in survival analysis.

The likelihood function for the Maximum Likelihood Estimation of the multi-state Markov model can be found in Kalbfleisch and Lawless (Citation1985). All calculations in this paper have been carried out with the package msm (Jackson, Citation2011) of the R software (R Core Team, Citation2019). For data wrangling we made use of the packages dplyr (Wickham et al., Citation2020) and reshape2 (Wickham, Citation2007).

4 Results

The model we will run has the transition intensities in (8) for r and s taking a value between 1 and 3, and with q3s=0 for all 285 individual verb stems (see Appendix A). The model has two covariates frequency and vowel pattern, but as these are both ternary factors, we need 4 betas. Low-frequency range and ABB pattern are treated as reference levels.

(8) qrszt=qrsbaselineeβrsmfZmft+βrshfZhft+βrsABCZABCt+βrsABAZABAt(8)

(mf = mid-frequency, hf: high-frequency, ABC: vowel pattern ABC, ABA: vowel pattern ABA)

Before we run the model, let us first look at the descriptive statistics. The dataset contains information on 285 ‘individuals’, casu quo verb stems, together seen at non-set intervals – technically a kind of ‘non-informative sample times’ (Jackson, Citation2019, p. 4) – for a total number of 14,314 observations. Most of the times, verbs just stay in the state they were in before. So the bulk of the observations will be on the diagonal in the state table ().

Table 1. State table of the observations.

We can observe that transitions from strong to weak slightly outnumber the transition from weak to strong, but then again, there are more strong observations than weak observations in total.

Fitting the model with two covariates, frequency and vowel pattern yields the baseline transition intensities in .

Table 2. Baseline transition intensities.

As in our dataset, verb stems are relatively unlikely to shift to the ‘dead’ state, the model has some trouble reliably calculating the transition intensities for transition to the ‘dead’ state (witness the large confidence intervals).Footnote5

Now let us turn to the effect of the covariates. Remember that these are expressed on the natural logarithm scale. Coefficients below 1 mean that the effect of the covariate is negative, lowering the transition intensity, and coefficients above 1 mean that the effect of the covariate is positive, boosting the transition intensity. The confidence interval should not span over the value 1.

For frequency, we can see that mid-frequency verbs have a lower probability to experience lexical death from either of the two states (strong or weak). For high-frequency verbs, the point estimates for the coefficient are even lower, but the confidence interval is too large to make the claim. Mid-frequency verbs have an increased probability of going from weak to strong, but not vice-versa. High-frequency verbs have a decreased probability to transition from strong to weak, but an increased probability to transition from weak to strong. Overall, this shows two things: first, the higher the frequency, the less likely the verbs shift to the weak inflection, a trend well-established in the literature, where frequency offers a protective effect against ‘regularization’ (Bybee, Citation2006, p. 715, who calls this the ‘conserving effect’). Second, that the higher the frequency of the verb, the more likely it shifts to the strong inflection state, if it was not already there. Here we concur with Fertig (Citation2013, pp. 131–132), who argues that high-frequency verbs more easily undergo analogical, as opposed to regularization changes.

For the vowel pattern, we see that both ABC and ABA verbs are more likely to transition from strong to weak, and less likely to transition from weak to strong than ABB verbs. All this echoes the greater resilience to weakening found in ABB verbs (De Smet & Van de Velde, Citation2019), and the multi-state Markov model has the additional benefit that it shows that the latter effect not only transpires as ABB being protected to weakening but also a special proneness to shift back to the strong inflection, if it has slipped to the weak inflection.Footnote6 Lexical death is not significantly different in ABC or ABA verbs, compared to ABB verbs.

The model can also be used to calculate a transition probability matrix for specified durations. What is the probability of a verb stem to shift from strong to weak after 100 years, after 500 years, after 1000 years … taking into account the effects of the covariates? give the respective transition probability matrices. The higher the number of years, the less the figures change. The main trend is that verbs are more likely to transition from weak to strong, than from strong to weak. Again, this does not mean that verbs are massively turning into strong verbs, of course, as we only looked at verbs that belong to the strong group anywhere in their history. The high transition probabilities of weak-to-strong are mainly due to strong verbs getting back in line.

Table 3. Effect of the covariates Frequency (low-mid-high) and vowel pattern (ABB-ABC-ABA).

Table 4. Transition probability matrix for t = 100 (years).

Table 5. Transition probability matrix for t = 500 (years).

Table 6. Transition probability matrix for t = 1000 (years).

The next question we can ask is how long verbs stay in each state. This is technically known as the ‘sojourn time’. As can be appreciated from , the verbs in our dataset stay longer in the strong state than in the weak state:

Table 7. Sojourn times (in years).

Model quality is visually assessed by comparing the observed and expected prevalence. In , it is clear that the model becomes worse after 700 years, and starts to overestimate the prevalence in State 1 (strong) and State 2 (weak), and underestimate the prevalence in State 3 (dead), but otherwise the fit appears to be good.

Figure 4. Prevalence plots for ‘strong’ (State 1), ‘weak’ (State 2), and ‘dead’ (State 3).

Figure 4. Prevalence plots for ‘strong’ (State 1), ‘weak’ (State 2), and ‘dead’ (State 3).

5 Conclusion

In this article, we approached language change with a technique that does justice to the peculiar nature of the time variable. We built a multivariate multi-state Markov model in continuous time, which can be thought of as an offshoot of survival analysis. Over the past decades, such models have been applied in medical contexts, but so far, they have not often been tested against linguistic data. Krylov (Citation1995) works out the mathematics of a model for semantic changes, but does not apply it to real data.

The advantage of this technique over more common techniques like generalized linear mixed models in which time is entered as a covariate, is that the Markov model can better capture the oscillation between different states. Unlike most generalized linear models it does not assume a monotonous linear, or S-shaped increase of decrease in two competing states, nor a U-shape. In this respect it can compete with generalized additive mixed models. The latter are data-hungry, and can lead to overfitting. Another advantage of the multi-state Markov Model is that it can work with competing risks.

This article is meant as a proof-of-concept. We build an illness-death model on Dutch preterite inflection, in which verbs can transition from the so-called ‘strong’ inflection to the so-called ‘weak’ inflection or vice versa, or experience lexical death. The results are compared with results from previous studies. Our Markov model seems to work well: it brings out the ‘conserving effect’, whereby frequency correlates with survival rates of a lexeme or a construction, and it brings out the protective effect of vowel patterns in the ablaut.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work was supported by the FWO Vlaanderen [PhD 3H160540] and the Special Research Fund (BOF) from the University of Leuven [Project 3H180263 ‘How predictable is language change?’].

Notes

1. This terminology slightly deviates from the more familiar distinction between ‘regular verbs’, which predictably form their preterite with the suffix -ed, and the ‘irregular verbs’ (all the other verbs). The problem is that some irregular verbs are technically speaking weak (e.g. English keep – kept), and are irregular because of the influence of sound laws and analogy, and, conversely, that strong verbs are not devoid of regularity, as the specific vowel change (‘ablaut’) can be predicted from the present-tense vowel and the consonant skeleton of the root, especially in older languages stages and in Germanic languages other than English. This article is not the place to delve into the intricacies involved; readers are referred to historical reference grammars where these matters are explained in detail (e.g. Prokosch, Citation1939; Ringe, Citation2006), or to De Smet and Van de Velde (Citation2019, pp. 141–143) for a concise discussion.

2. Occasionally, the vowel pattern of verbs that stay within their historical ablaut class also shifts diachronically. We followed the procedure in De Smet and Van de Velde (Citation2019), and assigned one fixed value per verb.

3. We discarded one verb stem, fuiven (‘party’), which was only attested once in the dataset, and hence does not impact the Markov model.

4. We set the limit at 1914, the end of what historians call ‘the long 19th century’.

5. We also ran a two-state model, without the death state (, above), and this barely affected the baseline transition intensities () or the covariate coefficients (, below). Future applications of the method tested in our proof-of-concept article would be well advised to use a phenomenon with higher death rates (lexical replacement). Here we stick to the three-state illness-death model. We adhere to the advice that model selection should be based on real-world or theoretical considerations, rather than deciding on the basis of statistical measures (see Harrell, Citation2015, pp. 67–72 in the context of regression).

6. As we have pointed out above, the dataset is biased towards weak verbs that became strong. The large number of verbs that happily stayed weak during their existence are not in our purview.

References

Appendix A.

Verb stem (types), observations per century (homonymous verbs disambiguated by English translation).

Appendix B.

verb stems experiencing lexical death (last seen before or in 1914)

beiden, belgen, bijden, blouwen, braden, dinsen, draden, drepen, dwaan, fijnen, grijnen, grijzen, houwen, klijven, klinken (‘hit’), kluiven, kruien, kruizen, kweden, kwelen, laden (‘invite’), lijden (‘declare’), lijen (‘borrow’), lijken (‘please’), louwen, malen (‘paint’), pijpen, pleken, pluiken, prenden, reken, rijgen, rijnen, rijten, scheren (‘arrange’), slijpen, sluiken, spanen, stenen, stuipen, tijgen (‘accuse’), vijsten, vlieten, vlijten, vreisen, wallen, waten, wellen, woepen, wouden, zijken, zijpen, zouten, zwelten, zwingen

(Homonymous verbs disambiguated by English translation)