Articles | Volume 8, issue 1
https://doi.org/10.5194/gchron-8-191-2026
https://doi.org/10.5194/gchron-8-191-2026
Research article
 | 
30 Mar 2026
Research article |  | 30 Mar 2026

The conflict between sampling resolution and stratigraphic constraints from a Bayesian perspective: OSL and radiocarbon case studies

Guillaume Guérin, Pierre Guitton-Boussion, Imène Bouafia, and Anne Philippe
Abstract

Bayesian modelling is often implemented in geochronology and its applications to geomorphology, archaeology, etc. The rationale behind such practices is the aim to improve robustness, precision and accuracy thanks to the use of prior knowledge regarding the studied sites, and in particular the order of samples constrained by stratigraphy. All chronological models tested in this study (OxCal, Chronomodel and BayLum) use the same mathematical model to handle stratigraphic constraints. However, this model has been shown to lead to estimation biases. First, this bias is illustrated with BayLum modelling on a high-resolution OSL dataset. Then, this paper compares statistical inferences obtained with the three above-mentioned modelling software on the Neolithic East mound of Çatalhöyük (Turkey). For this site, 49 radiocarbon ages were obtained with the aim to determine the start of occupations at this locality. Interestingly, age uncertainties are rather large, because of calibration curve plateaus. Therefore, the conditions for estimation biases are met. We discuss the behaviour of the different models and show that caution must be taken when modelling results are at odds with measurements. While OxCal, Chronomodel and BayLum are all affected by a spread in ages resulting from their common model of stratigraphic errors, Chronomodel suffers from a great loss of precision and OxCal, through the phase model, concentrates ages undesirably. We also conclude that the onset of occupations at Çatalhöyük was probably earlier than previously thought based on the OxCal model.

Share
1 Introduction

In geochronology, Bayesian modelling is often used to include prior knowledge for estimating time sequences. Perhaps the most important, and most commonly used prior knowledge, is the chronological order of events dictated by stratigraphy. In a typical stratigraphic sequence composed of several sedimentary layers, relative ordering arises naturally: if Layer a was deposited before Layer b, then their respective deposition ages ta and tb satisfy

tatb

Although this statement appears straightforward, its formal translation into a mathematical model is the central focus of the present work. To our knowledge, all existing chronological modeling approaches employ the same prior probability distribution on the ordered dates. We refer to this as the uniform order denoted by 𝒰*. Let P and A, such that P<A be the lower and upper bounds of the study period. If we consider n layers, we can thus define a strict temporal ordering as

t1,,tnU*P,A

with a corresponding density

(1) π t 1 , , t n = n ! A - P n if P t 1 t n A 0 otherwise

Such a prior is indeed implemented in Chronomodel (Lanos and Philippe, 2017; Lanos and Dufresne, 2019), where it forms the foundation of the hierarchical structure used to represent stratigraphic relationships. While the software is capable of incorporating partial ordering, the present discussion focuses exclusively on the case of complete ordering:

conditional to ξi:MiNgiξi,τi2+σgi2ξiiconditional to ti,σi2:ξiNti,σiiσi2shrinkageτ2it1,tnU*P,A

Where measurements (Mi,τi) are related to dated events ξi, via a calibration function gi (with error σgi2), which are in turn related to target events ti with error σi quantifying their discrepancy.

A similar implementation exists in BayLum (Combès and Philippe, 2017; Christophe et al., 2020; Philippe et al., 2019; Guérin et al., 2021), where various forms of stratigraphic constraints can be defined and expressed through a matrix S. In the specific case of a strict order, BayLum applies these constraints to compute the following OSL model:

D1,,DnNdiagtd˙,diagtΣdiagt+diagσ2e1,enU*logP,logAti=expeii{1,,n}

where:

  • D=D1,,Dn are the equivalent doses determined from luminescence measurements;

  • d˙=d˙1,,d˙n are the mean annual dose rates;

  • σ2=σ12,,σn2 are the measurement errors;

  • Σ is the covariance matrix capturing shared uncertainties in mean annual dose rate estimation d˙;

  • diag(t) is the n×n diagonal matrix with entries t1,,tn on the diagonal.

The BayLum prior has a density of the following form

πt1,,tn=n!logA/Pni=1n1tiifPt1tnA0otherwise.

In OxCal (Bronk Ramsey, 2009), stratigraphic relationships are represented through artificial parameters known as phases ψ1,,ψn+1. In this framework, dated samples are typically modeled as occurring between two undated events that mark the beginning and end of a phase encompassing those samples. Consequently, the chronological order is no longer imposed directly on the ages themselves but rather through the ordering of these phases. For instance, in the case of contiguous phases, the model can be expressed as

Pψ1t1ψ2ψntnψn+1A

Here, it is the phase boundaries ψi forall i that are ordered, while the sample ages are nested within them.

ψ1,,ψn+1U*P,AConditional to ψ1,,ψn:tiUψi,ψi+1i

In OxCal, P and A exist implicitly: by default, their values correspond to the bounds of the calibration curve. Alternatively, they can be explicitly defined using the After and Before functions, or by defining each phase boundary with a uniform distribution U(P,A), with all boundaries and phases wrapped inside a Sequence to impose ordering.

These three software programs address specific needs: OxCal was initially developed for radiocarbon (Hajdas et al., 2021), while Chronomodel was designed to combine various dating methods; finally, BayLum was specifically conceived for Optically Stimulated Luminescence (OSL) dating (Murray et al., 2021). (Note: it also allows radiocarbon calibration and modelling, like OxCal accommodates non-radiocarbon ages in the form of gaussian probability densities).

This choice of the uniform order prior was the subject of a study by Steier and Rom (2000), who concluded that the uniform order prior may lead to estimation biases in the following cases:

  1. Presence of outliers: this prior is not robust to outliers, whose influence can propagate and shift all estimated ages. However, the present study does not address this issue directly;

  2. Increasing number of samples within a fixed period: as the number of dated samples grows for a given time span, the estimated ages of the oldest and youngest samples tend to diverge from their true values;

  3. Minimal true age gaps relative to measurement uncertainties: when the temporal differences between successive events are small compared to measurement uncertainties, the older ages tend to be overestimated and the younger ones underestimated.

Furthermore, Steier and Rom (2000) noted that models employing the uniform order prior may yield deceptively low uncertainty estimates, thereby creating an illusion of higher precision. This occurs because the highest posterior density (HPD) intervals are artificially reduced. Indeed, narrower intervals appear more precise but at the expense of accuracy: in their simulations, the true age often fell outside the 95 % credible interval.

In practice, the two issues described in points (2) and (3) above (thus, excluding the question of outliers) arise when the sampling resolution is too high i.e, when the true age difference between successive samples is small compared to the error of the measures. In fact, the problems arise when the uncertainties are large, i.e. when the measures' marginals for successive sample are overlapping; in such cases, there is a strong probability that measures are inverted. To illustrate this, let us consider the simple case of Gaussian measurement errors:

M1,,MnNnt1tn;σ2In

and compute the probability of an inversion between two consecutive measurements:

PMi+1-Mi0=Φ-ti+1-ti2σ

where Φ denotes the cumulative distribution function of a standard normal distribution 𝒩(0,1). Two limiting cases emerge:

  • if σ|ti+1-ti|, we are in the tail of the distribution, yielding a very small probability of inversion;

  • if σ|ti+1-ti|, the difference term approaches zero, and the inversion probability tends towards 0.5, corresponding to half the area under the Gaussian curve.

From this, we may deduce that issues will arise more frequently with imprecise measurements, and thus should occur more with OSL than with radiocarbon, simply because OSL is less precise than radiocarbon1.

Hence, the loss of accuracy arises because the modelled ages are spread by the uniform order model (Eq. 1), i.e., the age of younger samples becomes underestimated while the age of the older samples gets overestimated by the strict ordering constraint (Steier and Rom, 2000). In the following, we refer to this observation as the spread effect.

The spread effect becomes evident when posterior distributions generated under a prior that incorporates stratigraphic constraints are compared with those obtained under a prior that does not. Under the uniform order prior, a characteristic pattern emerges: for two samples with statistically indistinguishable measured ages, the model enforces separation by generating two distinct posterior distributions. In doing so, it effectively rules out solutions where the samples share nearly identical true ages. This modelling artefact gains importance as chronological datasets grow denser. The continuing development of dating techniques and the increasing number of dated samples per study, particularly in high-resolution chronologies such as those reported by Stevens et al. (2018), intensify the problem, making the spread effect an increasingly significant source of estimation bias.

In reply to Steier and Rom (2000), Bronk Ramsey (2000) pointed to boundaries as a mean to overcome the spread effect. Such a model has been shown to lead to a concentration effect (Lanos and Philippe, 2018): the greater is n, the greater is the probability that ψn+1-ψ1 is small in the model outcome. (i.e., lower differences between ψn+1 and ψ1 are favoured by the model). In other words, the phase model tends to cluster all ages inside a phase: ψi+1tiψii. Another effect of the phase model is that phase boundaries ψi+1,ψi may extend well outside the range of measured ti value, all the more when we have only few age ti inside each phase ψi+1,ψi (e.g., Guérin et al., 2023).

Such behaviors are generally undesirable, which motivated Nicholls and Jones (2001) to propose an alternative prior in which the phase duration is explicitly modeled:

(2) d n U 0 , A - P

Since each software implements different strategies for representing stratigraphic information, the notion of duration can be interpreted in various ways depending on the modeling framework:

(3) d n = ψ n + 1 - ψ 1 with contiguous phases of OxCal d n = t n - t 1 in the case of a strict order d n = max t i - min t i in general cases

In OxCal, users have the option to apply either the standard uniform order prior (described above) or the uniform phase-based prior introduced by Nicholls and Jones (2001). When the latter is wanted, the user needs to set the parameter UniformSpanPrior to true (note: this is the default option and is implemented in the present work). This configuration defines the prior differently in OxCal (compared to what is presented above for this software). The following example below illustrates how a contiguous phases model is represented under this configuration:

(4) d n U 0 , A - P Conditional  to   d n : ψ 1 U P , A - d n  and ψ n + 1 = ψ 1 + d n Conditional to  ψ 1 , ψ n + 1 : ψ 2 , , ψ n U * ψ 1 , ψ n + 1 Conditional to  ψ 1 , , ψ n : t i U ψ i , ψ i + 1 i

The first purpose of this article is to illustrate the issue at stake using various datasets obtained from different methods (namely, OSL and radiocarbon), and to compare different Bayesian chronological models: OxCal, Chronomodel and BayLum. By comparison with OxCal, in Chronomodel the event model of Lanos and Philippe (2018) is the basis for chronological modelling; every date or age is indeed encapsulated in one event – and one event may contain several dates or ages (for example, speleothems or corals may be jointly dated by U/Th and radiocarbon). Thus, in Chronomodel the concentration effect should not appear; however, the event model introduces an additional time uncertainty, of a similar magnitude as the measured age uncertainty. Finally, BayLum is a somewhat simpler model regarding the modelling of ages in stratigraphic constraints, in the sense that only the spread effect should be visible and that no additional uncertainty is added to measured ages. To summarise, OxCal is affected by the spread and concentration effects, Chronomodel by the spread effect and by additional uncertainties, and BayLum by the spread effect only. The purpose of this paper is to illustrate the issue that may be encountered in chronological modelling of OSL and ?14C-based chronologies. We show how it can be recognised upon scrutiny by comparing ages modelled with or without stratigraphic constraints. To this end, we use two real datasets rather than modelled ones to emphasize the possibility to detect modelling artefacts without knowing the true age of each sample. First, an already published, high-resolution OSL dataset, is modelled using BayLum. Issues related to increased sampling resolution are demonstrated using the open-source ArchaeoPhases R package (Philippe and Vibet, 2020). To compare chronological models when confronted with high-resolution datasets, we turn to radiocarbon dating. Indeed, in BayLum OSL measurements are combined in a hierarchical model linking regenerative doses, individual equivalent dose estimation, the central dose parameter of interest, etc. Conversely, OSL ages can be included in OxCal and in Chronomodel, but only in the form of Gaussian probability densities (in practice, an age and its uncertainty). Therefore, to make the model comparison as straightforward as possible, we use a high-resolution radiocarbon dataset. Each of the three tested models indeed performs calibration of measured 14C ages, using similar Markov Chain Monte Carlo (MCMC) algorithms.

Before presenting the method, we should warn the reader that radiocarbon and OSL communities use different conventions, both in terms of time axis orientation and system of units: in radiocarbon, the x-axis goes from past (left) to present (right), whereas in OSL the x-axis is age, so increasing age (i.e., going right on the x-axis) goes back in the past. Thus, by convention the Bayesian tools developed in either community use different standards. The same applies to time units: radiocarbon uses calendar dates expressed in years BC, whereas in OSL the standard physical quantity is age, expressed in ka. As a consequence, we used the OSL conventions for the study of OSL ages – and the radiocarbon convention for the C-14 study. The only exception to this rule is when only BayLum (developed in the OSL community) is used to model a sequence dated with radiocarbon only (Fig. 9); for this illustrative figure, the OSL convention was adopted for radiocarbon.

2 Materials and methods

2.1 OSL dating and BayLum

Loess plateaus often offer long sedimentary sequences spanning long periods of time; as such, they provide valuable records of sediment accumulation together with climatic and environmental proxies. During periods of fast sediment accumulation, series of indistinguishable OSL ages may be obtained over several meters of depth. Such is the case of the Jingbian loess plateau, recently studied by Stevens et al. (2018). For our study, we selected of cluster of 10 samples deposited around 15 ka ago (Fig. 1 and Table S1); 7 of these 10 samples are consistent with the average age (14.6 ka) at one standard deviation and 3 at two standard deviations. In other words, all ages are consistent with a short, fast deposition event (short and fast compared to individual age uncertainties). Yet, they correspond to 90 cm of loess accumulation; therefore, their true ages are ordered following the stratigraphy. To investigate the issue posed by Eq. (1) in the modelling of stratigraphic constraints, we used BayLum to model the stratigraphic sequence. To take systematic errors into account (which is important when imposing the stratigraphic constraints, because some errors are not random), we built a Theta matrix following Guérin et al. (2021); such a matrix takes into account the covariation of ages in the MCMC algorithm, by defining to which extent systematic errors (on the laboratory source dose rate, but also on radioelement concentrations, etc.) affect a suite of OSL ages. Following the assumptions made by the authors of the original study (Stevens et al., 2018), we considered the following sources of systematic errors:

  • calibration of gamma spectrometers for the concentrations in K, U and Th (1.5 % uncertainty on each value);

  • calibration of the laboratory-source dose rates (3 % uncertainty);

  • internal dose rate (0.01 Gy ka−1, absolute uncertainty);

  • water content, taken to be 15 ± 5 %; we treated all the uncertainty associated to this parameter as corresponding to a systematic error.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f01

Figure 1OSL ages as published by Stevens et al. (2018) for the study of a loess Plateau in Jingbian, China. The subset of 10 samples analysed in the present study lies inside the red circle. In this figure, error bars correspond to one sigma, i.e., to the 68 % confidence interval.

As pointed out in the introduction (see issue number 2), for a fixed duration of events the number of dated samples is a key parameter in the issue at stake: the more samples there are, the wider is the resulting spread in ages modelled in order. Therefore, we investigated the effect of including various numbers of samples on the estimated duration between the deposition of the first and last samples. To quantify the effect, we used the open-source R package ArchaeoPhases to estimate this duration. We also looked at the joint probability between ages estimated with and without imposing stratigraphic constraints, to estimate the consistency between the model output and the input data, as a function of the number of samples included in the model.

2.2 Comparison of OxCal, BayLum and Chronomodel on a radiocarbon dataset

For the purpose of model comparison, we selected a high-resolution radiocarbon dataset, with large uncertainties. In general, measurements of the 14C/12C ratios display <1 % uncertainty, so we decided to focus on periods affected by plateaus in the calibration curve (Reimer et al., 2020) – i.e., periods for which similar 14C/12C ratios correspond to various true ages. One study that appears to correspond to these criteria is the chronological investigation led by Bayliss et al. (2015) at the famous archaeological site of Çatalhöyük, in Turkey. 49 radiocarbon ages from 46 samples (3 measurements were replicates and combined with their counterpart) were obtained from the stratigraphy, spanning a period from 7600 until 6600 BC. Table S2 provides all selected radiocarbon ages. If ages were distributed homogeneously,  20 years would separate two successive samples. This order of magnitude for the true age difference is much smaller than typical age uncertainties: analytical uncertainties on the 14C/12C ratio are  40–50 years, and after calibration these uncertainties are largely increased.

The purpose of the present study being to compare OxCal, Chronomodel and BayLum on the same dataset, we will not discuss here data selection and modelling choices made by Bayliss et al. (2015). Thus, we reproduced their model, based on their discussion but also their modelling choices. Indeed, OxCal offers the possibility to define phases, encapsulated phases (i.e., phases inside other, larger phases), hiatuses, terminus ante- and post-quem, boundaries, etc. For example, one may wonder why boundaries are defined in the model sometimes before, in other places between, and/or after phases. But this is not the purpose of our work here, so we take the arguments of Bayliss et al. (2015) at face value and stick to their preferred model.

By comparison with OxCal, it is important to note that Chronomodel and BayLum offer fewer options to model chronological sequences: in the latter two modelling programmes, events (in Chronomodel) and ages (in BayLum) are stratigraphically constrained according to the Harris matrix provided by Bayliss et al. (2015; their Fig. 3), but are not affected by the definition – or absence of definition – of boundaries and phases. This lack of modelling choices actually appears to be a positive feature in our view, in the sense that it tends to provide less arbitrary results. One clear example of issues posed by the phase model can be seen, e.g., in the work of McPherron et al. (2012), who modelled 10 radiocarbon samples from one unique layer. In their study, defining a phase – i.e., saying that there was a “before” and an “after” – affects the chronological inference by (i) concentrating the ages and (ii) excluding ages considered as outliers. In our view, there can be no outlier detection when no stratigraphic inversion is detected; as discussed in detail in Guérin et al. (2023), from first principles the start of a chronological phase should not possibly be defined after the oldest sample and before the youngest sample.

Figure 2 presents the 95 % credible interval (C.I.) of the ages calibrated with OxCal, Chronomodel and Baylum. For calibration, IntCal20 was used in OxCal and BayLum, while it is not available in the current version of Chronomodel (2.0). Hence, IntCal13 was used in Chronomodel. Most age intervals are very similar and most differences occur between Chronomodel on the one hand, BayLum and OxCal on the other. We attribute these differences to the change in the calibration curve. It should be noted here that Fig. 2 only includes the ages that are not considered as outliers in subsequent modelling (and in Bayliss et al., 2015); indeed, including here ages that are clearly off the stratigraphic record would render this figure difficult to read. However, all ages are given in Table S2.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f02

Figure 2Radiocarbon ages from Çatalhöyük calibrated with OxCal, Chronomodel and BayLum. Note: IntCal20 was used in OxCal and BayLum, while it is not available in the current version of Chronomodel. Hence, IntCal13 was used in Chronomodel. The ages appear in stratigraphic order and exclude outliers (see Sect. 2.2 for details).

Download

All calibrated ages shown in Fig. 2 belong to the Çatalhöyük mound period, except the two bottom-most samples that pre-date the mound. So, if we consider all ages from the mound, two sets of samples can be distinguished: one sample set at the bottom of the mound (from sample OxA-9893 to sample 109993) dates from between  7350 and  7050 BC, and a second set (from sample OxA-9776 to sample 103134) that dates from between  7200 and  6700 BC. To put things simply, the ages are rather imprecise because they correspond to two plateaus of the calibration curve; and, according to the unmodelled data, the mound was built after 7350 BC and stopped being occupied before 6700 BC. However, Bayliss et al. (2015) concluded from their chronological model that deposit from the mound started to accumulate between 7165 and 7085 BC (at the 95 % confidence level), which is in striking contrast with the measurements from the mound itself. This modelling result then has significant archaeological implications regarding the appearance of pottery and sheep domestication in the area. Furthermore, it raises questions about a 4-century time gap between the onset of Çatalhöyük and nearby archaeological remains found in Boncuklu (see Bayliss et al., 2015, and references therein). The questions we want to address is (i) that of the consistency of the modelling outcome, (ii) to determine whether other models would provide a different inference, and most importantly (iii) discuss the reasons explaining the potential discrepancies obtained with various models.

3 Results

3.1 Loess plateau dated by OSL

Figure 3 shows the output of BayLum, i.e., the modelled sequence obtained after 10 million iterations (note: convergence of the MCMC chains was assessed using the Gelman-Rubin statistic; convergence only required 5000 iterations when the ordering constraint is not imposed. The fact that so many iterations were required when modelling ages under stratigraphic constraints already suggests that the results should be taken with caution).

For clarity, on this figure the average age of the 10 samples estimated with BayLum, without imposing the stratigraphic constraints, is indicated (14.7 ka, instead of 14.6 ka estimated with the ages published by Stevens et al., 2018; we regard this 1 % difference as insignificant). Visually, as expected the ages now appear in stratigraphic order; in addition, the ten ages seem to form a smooth sequence of continuous sediment accumulation, in sharp contrast with Fig. 1. One may also note that only 5 ages agree with the average of unmodelled ages (NB: by unmodelled, here we refer to ages estimated without imposing the stratigraphic constraints) at the 68 % credibility level; 4 at the 95 % level, and 1 age does not include 14.7 ka at the 95 % level. In other words, it seems that imposing stratigraphic constraints leads to ages in poor agreement with the measured data.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f03

Figure 3Jingbian OSL ages estimated with BayLum, taking into account stratigraphic constraints and systematic errors.

Download

Figure 4 shows four series of OSL ages estimated with BayLum using stratigraphic constraints and systematic errors. To investigate the influence of sampling density, we made the number of samples vary from 2 (keeping only the top most and bottom most samples) up to 10 (case shown in Fig. 3). In each case, we kept the top most and bottom most samples. Intermediate calculations included 4 and 7 samples, respectively. The convergence of MCMC chains was assessed using the Gelman and Rubin statistic; it is worth emphasizing here that calculation times (i.e., the number of iterations) needed to reach convergence increased exponentially with the number of samples – well beyond the additional time required to account for the greater number of variables. We attribute this feature to the decreasing agreement between the input data and the modelling output: BayLum struggles to find an appropriate solution combining the data and the stratigraphic order, which is not surprising when considering the measured ages (Fig. 1). When looking at Fig. 4, it appears that the sampling resolution strongly affects the modelling results; more precisely, the spread effect is increased when the number of samples is increased, as was expected based on the study of Steier and Rom (2000). Indeed, the difference between the two extreme ages (top and bottom) appears to increase when intermediate samples are analysed. Moreover, when the number of samples included in the modelling is increased, it appears that the credible intervals become more and more shifted from the age intervals estimated independently.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f04

Figure 4Series of OSL ages estimated with BayLum, taking into account stratigraphic constraints and systematic errors. In all series, the top most and bottom most samples are included in the calculation. However, the total number of samples is varied from one calculation to another, from 2 up to 10. Note that for clarity, the y-coordinates of each sample are slightly shifted when the total number of samples included in the model is varied. The true depth values correspond to the black lines.

Download

To investigate this issue further, we looked at the joint probability, i.e., the probability that the ages modelled in stratigraphy fall in the 95 % C.I. of the ages estimated independently. To do this, we use the MCMC sample used to estimate the posterior probability density of the ages. We then count the fraction of iterations for which all individual ages belong to their initial 95 % C.I., estimated without imposing stratigraphic constraints. Ideally, an increase in the number of samples would lead to a gain in the precision of the statistical inference, i.e., the length of the C.I. should decrease whilst the model output stays consistent with the data – so the joint probability should stay close to 1. In other words, the joint probability ranges from 0 to 1 and is a measure of the agreement between the model output and the data used as input – 1 indicating perfect agreement and 0 very poor agreement (Nb: this joint probability is similar to OxCal's agreement index; Bronk Ramsey, 1995). Figure 5 shows the evolution of the joint probability as a function of the number of samples included in the calculation, following the same procedure as in Fig. 4; in Fig. 5, we normalised this joint probability by 0.95n, where n is the number of modelled samples (therefore, if the stratigraphic model yielded identical results compared to the unmodelled data, this normalised probability would be equal to 1). The joint probability of ages decreases when the number of samples modelled in stratigraphy is increased, demonstrating that the model output becomes less and less consistent with the input data. Of course, such a situation might occur if data is inconsistent with the assumption underlying the model, for example if outliers are present, or if – for some reason – ages decrease with depth. However, in our case the data simply suggest a fast accumulation of sediment, all samples being deposited around 14.7 ka. As pointed out by Steier and Rom (2000) in the case of radiocarbon, it appears that modelling of OSL ages in stratigraphy using BayLum, and thus Eq. (1), leads to the same spread effect.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f05

Figure 5Normalised joint probability between BayLum ages: (i) estimated independently and (ii) estimated under stratigraphic constraints. In practice, the joint probability corresponds to the probability that each iteration of the MCMC, when calculating ages in stratigraphic constraints, falls in the 95 % C.I. of the ages estimated independently (i.e., without imposing stratigraphic constraints). To assess the stratigraphic model performance, we normalise this probability by 0.95n, where n is the number of samples (therefore, if the stratigraphic model yielded identical results compared to the data estimated without constraints, this normalised probability would be 1).

Download

Finally, we also estimated the duration between deposition of the first and last samples considered here, as a function of the number of samples included in the model (in a similar fashion as for Figs. 4 and 5). To this end, we used the function PhaseStatistics of the R package ArchaeoPhases (Philippe and Vibet, 2020). Based on visualisation of the data (Fig. 1), we expect a short duration (compared to measured uncertainties). However, it appears on Fig. 6 that an instantaneous deposition event (i.e., a phase of null duration) is only included in the 95 % C.I. of the duration in the case where only 2 samples are included in our BayLum model. Overall, both the range of values and the length of the C.I. increase with the number of modelled samples. In other words, when including more samples, and so more information, the precision of the estimate decreases and the short duration values become less and less credible, which is in striking contradiction with the data.

To summarise, the model of stratigraphic constraints (Eq. 1) implemented in BayLum is not robust; neither is it precise (since the precision of the estimated duration increases with the number of modelled samples), nor accurate (since short phases are not amongst the most credible values, contrary to what the data suggest).

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f06

Figure 695 % credible interval of the duration between the first and last sample deposition, based on the ages shown in Fig. 4. It appears that the range of credible values increases when the number of samples modelled in stratigraphy is increased. In addition, the length of the 95 % credible interval is also increased, which corresponds to a counterintuitive loss in precision when more data are used for the estimation.

Download

3.2 Radiocarbon dating at Çatalhöyük

3.2.1 OxCal

Figure 7 presents the sequence of 46 radiocarbon ages modelled with OxCal. Seven ages appear as outliers and are shown in red on this figure, as in the original study. All seven outliers correspond to hackberry seed samples (Celtis sp.), except one bone from a stillborn sheep (UCIAMS-109995). The reason why these samples yielded anomalous ages was unknown in the original study and it still is today. In the following, unless explicitly stated otherwise, these samples are excluded from our dataset.

The concentration effect is particularly visible at the bottom of the sequence, where large parts of the probability density distributions of the calibrated radiocarbon ages (towards older ages, left of Fig. 7) are excluded of the model inference (between samples OxA-9893 and OxA-23251, bottom black bracket in Fig. 7). Yet, only two ages (the two bottom most samples, that predate the onset of the mound) constrain these ages towards younger values; and the two bottom-most samples are affected by very large uncertainties due to a calibration plateau. These two bottom-most samples are PL-980252A, whose age (95 % C.I.) almost entirely lies outside the calibrated age (95 % C.I.) of all samples above – so it does not bring any really useful information regarding the start of the mound; and sample AA-27982, whose calibrated age (defined here as its highest posterior density (HPD) interval) partly overlaps the density distributions of samples above. The model completely removes the possibility that the age of sample AA-27982 is greater than 7300 BC, although approximately one third of the HPD interval extends beyond this date. This concentration effect is even more visible on Fig. 2 of Bayliss et al. (2015)2. We see no reason to this behaviour other than a modelling artefact. Simply looking at the calibrated data, one would think that the start of the mound is plausible at 7300 BC, since 8 samples from the mound itself include this date in their 95 % credible interval (C.I.).

Another concentration effect is also visible in the range between samples OxA-9776 and UCIAMS-103138 (top black bracket): all these samples yielded similar, almost indistinguishable ages with large uncertainties; yet, the outcome of OxCal for this samples is very small credible intervals concentrated on a tiny portion of the calibrated age intervals.

Finally, the spread effect is also quite clearly visible, on the top part of the sequence (from sample OxA-9776 upwards; blue bracket in Fig. 7): essentially, considering measurement and calibration only, all these samples have indistinguishable ages, because of the plateau of the calibration curve. To put it in simple words, most – if not all – of these samples have a calibrated credible age interval spanning between 7100 and 6700 BC. Saying that some are older than others cannot reasonably end up in saying that some of these samples are very precisely dated to between 7100 and 7000 BC, which is the outcome of the OxCal model. In our view, it does not appear sensible to hope that a stratigraphic model may resolve calibration issues, and in particular plateaus in the calibration curve.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f07

Figure 7OxCal model for the sequence of Çatalhöyük. Empty curves correspond to calibrated radiocarbon ages, while filled curves correspond to modelled ages according to the stratigraphy. The phase structure reproduces that of the initial publication (Bayliss et al., 2015). Red and brown: outliers; grey: phase boundaries; green: all individual ages consistent with the model. Note that the boundary `Start Mound' occurred after the two bottom-most dates, which are terminus post quem. The black brackets indicate visible concentration effects, whereas the blue bracket indicates a visible spread effect (see text for details).

3.2.2 Chronomodel

Figure 8 presents the chronological inference obtained with Chronomodel for the bottom part of the sequence (note: replicate measurements of a sample were included in one event). It should be noted here that including more samples on this figure would make the curves and text unreadable. Therefore, we chose to represent the bottom (rather than the top) part of the sequence, because this is the part which has the most important archaeological implications (at least as of Bayliss et al., 2015). However, it is important to state here that all samples were included in the model – even though the results are not all presented here in figures. It should be emphasized that most, if not all of our following observations and comments are illustrated by Fig. 8, even though it only partially reflects our modelling results.

As expected, Chronomodel easily accommodates outliers thanks to the event model. However, the spread effect is rather pronounced: quite quickly along the stratigraphy, the age probability densities of subsequent samples are indeed not overlapping, contrary to what is observed on Fig. 1 showing the calibrated ages. This is not surprising because the event model, which is the modelling basis in Chronomodel, inflates analytical uncertainties and so is expected to increase the modelling artefact (see introduction).

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f08

Figure 8Part of the chronological model obtained with Chronomodel for the sequence of Çatalhöyük, showing the 17 bottom-most samples. The x-axis represents time BC (in years).

Download

3.2.3 BayLum

Before using BayLum, all radiocarbon ages obtained from hackberry endocarps were excluded from our analyses, as well as sample UCIAMS-109995, since they all appeared as outliers in the OxCal model. It should indeed be noted that the algorithm for outlier detection in BayLum appears to be less efficient than that of OxCal. However, the chronological inference obtained with OxCal also ignores these outlier ages, so the comparison between chronological inferences obtained with OxCal and BayLum remains valid.

Figure 9 shows the output of BayLum for the 40 remaining samples, obtained after 500,000 iterations. Here also, the “spread effect” is visible – to a comparable extent to that observed with OxCal, and so less than with Chronomodel. For clarity, a blue bracket is shown in Fig. 9 where the spread effect is most visible: all calibrated ages for samples from sample OxA-9776 up to sample UCIAMS-98210 are essentially the same, but after modelling in stratigraphy the age intervals become a sequence of disjointed, short intervals compared to the initial estimates.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f09

Figure 9Ages for the sequence of Çatalhöyük modelled with BayLum. Note that the output of BayLum only indicates the sample labels for one every two samples (the reader is referred to Table S2 and Figs. 2, 7 and 8 for the identity of the other samples). The blue bracket highlights the top part of the stratigraphy, in which the spread in ages is particularly visible since all calibrated ages (without stratigraphic modelling) are essentially indistinguishable. Note that in this area three samples (the three samples with the largest C.I.) are only partially constrained (cf. use of the function “After” in OxCal by Bayliss et al., 2015).

Download

4 Discussion: Stratigraphic constraints across different models and methods

Steier and Rom (2000) identified issues posed by Eq. (1) when used to model radiocarbon ages using OxCal. With the first sample set of the present study, we could illustrate similar problems arising when using the same equation, but using BayLum to model OSL samples in stratigraphy.

4.1 Çatalhöyük ages

The rest of the first part of this discussion is, therefore, dedicated to the comparison of models when dealing with the same dataset – the radiocarbon-dated sequence of Çatalhöyük. Figure 10 presents the comparison between OxCal, Chronomodel and BayLum. Not knowing the true age values since no historical sources attest to the occupations of the site of Çatalhöyük, it is difficult to make definitive statements. Nevertheless, several remarks can be made.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f10

Figure 10Comparison of radiocarbon ages modelled with OxCal, Chronomodel and Baylum (95 % C.I.).

Download

A first observation is that, for four samples towards the bottom (from OxA-9893 to OxA-9777), there is no overlap between the 95 % C.I. obtained with OxCal and neither that obtained with BayLum nor that estimated with Chronomodel. In other words, it is very likely that at least one model provides flawed results, i.e., the true age value lies outside the age interval estimated by at least one model.

Second, some samples display extremely small 95 % C.I. age intervals, for example UCIAMS-109991 and UCIAMS-109992 have an age uncertainty corresponding 23 and 24 years, respectively, according to OxCal – which corresponds to  0.05 % uncertainty (expressed as one quarter of the 95 % C.I. divided by the age, in our case  9000 years). Note that the length of these 95 % C.I. is only slightly larger according to BayLum: 56 and 33 years, respectively. In a period characterised by a plateau of the calibration curve, this observation should raise a flag about the validity of the model – such a precision level seems doubtful.

A third observation is that ages estimated with Chronomodel display significantly larger credible age intervals, i.e., greater uncertainties, compared to BayLum and OxCal. In other words, Chronomodel appears to lie on the cautious side of things – which is perhaps an advantage given our first observation, namely that at least one model gets (at least part of) the chronology wrong.

4.2 Çatalhöyük: start of the mound (onset of occupations)

Bayliss et al. (2015) were particularly interested in the chronology of the start of the mound, which has important archaeological implications. From this perspective, it is interesting to compare the radiocarbon calibrated ages with the statistical inferences obtained with OxCal, Chronomodel and BayLum (Figs. 11 and 12).

In OxCal, the boundary “start mound” is dated to between 7144 and 7079 BC (NB: we base our statements on the 95 % C.I. in this paragraph). According to Chronomodel, the earliest sample (sample OxA-9893) from the mound, i.e., the sample that is used to determine when occupations started, is dated to between 7487 and 7221 BC (95 % C.I.). Finally, according to BayLum, the age of this sample is dated to between 7373 and 7217 BC. Let us now compare these statistical inferences with the calibrated ages obtained for samples from the bottom of the mound. The bottom-most sample from the mound is sample OxA-9893, dated to 7326–7049 BC (calibrated age). As a consequence, the start of the mound cannot, at the 95 % credibility level, be earlier than 7326 BC (NB: in principle it can, because one cannot be sure that this sample is the very first remain left by people occupying the site; but any modelling solution earlier than that would by definition extrapolate the data). It is important to reiterate here our previous observation that this age interval is very similar to the age intervals of the seven overlying samples (Table 1). Based on this sample set, one may assert that the start of the mound cannot – at the 95 % credibility level – have occurred after 7077 BC (this is the younger age limit for sample UCIAMS-109994).

Table 1Calibrated ages of the eight earliest samples from the mound of Çatalhöyük, the two samples predating the mound (in italics), and comparison with statistical inferences (in bold) concerning the start of the mound according to OxCal, Chronomodel and BayLum.

Download Print Version | Download XLSX

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f11

Figure 11Bottom of the sequence: calibrated ages (using OxCal) in comparison with ages modelled with OxCal, Chronomodel and BayLum. Sample OxA-9893 (third from the bottom) is the oldest from the mound and thus serves as a proxy for the start of the mound in BayLum and Chronomodel. The onset of the mound estimated with OxCal is also shown.

Download

From this perspective, at the 95 % credibility level, (i) Chronomodel appears to suggest an age interval that is too old – about 50 % of this interval correspond to ages older than 7326 BC; (ii) The date for the start of the mound inferred with BayLum is the closest to the age interval, although both BayLum and Chronomodel exclude a “recent start” (after  7220 BC); (iii) OxCal completely excludes an “old start” for the mound, in striking contradiction with the data (see Fig. 12). Indeed, all individual ages from the bottom most layer have part of their 95 % C.I. extending beyond 7144 BC. The sample that appears most recent (OxA-23251), in the sense that the older boundary of its 95 % C.I. is the most recent date of all C.I. from this layer, has a 95 % C.I. extending from 7314 to 7048 BC. Yet, according to OxCal, at the 95 % credibility level, the phase cannot have started earlier than 7144 (the 95 % C.I. is 7144–7079 BC). In our view, this is clear evidence that, in this case, OxCal provides results that poorly reflect the available chronological information, i.e., the calibrated ages. It is interesting to look here at the two samples from below the mound; one might indeed assume that such samples would explain why OxCal excludes early ages for the start of the mound. However, their ages (7471–7047 and 7589–7186 BC) are not precise enough to exclude an early start of the mound. In other words, one is typically here in the case of an artificial gain in precision obtained with OxCal (as already highlighted by Steier and Rom, 2000), in contradiction with the data.

https://gchron.copernicus.org/articles/8/191/2026/gchron-8-191-2026-f12

Figure 12Comparison of calibrated radiocarbon ages (95 % credible intervals in black) with the start mound boundary, as estimated with OxCal (light green). All samples on this figure (except AA-27982 and PL-980525A, at the bottom) belong to the lowermost layer of the mound, i.e., correspond to the layer yielding the earliest human occupations of the site. Samples AA-27982 and PL-980525A pre-date these human occupations. The dashed purple line corresponds to 7314 BC. This is the earliest date all individual ages from the lowermost archaeological layer are consistent with, at the 95 % credibility level. Thus, excluding a start around or shortly after this date seems very unreasonable.

Download

4.3 Comparison of models: overview

Table 2 summarises our observations in the form of a list of modelling artefacts: all three models (OxCal, Chronomodel and BayLum) are affected by the spread effect due to what appears to be poor modelling of stratigraphic constraints (Eq. 1); OxCal is affected by the concentration effect because of the uniform phase model (Eq. 2); Chronomodel adds extra uncertainty that is not quantified during measurement (note that this may actually be an asset if analytical uncertainties were underestimated, or if an unknown source of error is at play).

Table 2List of modelling artefacts that may be encountered in chronological modelling, based on our study of radiocarbon dating at Çatalhöyük.

Download Print Version | Download XLSX

By definition, all models are wrong – a model is a mean to simplify experiments (real life) so that a theory can be designed. In addition, modelling choices come into play. As a result, no particular model is fundamentally bad or good, thus the purpose of this paper is not to tell readers to stop using all the above-mentioned models; it rather is to illustrate some modelling artefacts that should be given attention. In our second case study, let us first examine the data available to define the start of the mound (the intervals in black in Fig. 6 represent the 95 % credible interval of all ages relevant for this question). We observe (i) that, at the 95 % credibility level, the ages preceding the start of the mound fall between 7589 and 7186 (sample PL-980525A), and between 7471 and 7047 BC (sample AA-27982), with a near-uniform probability density (so all ages within this interval have similar probabilities; see Fig. 2); (ii) all 95 % C.I. for the samples from the first layer of the mound are very similar (Fig. 6) and are included in the interval [7461, or 7335 if one excludes ample PxA-9778; 7047] BC. From these observations, excluding a start of the mound around 7314 BC seems very unreasonable, because all samples from the archaeological layer are consistent with this age (Fig. 12), and because the only samples that might exclude such an early start, i.e. the samples below the archaeological layer, are perfectly consistent with an earlier age (with a probability of  35 % for the youngest sample AA-27982).

It follows that none of the three tested models gives a solution to the question of the start of the mound that is consistent with the data: OxCal because it gives a way too recent (and too precise) date, Chronomodel because it gives a high probability to a start of the mound much earlier than any measured age suggests – while excluding a recent start of the mound (compared to measured age ranges), and BayLum for the same reasons as Chronomodel (although to a marginal or negligible level regarding the extrapolation towards older ages). As frustrating as it may be, in our view none of the tested models can tell us anything better than the actual data themselves: the mound started to build between 7326 and 7077 BC. Therefore, the time gap identified by Bayliss et al. (2015) between the occupations in Boncuklu (dated to no later than 7500 BC – but even there, this date is open to discussion since the site's latest layers were truncated; see discussion in Bayliss et al., 2015) and Çatalhöyük may not have been as long as the 400 years claimed by these authors. At least, the data do not support their claim of a late ( 7100 BC) start for the occupations of Çatalhöyük.

5 Conclusion

All chronological models tested in this study – namely OxCal, Chronomodel and BayLum – are affected by the spread effect. This effect tends to create an estimation bias when modelling a short period with large uncertainties, i.e., when age uncertainties are large compared to the actual time differences between samples. This effect is particularly visible for the sequence of Çatalhöyük. Consequently, all three models provide statistical inferences that are, to some degree, inconsistent with the measurements. Of these three models, it appears that for Çatalhöyük, it is Baylum that provides the results that are more consistent with measurements. Notwithstanding the generality of this statement, our results suggest that other features of OxCal and Chronomodel explain their poorer performance: the phase model in OxCal, which concentrates ages inside a phase; and the event model in Chronomodel, which leads to increased uncertainty and thus better accuracy/robustness, but which increases the spread effect. This being said, BayLum clearly leads to aberrant results when modelling sequences of indistinguishable ages, as exemplified with the OSL dataset from Jingbian.

Overall, this article should be viewed as a warning message: when testing any chronological model, it is of utmost importance to compare the model outcome with the input data. Statistical artefacts that should be given special attention are the spread effect (common to OxCal, Chronomodel and BayLum), the concentration effect when using the phase model implemented in OxCal, and unwanted loss of precision when using the event model implemented in Chronomodel. Obviously, chronological modelling aims at improving the statistical inferences that can be attained from a given dataset – but it should not lead to inconsistent results.

Finally, comes the following fundamental question: what are we using Bayesian chronological models for? The general consensus it that they are implemented to make use of prior observations, such as stratigraphy, to refine the precision, accuracy and robustness of numerical chronologies. For example, one may hope to gain in precision without losing accuracy when including ordering constraints, by somehow correcting measurement errors. Our study shows that this goal is difficult to reach and that using models to correct measurements appears to be dangerous; at least, existing models of stratigraphic constraints appear to lead to problematic inferences when confronted with high-resolution datasets. Reducing uncertainties may indeed lead to a loss of robustness and to conclusions that are inconsistent with the data, at least for the examples studied here.

Code availability

All code used for this research is available as Supplement.

Data availability

All radiocarbon data were published by Bayliss et al. (2015) and are available in Supplement. OSL data were acquired by Buylaert et al. and published by Stevens et al. (2018). Therefore, these experimental data do not belong to the authors of the present manuscript. The authors express their gratitude to J.-P. Buylaert for sharing their data.

Supplement

The supplement related to this article is available online at https://doi.org/10.5194/gchron-8-191-2026-supplement.

Author contributions

Conceptualization: GG, AP. Formal analysis: GG, PGB, IB. Funding acquisition: GG. Investigation: all. Methodology: GG, IB, AP. Project administration: GG, AP. Software: all. Supervision: GG, AP. Visualization: all. Writing: GG, IB, AP.

Competing interests

The contact author has declared that none of the authors has any competing interests.

Disclaimer

Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. The authors bear the ultimate responsibility for providing appropriate place names. Views expressed in the text are those of the authors and do not necessarily reflect the views of the publisher.

Financial support

This research has been supported by the European Research Council, H2020 programme (grant no. 851793 awarded to G. Guérin, “Quina World” project).

Review statement

This paper was edited by Michael Dietze and reviewed by two anonymous referees.

References

Bayliss, A., Brock, F., Farid, S., Hodder, I., Southon, J., and Taylor, R. E.; Getting to the bottom of it all: a Bayesian approach to dating the start of Çatalhöyük, J. World Prehist., 28, 1–26, 2015. 

Bronk Ramsey, C.: Radiocarbon calibration and analysis of stratigraphy: the OxCal program, Radiocarbon, 37, 425–430, 1995. 

Bronk Ramsey, C.: Comment on “The use of Bayesian statistics for 14C dates of chronologically ordered samples: a critical analysis”, Radiocarbon, 42, 199–202, 2000. 

Bronk Ramsey, C.: Bayesian analysis of radiocarbon dates, Radiocarbon, 51, 337–360, 2009. 

Christophe, C., Philippe, A., Kreutzer, S., and Guerin, G.: BayLum: Chronological Bayesian Models Integrating Optically Stimulated Luminescence and Radiocarbon Age Dating. R package version 0.2.0, https://CRAN.R-project.org/package=BayLum (last access: 20 June 2024), 2020. 

Combès, B. and Philippe, A.: Bayesian analysis of individual and systematic multiplicative errors for estimating ages with stratigraphic constraints in optically stimulated luminescence dating, Quat. Geochronol., 39, 24–34, 2017. 

Guérin, G. Aldeias, V., Baumgarten, F., Goldberg, P., Gómez-Olivencia, A., Lahaye, C., Madelaine, S., Maureille, B., Philippe, A., Sandgathe, D., Talamo, S., Thomsen, K. J., Turq, A., and Balzeau, A.: A Third Neanderthal Individual from La Ferrassie Dated to the End of the Middle Palaeolithic, PaleoAnthropology, 2023, 98–118, 2023. 

Guérin, G., Lahaye, C., Heydari, M., Autzen, M., Buylaert, J.-P., Guibert, P., Jain, M., Kreutzer, S., Lebrun, B., Murray, A. S., Thomsen, K. J., Urbanova, P., and Philippe, A.: Towards an improvement of optically stimulated luminescence (OSL) age uncertainties: modelling OSL ages with systematic errors, stratigraphic constraints and radiocarbon ages using the R package BayLum, Geochronology, 3, 229–245, https://doi.org/10.5194/gchron-3-229-2021, 2021. 

Hajdas, I., Ascough, P., Garnett, M. H., Fallon, S. J., Pearson, C. L., Quarta, G., Spalding, K. L., Yamaguchi, H., and Yoneda, M.: Radiocarbon dating, Nature Reviews Methods Primers, 1, 62, https://doi.org/10.1038/s43586-021-00058-7, 2021. 

Lanos, P. and Dufresne, P.: ChronoModel version 2.0: Software for Chronological Modelling of Archaeological Data using Bayesian Statistics, https://chronomodel.com (last access: 29 September 2023), 2019. 

Lanos, P. and Philippe, A.: Hierarchical Bayesian modeling for combining dates in archeological context, Journal de la Société Française de Statistique, 158, 72–88, 2017. 

Lanos, P. and Philippe, A.: Event Date Model: A Robust Bayesian Tool for Chronology Building, Communications for Statistical Applications and Methods, 25, 131–57, 2018. 

McPherron, S. P., Talamo, S., Goldberg, P., Niven, L., S. D., Richards, M. P., Richter, D., Turq, A., and Dibble, H. L.: Radiocarbon dates for the late Middle Palaeolithic at Pech de l'Azé IV, France, J. Archaeol. Sci., 39, 3436–3442, 2012. 

Murray, A., Arnold, L. J., Buylaert, J. P., Guérin, G., Qin, J., Singhvi, A. K., Smedley, R., and Thomsen, K. J.: Optically stimulated luminescence dating using quartz, Nature Reviews Methods Primers, 1, 1–31, 2021. 

Nicholls, G. and Jones, M.: Radiocarbon dating with temporal order constraints, J. R. Stat. Soc. C-Appl., 50, 503–521, 2001.  

Philippe, A. and Vibet, M. A.: Analysis of archaeological phases using the r package archaeophases, J. Stat. Softw., 93, 1–25, 2020. 

Philippe, A., Guérin, G., and Kreutzer, S.: BayLum-An R package for Bayesian analysis of OSL ages: An introduction, Quat. Geochronol., 49, 16–24, 2019. 

Reimer, P. J., Austin, W. E., Bard, E., Bayliss, A., Blackwell, P. G., Ramsey, C. B., Butzin, M., Cheng, H., Edwards, R. L., Friedrich, M., and Grootes, P. M.: The IntCal20 northern hemisphere radiocarbon age calibration curve (0–55 cal kBP), Radiocarbon, 62, 725–757, 2020. 

Reimer, P., Bard, E., Bayliss, A., Beck, J. W., Blackwell, P. G., Ramsey, C. B., Buck, C. E., Cheng, H., Edwards, R. L., Friedrich, M., and Grootes, P. M.: IntCal13 and Marine13 Radiocarbon Age Calibration Curves 0–50,000 Years cal BP, Radiocarbon, 55, 1869–1887, 2013. 

Steier, P. and Rom, W.: The use of Bayesian statistics for 14C dates of chronologically ordered samples: a critical analysis, Radiocarbon, 42, 183–198, 2000. 

Stevens, T., Buylaert, J. P., Thiel, C., Újvári, G., Yi, S., Murray, A. S., Frechen, M., and Lu, H.: Ice-volume-forced erosion of the Chinese Loess Plateau global Quaternary stratotype site, Nat. Commun., 9, 983–983, 2018. 

1

Note: OSL and radiocarbon being the two most widely used dating methods for the study of the past 100 000 years, we will essentially focus on these methods in this article; nevertheless, all dating methods must be concerned by the problem.

2

Marginal differences seem to occur between the OxCal model of Bayliss et al. (2015) and ours, presumably because we used IntCal20 (Reimer et al., 2020) instead of IntCal13 (Reimer et al., 2013) used by Bayliss et al. (2015) to calibrate radiocarbon ages.

Download
Short summary
Bayesian modelling is often used to refine numerically dated chronological sequences, e.g., by making use of stratigraphic constraints. First, a high-resolution dataset based on luminescence dating is modelled with the dedicated R package BayLum. Then, three Bayesian modelling tools – namely BayLum, Chronomodel and OxCal – are compared using a high-resolution, radiocarbon dataset. Modelling artefacts are identified; the strengths and weaknesses of the models are discussed.
Share