We model the underlying shared calendar age density f(θ) as an infinite and unknown mixture of individual calendar age clusters/phases: f(θ) = w1Cluster1 + w2Cluster2 + w3Cluster3 + … Each calendar age cluster in the mixture has a normal distribution with a different location and spread (i.e., an unknown mean μj and precision τj2). Each object is then considered to have been drawn from one of the (infinite) clusters that together constitute the overall f(θ).
Such a model allows considerable flexibility in the estimation of the joint calendar age density f(θ) — not only allowing us to build simple mixtures but also approximate more complex distributions (see illustration below). In some cases, this mix of normal densities may represent true and distinct underlying normal archaeological phases, in which case additional practical inference may be possible. However this is not required for the method to provide good estimation of a wide range of underlying f(θ) distributions.
The probability that a particular sample is drawn from a particular cluster will depend upon the relative weight wj given to that specific cluster. It will be more likely that an object will come from some clusters than others. Given an object belongs to a particular cluster, its prior calendar age will then be normally distributed with the mean μj and precision τj2 of that cluster.
To estimate the shared calendar age density f(θ) based upon our set of 14C observations, X1, …, XN, we need to estimate:
This requires us to also calibrate the 14C determinations to obtain their calendar ages θ1, …, θN. Since we assume that the calendar ages of each object arise from the shared density, this must be performed simultaneously to the estimation of f(θ).
We use Markov Chain Monte Carlo (MCMC) to iterate, for k = 1, …, M, between:
After running the sampler for a large number of iterations (until we are sufficiently confident that the MCMC has converged) we obtain estimates for the calendar age θi of each sample, and an estimate for the shared calendar age density f̂(θ) from which they arose. These latter estimates of the shared calendar age density are called predictive estimates, i.e., they provide estimates of the calendar age of a hypothetical new sample (based on the set of N samples that we have observed).
Critically, with our Bayesian non-parametric method, the number of calendar age clusters that are represented in the observed data is unknown (and is allowed to vary in each MCMC step). This flexibility is different, and offers a substantial advantage, from other methods that require the number of clusters to be known a priori. For full technical details of the models used, and explanation of the model parameters, see Heaton (2022).
The MCMC updating is performed within an overall Gibbs MCMC scheme. There are two different schemes provided to update the DPMM — a Polya Urn approach (Neal 2000) which integrates out the mixing weights of each cluster; and a slice sampling approach in which they are explicitly retained (Walker 2007).
Run using the Polya Urn method (our recommended approach based upon testing):
polya_urn_output <- PolyaUrnBivarDirichlet(
rc_determinations = kerr$c14_age,
rc_sigmas = kerr$c14_sig,
calibration_curve = intcal20,
n_iter = 1e5,
n_thin = 5)
or the Walker method as follows:
walker_output <- WalkerBivarDirichlet(
rc_determinations = kerr$c14_age,
rc_sigmas = kerr$c14_sig,
calibration_curve = intcal20,
n_iter = 1e5,
n_thin = 5)
Note: This example runs the MCMC for our default choice of 100,000 iterations. However, as we discuss below, for this challenging dataset we need a greater number of iterations to be confident of convergence. We always suggest running the MCMC for at least 100,000 iterations to arrive at the converged results. However, for some complex datasets, longer runs may be required. More detail on assessing convergence of the MCMC can be found in the determining convergence vignette
Both of these methods will output a list containing the sampler outputs at every nthin iteration, with the values of the model parameters and the calendar ages.
Our sampler provides three outputs of particular interest.
The output data contains information to allow calculation of the
predictive distribution for the calendar age of a new, as yet
undiscovered, object. This density estimate summarises the calendar ages
of all objects. It is generated using the posterior sampled values of
the DPMM component of our MCMC sampler. This calendar age density can be
calculated and plotted using
PlotPredictiveCalendarAgeDensity()
. The pointwise mean of
f̂(θ) will be plotted,
together with a corresponding interval at (a user-specified) probability
level.
The function allows calculation using multiple outputs so that their results can be compared. For example below we compare the results from the two sampler methods above.
densities <- PlotPredictiveCalendarAgeDensity(
output_data = list(polya_urn_output, walker_output),
denscale = 2.5)
We also have the option to plot the SPD, to plot in the F14C scale, and to change the confidence intervals on the plot.
densities <- PlotPredictiveCalendarAgeDensity(
output_data = polya_urn_output,
denscale = 2.5,
show_SPD = TRUE,
interval_width = "bespoke",
bespoke_probability = 0.8,
plot_14C_age = FALSE)
Note: The fact that the two different MCMC samplers
do not provide matching probability intervals should flag to us that we
might not have reached convergence, and need to run the MCMC for longer.
Our investigations generally showed that
PolyaUrnBivarDirichlet()
is better at reaching convergence,
and so we recommend its use over WalkerBivarDirichlet()
. A
longer run of 1,000,000 iterations indicates that the plotted Polya Urn
output (in green) above is an accurate representation of the predictive
distribution.
Around 1176 cal yr BP, we see a substantial change between the 95% intervals for the summarised (predictive) estimate of the joint f(θ) (which have a large spike) and the 80% intervals (which do not and are smooth). This is not an error, but rather highlights a benefit of the DPMM method whereby the number of clusters needed to represent the data is allowed to vary. This feature occurs because the method is unsure if the observed data support an additional (highly-concentrated) cluster located around this time period. In some iterations of the MCMC, such a cluster will be included; but for the majority of iterations, the method believes it is not required. Since the plotted 80% interval does not contain the spike, but the 95% does, it is likely that the method thinks there is a 2.5–10% chance of such a distinct and highly-concentrated cluster (as this is the proportion of the MCMC iterations containing one). If more detailed inference is needed, one could look at the actual individual MCMC iterations to estimate how likely such a highly-concentrated cluster, resulting in a sudden spike in samples, is.
Aside: The sharp jump in the IntCal20 calibration curve at 1176 cal yr BP (774 cal AD) is due to an extreme solar particle event (ESPE) also known as a Miyake Event (Miyake et al. 2012).
The output data also includes the calendar age estimate for each
14C sample. We can use this to determine the posterior
distribution of the calendar age for each sample. Note that the calendar
age estimates use the joint information provided by all the
14C determinations (as opposed to solely the 14C
determination of the single object that would be found using
CalibrateSingleDetermination()
) on the understanding the
calendar ages of the objects are related.
You can calculate and plot this using
PlotCalendarAgeDensityIndividualSample()
- for example to
calculate the posterior calendar age distribution for the 21st
14C determination:
The highest posterior density range for a given probability and the
unmodelled density (i.e., the result of
CalibrateSingleDetermination()
) can also be shown on the
plot by specifying this in the arguments, as shown below.
The output data also contains information about the cluster allocation of each sampled object, which we can use to build the probability for there being a given number of total clusters. If we believe the underlying individual clusters in the model have inherent meaning in terms of representing genuine and distinct periods of site usage, as opposed to simply providing a tool to enable a non-parametric density estimate, this information may be archaeologically useful.
For those plotting functions which present calendar ages (i.e.,
PlotCalendarAgeDensityIndividualSample()
and
PlotPredictiveCalendarAgeDensity()
) we can change the
calendar age scale shown on the x-axis. The default is to plot on the
cal yr BP scale (as in the examples above). To instead plot in
cal AD, set plot_cal_age_scale = "AD"
; while for
cal BC, set plot_cal_age_scale = "AD"
, e.g.,
The current implementation of our Bayesian non-parametric approach only supports normally-distributed clusters as the components in the overall calendar age mixture distribution f(θ). While this still allows a great deal of flexibility in the modelling, as many distributions can be well approximated by normals, there are certain distributions f(θ) for which they will struggle. In particular, you cannot approximate a uniform phase well with a mixture of normal distributions.
If the underlying shared calendar age density f(θ) is close to a uniform
phase, or a mixture of uniform phases, then the current
(normally-distributed cluster component) Bayesian non-parametric method
is unlikely to work optimally and provide reliable summaries. In such
cases, we advise use of PPcalibrate()
. This alternative
approach is ideally suited to such situations.
The inhomogeneous Poisson process/changepoint approach taken by
PPcalibrate()
implicitly assumes a shared underlying
calendar age model for f(θ) that consists
precisely of an unknown mixture of uniform phases. Implementing
PPcalibrate()
and plotting the posterior rate of the
Poisson process (see
vignette) will provide an estimate of that shared calendar age
density.