Journal topic
Clim. Past, 15, 1999–2017, 2019
https://doi.org/10.5194/cp-15-1999-2019
Clim. Past, 15, 1999–2017, 2019
https://doi.org/10.5194/cp-15-1999-2019

Research article 17 Dec 2019

Research article | 17 Dec 2019

# Spiky fluctuations and scaling in high-resolution EPICA ice core dust fluxes

Spiky fluctuations and scaling in high-resolution EPICA ice core dust fluxes
Shaun Lovejoy1 and Fabrice Lambert2,3 Shaun Lovejoy and Fabrice Lambert
• 1Physics Department, McGill University, 3600 University St., Montréal, QC, H3A 2T8, Canada
• 2Department of Physical Geography, Pontificia Universidad Catolica de Chile, Vicuna Mackenna 4860, Santiago, Chile
• 3Millennium Nucleus Paleoclimate, Santiago, Chile

Correspondence: Fabrice Lambert (lambert@uc.cl)

Abstract

Atmospheric variability as a function of scale has been divided in various dynamical regimes with alternating increasing and decreasing fluctuations: weather, macroweather, climate, macroclimate, and megaclimate. Although a vast amount of data are available at small scales, the larger picture is not well constrained due to the scarcity and low resolution of long paleoclimatic time series. Using statistical techniques originally developed for the study of turbulence, we analyse the fluctuations of a centimetric-resolution dust flux time series from the EPICA Dome C ice core in Antarctica that spans the past 800 000 years. The temporal resolution ranges from annual at the top of the core to 25 years at the bottom, enabling the detailed statistical analysis and comparison of eight glaciation cycles and the subdivision of each cycle into eight consecutive phases. The unique span and resolution of the dataset allows us to analyse the macroweather and climate scales in detail.

We find that the interglacial and glacial maximum phases of each cycle showed particularly large macroweather to climate transition scale τc (around 2 kyr), whereas mid-glacial phases feature centennial transition scales (average of 300 years). This suggests that interglacials and glacial maxima are exceptionally stable when compared with the rest of a glacial cycle. The Holocene (with τc≈7.9 kyr) had a particularly large τc, but it was not an outlier when compared with the phases 1 and 2 of other cycles.

We hypothesize that dust variability at larger (climate) scales appears to be predominantly driven by slow changes in glaciers and vegetation cover, whereas at small (macroweather) scales atmospheric processes and changes in the hydrological cycles are the main drivers.

For each phase, we quantified the drift, intermittency, amplitude, and extremeness of the variability. Phases close to the interglacials (1, 2, 8) show low drift, moderate intermittency, and strong extremes, while the “glacial” middle phases 3–7 display strong drift, weak intermittency, and weaker extremes. In other words, our results suggest that glacial maxima, interglacials, and glacial inceptions were characterized by relatively stable atmospheric conditions but punctuated by frequent and severe droughts, whereas the mid-glacial climate was inherently more unstable.

1 Introduction

Over the late Pleistocene, surface temperature variability is strongly modulated by insolation, both at orbital (Jouzel et al., 2007) and daily timescales. In between these two scales, temperature variability has been shown to scale according to power law relationships, thus evidencing a continuum of variability at all frequencies (Huybers and Curry, 2006). However, although a vast amount of high-resolution data exist for modern conditions, our knowledge of climatic variability at glacial–interglacial timescales is usually limited by the lower resolution of paleoclimatic archive records, thus restricting high-frequency analyses during older time sections. Previous analyses using marine and terrestrial temperature proxies from both hemispheres suggest a generally stormier and more variable atmosphere during glacial times than during interglacials (Ditlevsen et al., 1996; Rehfeld et al., 2018).

One of the difficulties in characterizing climate variability is that ice core paleotemperature reconstructions rapidly lose their resolutions as we move to the bottom of the ice column. Figure 1 shows this visually for the EPICA Dome C Antarctic ice core temperature proxy (5787 measurements in all); the curve becomes noticeably smoother as we move back in time. In terms of data points, the most recent 100 kyr period has more than 3000 points (≈30-year resolution), whereas the most ancient 100 kyr period has only 137 (≈730-year resolution). This implies that while the most recent glacial–interglacial cycle can be perceived with reasonable detail, it is hard to compare it quantitatively to previous cycles or to deduce any general cycle characteristics.

Figure 1Temperature (blue) and dust flux (red) from the EPICA Dome C ice core (Jouzel et al., 2007; Lambert et al., 2012a). The dust flux time series has 32 000 regularly spaced points (25-year resolution); the temperature series has 5752 points. The temperature data are irregularly spaced and lose resolution as we go back into the past (number of temperature data points in successive ice ages: 3022, 1117, 521, 267, 199, 331, 134, 146). In both cases we can make out the glacial cycles, but they are at best only quasi-periodic.

Fluctuation analysis (Lovejoy, 2017; Lovejoy and Schertzer, 2013; Nilsen et al., 2016) gives a relatively simple picture of atmospheric temperature variability (Fig. 2). The figure shows a series of regimes each with variability alternately increasing and decreasing with scale. From left to right we see weather-scale variability, in which fluctuations tend to persist, building up with scale (they are unstable) and increasing up to the lifetime of planetary structures (about 10 d). This is followed by a macroweather regime with fluctuations tending to cancel each other out, decreasing with scale and displaying stable behaviour. In the last century, anthropogenically forced temperature changes (mostly from greenhouse gases) dominate the natural (internal macroweather) variability at scales longer than about 10–20 years. The figure shows that in pre-industrial periods, the lower-frequency climate regime starts somewhere between 100 and 1000 years (the macroweather–climate transition scale τc), indicating that different long-frequency processes become dominant. The macroweather–climate transition scale marks a change of regime where the dominant high-frequency processes associated with weather processes (and reproduced by GCMs, general circulation models, in control runs) give way to a different regime, where the variability is dominated by either the responses to external forcings or to slow internal sources of variability that were too weak to be important at higher frequencies. Further to the right of Fig. 2, we can see the broad peak associated with the glacial cycles at about 50 kyr (half the 100 kyr period), and then at very low frequencies, the megaclimate regime again shows increasing variability with scale. In between the climate and megaclimate regimes, the fluctuations decrease with scale over a relatively short range from about 100 to 500 kyr. However, the temperature fluctuations shown in Fig. 2 display average behaviour, which can potentially hide large variations from epoch to epoch. In this paper, we use a uniquely long and high-resolution paleo-dataset to analyse the macroweather and climate scales in detail.

Figure 2A composite showing root mean square (rms) Haar fluctuations (ΔT in units of C) black and rms dust fluctuations analysed in this paper (red, in units of milligrams per square metre per year; Lambert et al., 2012a). From left to right: thermistor temperatures at 0.0167 s resolution (Lovejoy, 2018), hourly temperatures from Landers, Wyoming (Lovejoy, 2015), daily temperatures from 75 N (Lovejoy, 2015), EPICA Dome C temperatures (Jouzel et al., 2007), and two marine benthic stacks (Veizer et al., 1999; Zachos et al., 2001). The macroweather–climate transition is not in phase between the different records because the left ones (industrial side) are influenced by anthropogenic climate change, while the right data are pre-industrial natural variability. As elsewhere in this paper, the fluctuations were multiplied by the canonical calibration constant of 2 so that when the slopes are positive, the fluctuations are close to difference fluctuations. The various scaling regimes are indicated at the bottom. Adapted from Lovejoy (2017).

We focus on the EPICA Dome C dust flux record, which has a 55 times higher resolution than the deuterium record, including high resolution over even the oldest cycle (Lambert et al., 2012a, Fig. 1). Antarctic dust fluxes are well correlated with temperature at orbital frequencies (Lambert et al., 2008; Ridgwell, 2003). But the fluxes are also affected by climatic conditions at the source and during transport (Lambert et al., 2008; Maher et al., 2010). The dust data used here can therefore be thought of as a more “holistic” climatic parameter that includes not only temperature changes but describes atmospheric variability as a whole (including wind strength and patterns and the hydrological cycle).

2 Method

In order to proceed to a further quantitative analysis of the types of statistical variability and of the macroweather–climate transition scale, we need to make some definitions. A commonly used way of quantifying fluctuations is the Fourier analysis. It quantifies the contribution of each frequency range to the total variance of the process. However, the interpretation of the spectrum is neither intuitive nor straightforward (Sect. 2.3). The highly non-Gaussian spikiness for both dust flux and its logarithm (e.g. Fig. 3b, c), implies strong – but stochastic – Fourier space spikes. Indeed, Lovejoy (2018) found that the probability distributions of spectral amplitudes can themselves be power laws. This has important implications for interpreting spectra, especially those estimated from single series (“periodograms”): if the spectral amplitudes are highly non-Gaussian, then we will typically see strong spectral spikes whose origin is purely random. This makes it very tempting to attribute quasi-oscillatory processes to what are in fact random spectral peaks. It therefore makes sense to consider the real (rather than Fourier) space variability (fluctuations). The problem here is that the spectrum is a second-order statistical moment (the spectrum is the Fourier transform of the autocorrelation function). While second-order moments are sufficient for characterizing the variability of Gaussian processes, in the more general and usual case – especially with the highly variable dust fluxes – we need to quantify statistics of higher orders, in particular, the higher-order statistics that characterize the extremes. Here, we will use two simple concepts to describe variability and intermittency (or spikiness) of the data.

The theoretical framework that we use in this paper is that of scaling, multifractals, and the outcome of decades of research attempting to understand turbulent intermittency. Intermittent, spiky transitions – characterized by different scaling exponents for different statistical moments – turn out to be the generic consequence of turbulent cascade processes. Although the cascades are multiplicative, the extreme probabilities generally turn out to be power laws (Mandelbrot, 1974; Schertzer and Lovejoy, 1987), not log-normals (as was originally proposed by Kolmogorov, 1962). The analyses are based on scaling regimes and their statistical characteristics. Because scaling is a symmetry (in this case invariance of exponents under dilations in time), in a dynamical regime in which two different components – such as temperature and dust – are strongly coupled parts of the system, each may have different scaling properties but both should respect the scale symmetry including the transition scale at which the symmetry breaks down. Therefore, the broad conclusions of our dust flux analyses – scaling regimes and their break points and stability or instability – are expected to be valid for the more usual climate parameters including the temperature. Although it is beyond our present scope, we will explore the scale-by-scale relationship between EPICA dust fluxes and temperatures in a future publication.

## 2.1 Haar fluctuations

The basic tool we use to characterize variability in real space of a series F(t) is the Haar fluctuation, which is simply the absolute difference of the mean over the first and second halves of an interval:

$\begin{array}{}\text{(1)}& \mathrm{\Delta }F\left(\mathrm{\Delta }t\right)=\frac{\mathrm{2}}{\mathrm{\Delta }t}\underset{t-\mathrm{\Delta }t/\mathrm{2}}{\overset{t}{\int }}F\left(t{}^{\prime }\right)\mathrm{d}t{}^{\prime }-\frac{\mathrm{2}}{\mathrm{\Delta }t}\underset{t-\mathrm{\Delta }t}{\overset{t-\mathrm{\Delta }t/\mathrm{2}}{\int }}F\left(t{}^{\prime }\right)\mathrm{d}t{}^{\prime }.\end{array}$

We can characterize the fluctuations by their statistics. For example, by analysing the whole dataset using intervals of various lengths, we can thus define the variability as a function of scale (i.e. interval length). If over a range of timescales Δt, there is no characteristic time, then this relationship is a power law, and the mean absolute fluctuation varies as

$\begin{array}{}\text{(2)}& 〈\left|\mathrm{\Delta }F\left(\mathrm{\Delta }t\right)\right|〉\propto \mathrm{\Delta }{t}^{H},\end{array}$

where “〈 〉” indicates ensemble average – here an average over all the available disjointed intervals. A positive H implies that the average fluctuations increase with scale. This situation corresponds to unstable behaviour identified with the climate regime. In contrast, when H is negative, variability converges towards a mean state with increasing scale. This is the situation found in the stable macroweather regime. Haar fluctuations are useful for the exponent range $-\mathrm{1}, which is valid for the dust series, and indeed for almost all geodata analysed to date.

More generally, we can consider other statistical moments of the fluctuations, the “generalized structure functions”, Sqt):

$\begin{array}{}\text{(3)}& {S}_{q}\left(\mathrm{\Delta }t\right)=〈{\left|\mathrm{\Delta }F\left(\mathrm{\Delta }t\right)\right|}^{q}〉\propto \mathrm{\Delta }{t}^{\mathit{\xi }\left(q\right)}.\end{array}$

If the fluctuations are from a Gaussian process, then their exponent function is linear: ξ(q)=qH. More generally however, ξ(q) is concave and it is important to characterize this, since the non-linearity in ξ(q) is due to intermittency, i.e. sudden, spiky transitions (for more details on Haar fluctuations and intermittency, we refer to Lovejoy and Schertzer, 2012). We therefore decompose ξ(q) into a linear and a non-linear (convex) part K(q), with K(1)=0:

$\begin{array}{}\text{(4)}& \mathit{\xi }\left(q\right)=qH-K\left(q\right),\end{array}$

so that K(q)=0 for quasi-Gaussian processes. Since the spectrum is a second-order moment, the spectrum of a scaling process at frequency ω is a power law:

$\begin{array}{}\text{(5)}& E\left(\mathit{\omega }\right)\approx {\mathit{\omega }}^{-\mathit{\beta }},\end{array}$

where the spectral exponent $\mathit{\beta }=\mathrm{1}+\mathit{\xi }\left(\mathrm{2}\right)=\mathrm{1}+\mathrm{2}H-K\left(\mathrm{2}\right)$; K(2) is therefore sometimes termed the “spectral intermittency correction”.

## 2.2 Intermittency

A simple way to quantify the intermittency is thus to compare the mean and root mean square (rms) Haar fluctuations:

$\begin{array}{}\text{(6)}& & {S}_{\mathrm{1}}\left(\mathrm{\Delta }t\right)=〈\left|\left(\mathrm{\Delta }F\left(\mathrm{\Delta }t\right)\right)\right|〉\propto \mathrm{\Delta }{t}^{\mathit{\xi }\left(\mathrm{1}\right)}=\mathrm{\Delta }{t}^{H},\text{(7)}& & {S}_{\mathrm{2}}{\left(\mathrm{\Delta }t\right)}^{\mathrm{1}/\mathrm{2}}={〈{\left(\mathrm{\Delta }F\left(\mathrm{\Delta }t\right)\right)}^{\mathrm{2}}〉}^{\mathrm{1}/\mathrm{2}}\propto \mathrm{\Delta }{t}^{\mathit{\xi }\left(\mathrm{2}\right)/\mathrm{2}}=\mathrm{\Delta }{t}^{H-K\left(\mathrm{2}\right)/\mathrm{2}},\end{array}$

with the ratio

$\begin{array}{}\text{(8)}& \begin{array}{rl}& {S}_{\mathrm{1}}\left(\mathrm{\Delta }t\right)/{S}_{\mathrm{2}}{\left(\mathrm{\Delta }t\right)}^{\mathrm{1}/\mathrm{2}}\\ & \phantom{\rule{0.25em}{0ex}}\phantom{\rule{0.25em}{0ex}}\phantom{\rule{0.25em}{0ex}}\phantom{\rule{0.25em}{0ex}}=〈\left|\mathrm{\Delta }F\left(\mathrm{\Delta }t\right)\right|〉/{〈{\left(\mathrm{\Delta }F\left(\mathrm{\Delta }t\right)\right)}^{\mathrm{2}}〉}^{\mathrm{1}/\mathrm{2}}\propto \mathrm{\Delta }{t}^{K\left(\mathrm{2}\right)/\mathrm{2}},\end{array}\end{array}$

where we estimate St) using all available disjointed intervals of size Δt. These expressions are valid in a scaling regime (Eq. 3). Since the number of disjoint intervals decreases as Δt increases, so does the sample size; hence, the statistics are less reliable at large Δt.

For theoretical reasons (Lovejoy and Schertzer, 2013; Schertzer and Lovejoy, 1987), it turns out that the intermittency near the mean (q=1) is best quantified by the parameter ${C}_{\mathrm{1}}=K{}^{\prime }\left(\mathrm{1}\right)$. Since K(1)=0 is a basic property, it turns out that for log-normal multifractals (approximately relevant here), the ratio exponent $K\left(\mathrm{2}\right)/\mathrm{2}\approx {C}_{\mathrm{1}}$.

While the mean-to-rms ratio is an intuitive statistic, it does not give a direct estimate of C1: a more accurate estimate of C1 uses the intermittency function Gt):

$\begin{array}{}\text{(9)}& \begin{array}{rl}& G\left(\mathrm{\Delta }t\right)=\underset{\mathrm{\Delta }q\to \mathrm{0}}{lim}〈\mathrm{\Delta }F〉{\left[\frac{〈\mathrm{\Delta }{F}^{\mathrm{1}-\mathrm{\Delta }q}〉}{〈\mathrm{\Delta }{F}^{\mathrm{1}+\mathrm{\Delta }q}〉}\right]}^{\mathrm{1}/\left(\mathrm{2}\mathrm{\Delta }q\right)}\\ & \phantom{\rule{0.25em}{0ex}}\phantom{\rule{0.25em}{0ex}}\phantom{\rule{0.25em}{0ex}}\phantom{\rule{0.25em}{0ex}}\propto \mathrm{\Delta }{t}^{\mathit{\xi }\left(\mathrm{1}\right)-\mathit{\xi }{}^{\prime }\left(\mathrm{1}\right)}=\mathrm{\Delta }{t}^{{C}_{\mathrm{1}}}\end{array}\end{array}$

whose exponent is C1. Equation (9) is exact in the limit Δq→0. The intermittency exponent C1 quantifies the rate at which the clustering near the mean builds up as a function of the range of scales over which the dynamical processes act; it only partially quantifies the spikiness. For this, we need other exponents, in particular the exponent qD that characterizes the tails of the probability distributions. This is because scaling in space and/or time generically gives rise to power law probability distributions (Mandelbrot, 1974; Schertzer and Lovejoy, 1987). Specifically, the probability (Pr) of a random dust flux fluctuation ΔF exceeding a fixed threshold s is

$\begin{array}{}\text{(10)}& \mathit{\text{Pr}}\left(\mathrm{\Delta }F>s\right)\approx {s}^{-{q}_{\mathrm{D}}};\phantom{\rule{0.25em}{0ex}}\phantom{\rule{0.25em}{0ex}}s\gg \mathrm{1},\end{array}$

where the exponent qD characterizes the extremes; for example, qD≈5 has been estimated for wind or temperature (Lovejoy and Schertzer, 1986) and for paleotemperatures (Lovejoy and Schertzer, 2013), whereas qD=3 for precipitation (Lovejoy et al., 2012). A qualitative classification of probability distributions describes classical exponentially tailed distributions (such as the Gaussian) as “thin-tailed”, log-normal (and log-Levy) distributions as “long-tailed”, and power law distributions as “fat-tailed”. Whereas thin and long-tailed distributions have convergence of all statistical moments, power distributions only have finite moments for orders q<qD.

## 2.3 How fluctuations help interpret spectra

Although spectra may be familiar, their physical interpretations are nontrivial, a fact that was underscored in Lovejoy (2015). In a scaling regime – a good approximation to the macroweather and climate regimes discussed here – the spectrum is a power law form (Eq. 5) where the spectral exponent β characterizes the spectral density. Although β tells us how quickly the variance changes per frequency interval, its physical significance is neither intuitive nor obvious. Integrating the spectrum over a frequency range is already easier to understand; it is the total variance of the process contributed by the range. Therefore, we already see that β−1 (the exponent of the integrated spectrum) is more directly relevant than β. But even to understand this, we need to consider whether over a range of frequencies, the process is dominated by either high or low frequencies. For this, we can compare the total variance contributed by neighbouring octaves. For a power law spectrum, the variance ratio of one octave to its neighbouring higher-frequency octave is 21−β. From this, we see that β>1 yields a ratio ${\mathrm{2}}^{\mathrm{1}-\mathit{\beta }}<\mathrm{1}$ implying low-frequency dominance, whereas when β<1, we have ${\mathrm{2}}^{\mathrm{1}-\mathit{\beta }}>\mathrm{1}$ and high-frequency dominance.

But what does low-frequency or high-frequency “dominance” mean physically? For this, it is easier to consider the situation in real space using fluctuations; the simplest relevant fluctuations are the Haar fluctuations ΔF discussed in Sect. 2.2 that vary with time interval Δt as ΔF≈ΔtH. We saw that the exponents in real and spectral space were simply related by $\mathit{\beta }=\mathrm{1}+\mathrm{2}H-K\left(\mathrm{2}\right)$, where K(2)>0 due to the spikiness (intermittency). This formula leads to two important conclusions. First, if we ignore intermittency (putting C1=0, hence K(2)=0) and assume that the mean fluctuations scale with the same exponent as the rms fluctuations, then $H=\left(\mathit{\beta }-\mathrm{1}\right)/\mathrm{2}$, showing again that it is the sign of β−1 that is fundamental: β>1 implies H>0; hence, fluctuations grow with scale and the process “drifts” or “wanders”; it is unstable. Conversely β<1 implies H<0; hence, fluctuations decrease with scale and the process “cancels” and “converges”; it is “stable”. The second conclusion is that if intermittency is strong (here we typically have C1≈0.1, K(2)≈0.2), then the relationship between the second- and first-order statistical moments is a little more complex so that for example, with these values and a β≈0.9, we would have high frequencies dominating the variance (β<1) but low frequencies dominating the mean (H>0).

## 2.4 Dust flux data

The dust flux data used in this study are based on a linear combination of insoluble particles, calcium, and non-sea-salt calcium concentrations (Lambert et al., 2012a). Because missing-data gaps in the three original datasets were linearly interpolated prior to the PCA (principal component analysis), high-frequency variability can sometimes be underestimated in short sections that feature a gap in one of the three original datasets. This occurs in about 25 % of all dust flux data points, although half of those are concentrated in the first 760 m of the core (0–43 kyr BP), when an older, less reliable dust-measuring device was used. Below 760 m these occurrences are evenly distributed and do not affect our analysis. Due to the sometimes slightly underestimated variability, the analysis shown here is a conservative estimate (Lambert et al., 2012a).

Unlike water isotopes that diffuse and lose their temporal resolution in the bottom section of an ice core at high pressures and densities, the relatively large dust particles diffuse much less and have been used to estimate the dust flux over every centimetre of the 3.2 km long EPICA core (298 203 valid data points; Lambert et al., 2012a). The temporal resolution of this series varies from 0.81 to 11.1 years (the averages over the most recent and the most ancient 100 kyr respectively). The worst temporal resolution of 25 yr cm−1 occurs around 3050 m depth, with the result that at that resolution, there are virtually no missing data points in the whole record (Fig. 1). We note, however, that the dust flux used here is a construct of concentrations at 1 cm resolution and accumulation rates at 55 cm resolution that were linearly interpolated to match the dust concentration resolution.

3 Results

## 3.1 Looking at the data

Unlike temperature for the water isotopes, polar dust flux records cannot be assigned to one particular atmospheric variable, like temperature for the water isotopes. At any given moment, the amount of dust deposited in East Antarctica will depend on the size and vegetation cover of the source region (mostly Patagonia for East Antarctic dust; Delmonte et al., 2008), on the amount of dust available in the source region (can depend on the presence of glaciers), on the strength of the prevailing winds between South America and Antarctica, and on the strength of the hydrological cycle (more precipitation will wash out more dust from the atmosphere; Lambert et al., 2008). Over large scales it is thought that temperature-driven moisture condensation may be the major process driving low-frequency variability (Markle et al., 2018), although that may not be true everywhere (Schüpbach et al., 2018). High- and low-frequency variability in the dust flux record is likely driven by different processes. For example, dust source conditions related to glaciers and vegetation cover may not have influenced high-frequency variability due to their relatively slow rate of change. On the other hand, volcanic eruption or extreme events related to the hydrological cycle may produce high-frequency signals in the record. A single dust peak within a low background may therefore reflect a short-term atmospheric disturbance like an eruption or drought over South America or low precipitation over the Southern Ocean. The analysis presented here focuses heavily on the occurrence of dust fluctuations, the physical interpretation of which will depend on the scale of the phenomenon.

Figure 3(a) Zooming out of the Holocene dust fluxes by octaves, by doubling the depth resolution from 1 cm (upper left) to 11 m (lower right) resolution. Starting at the left and moving to the right and from top to bottom (see the ellipses on the first three in the sequence), we zoom out by factors of 2 in depth maintaining exactly 290 data points (effectively non-dimensionalizing the depth; the small number of missing data points were not interpolated so that the final resolution is not exactly 210 cm = 10.24 m). The temporal resolution is not exactly doubled due to the squashing of the ice column; the total duration (in years) of each section is indicated in each plot; the average temporal resolution of plots is 0.24, 0.48, 0.98, 2.02, 4.32, 10.1 24.5, 54.1, 184, 434, and 2710 years. In order to fit all the curves on the same vertical scale, the dust fluxes were normalized by their mean over each segment. The means (in milligrams per square metre per year) are 0.44, 0.38, 0.30, 0.36, 0.35, 0.33, 0.34, 0.39, 2.48, 2.18, and 2.41, i.e. the first eight plots have nearly the same vertical scales, whereas the last three are about 6 times larger in range. This means that all the plots except the last three are at nearly constant normalization. (b) Same as (a) but for the absolute changes between neighbouring values in dust flux normalized by the corresponding mean over the segment (290 points). The horizontal lines indicate the Gaussian probability levels for $p=\mathrm{1}/\mathrm{290}$ (representing the mean extreme for a 290-point segment; full line) as well as $p={\mathrm{10}}^{-\mathrm{6}}$ (lower dashed line) and $p={\mathrm{10}}^{-\mathrm{9}}$ (upper dashed line). (c) Same as (a) but for the absolute changes between neighbouring values in the logarithms of dust flux normalized by the corresponding mean over the segment (290 points). The horizontal lines indicate the Gaussian probability levels for $p=\mathrm{1}/\mathrm{290}$ (representing the mean extreme for a 290-point segment; full line) as well as $p={\mathrm{10}}^{-\mathrm{5}}$ (lower dashed line) and $p={\mathrm{10}}^{-\mathrm{8}}$ (upper dashed line, not the same as in b).

Figure 3a shows a succession of 10 factor-of-2 “blow downs” (upper left to lower right at 11 different resolutions). In order to avoid smoothing, the data were “zoomed” in depth rather than time, but the point is clear: the signal is very roughly scale invariant, at no stage is there any sign of obvious smoothing, and the quasi-periodic 100 kyr oscillations are the only obvious timescale (we quantify this below). In comparison with more common paleoclimate signals, such as temperature proxies – which are apparently smoother but with spiky transitions – the dust flux itself is already quite spiky. However, it also displays spiky transitions. In Fig. 3b we show the absolute change in dust flux, and one can visually see the strong spikiness associated with strongly non-Gaussian variability: the intermittency. At each resolution, the solid line indicates the maximum spike expected if the process was Gaussian, and the upper dashed lines show the expected level for a (Gaussian) spike with probability 10−6. Again, without sophisticated analysis, we can see that the spikes are wildly non-Gaussian, frequently exceeding the 10−6 level even though each segment has only 290 points, with the spikiness being nearly independent of resolution.

Taking the logarithms of the dust flux is common practice since it reduces the extremes and makes the signal closer to the temperature and other more familiar atmospheric parameters. We therefore show the corresponding spike plot for the log-transformed data (Fig. 3c). Although the extreme spikes are indeed less extreme (see also Fig. 6a, b), we see that the transformation has not qualitatively changed the situation, with spikes still regularly exceeding (log-) Gaussian probability levels of 10−5 and occasionally 10−8.

## 3.2 Spectra

Figure 4 shows various spectral analyses (for the corresponding fluctuation analyses, see Fig. 5). There is a clear periodicity at about (100 kyr)−1. In the double power law fit (the smooth line plot, Eq. 11), the transition frequencies are a little lower: ${\mathit{\omega }}_{\mathrm{0}}=\left(\mathrm{160}\phantom{\rule{0.125em}{0ex}}\mathrm{kyr}{\right)}^{-\mathrm{1}}$ (flux) and ${\mathit{\omega }}_{\mathrm{c}}=\left(\mathrm{145}\phantom{\rule{0.125em}{0ex}}\mathrm{kyr}{\right)}^{-\mathrm{1}}$ (log flux), although a Gaussian fit near the max gives a spike at (94±9 kyr)−1. Note that it is actually a little bit “wide” (two peaks); hence, it is not perfectly periodic, and the amplitude is only about a factor 4 above the background. In comparison, the amplitude of the annual temperature frequency peak is several thousand times above the background (depending on the location) and is narrower (not shown).

Figure 4Log–log plot of the Fourier spectrum of the (25 year)−1 resolution dust concentration in frequency units of kyr−1 (red) and the same but of the logarithms of the flux (blue). Also shown is the average spectrum of the 5-year resolution data over the last 400 kyr (green). For the latter, the periodograms of each the four most recent 100 kyr cycles were averaged, but the full spectral resolution (5 years)−1 was retained. The beta parameters are the exponents of the theoretical spectrum (see main text, the negative of the logarithmic slope) for the macroclimate (−2.5), climate (1.7), and macroweather (0.8) regimes. The spectra were analysed using FFT (fast Fourier transform) with standard Hanning windows.

Figure 5The Haar fluctuation analysis of the entire 800 kyr dust flux dataset (thin lines). The dashed black and solid pink lines (top pair) represent rms fluctuations for dimensional and non-dimensional time respectively. The solid black and blue curves are the same but for the mean absolute (q=1) fluctuations. The curves with non-dimensional time lags have nominal (average) resolutions of 25 years and the fluctuation statistics are averaged over the eight cycles. The thick black line shows the Haar fluctuations for the most recent 400 kyr at a 5-year resolution. Note that the peak in the curves occurs as expected at Δt≈50 kyr, i.e. at about a half cycle; the horizontal dashed line shows that at this scale – corresponding to the largest difference in phases – the change in the mean absolute dust flux is about ±3 mg m−2 yr−1. Also shown (dashed vertical line) is the (average) timescale τc≈250 years at which the transition from macroweather to climate occurs. Several reference lines (with the slopes or exponents indicated) are shown showing approximate scaling behaviours.

Since this is a log–log plot, power laws appear as straight lines. We show in the figure the fits to the bi-scaling function

$\begin{array}{}\text{(11)}& E\left(\mathit{\omega }\right)=\frac{a}{{\left(\mathit{\omega }/{\mathit{\omega }}_{\mathrm{0}}\right)}^{{\mathit{\beta }}_{\mathrm{h}}}+{\left(\mathit{\omega }/{\mathit{\omega }}_{\mathrm{0}}\right)}^{{\mathit{\beta }}_{\mathrm{l}}}}\end{array}$

that smoothly transitions between a spectrum with $E\left(\mathit{\omega }\right)\approx {\mathit{\omega }}^{-{\mathit{\beta }}_{\mathrm{h}}}$ at ω>ω0 and $E\left(\mathit{\omega }\right)\approx {\mathit{\omega }}^{-{\mathit{\beta }}_{\mathrm{l}}}$ at ω<ω0. The figure shows the regressions with ${\mathit{\beta }}_{\mathrm{l}}=-\mathrm{2.5}$, βh=1.7, and a=7.5 (mg m−2 yr−1)2 yr, ${\mathit{\omega }}_{\mathrm{0}}\approx \left(\mathrm{145}\phantom{\rule{0.125em}{0ex}}\mathrm{kyr}{\right)}^{-\mathrm{1}}$ for the fluxes, and a=0.375 years−1, ${\mathit{\omega }}_{\mathrm{0}}\approx \left(\mathrm{160}\phantom{\rule{0.125em}{0ex}}\mathrm{kyr}{\right)}^{-\mathrm{1}}$ for the logarithms of fluxes. According to the figure, the high-frequency climate regime scaling continues to about (300 years)−1 before flattening to a very high-frequency scaling (βm≈0.8) “macroweather” regime (Lovejoy and Schertzer, 2013). The scaling exponents βh=1.7 and βm=0.8 corresponding to the climate and macroweather regime respectively may be compared with the values 2.1 and 0.4 for the EPICA paleotemperatures discussed in a future publication (compare, however, the red and black curves in Fig. 2). These results show that temperature and dust variability are of the same statistical type so that it is likely that the dust signal is a real climate signal – yet the significant differences in their exponents shows that it has a different information content.

The variability shown in Fig. 4 can be interpreted broadly or in detail. A clear feature is the spectral maximum at around (100 kyr)−1. The broad bispectral scaling model (Eq. 11) of the peak already accounts for 96 % of the spectral energy (variance), leaving only 4 % for the (extra) contribution from the (near-) (100 kyr)−1 orbital frequency (using the logarithm of the flux gives a similar result). Alternatively, with a narrow Gaussian-shaped spectral spike model, the spike is localized at (94±9 kyr)−1 and contributes a total of 31 % of the total variance. However, not all of this is above what we would expect from a scaling background; the exact amount depends on how the background is defined. For example, over the range from the 6th- to the 11th-highest frequencies in this discrete spectrum (from (133 kyr)−1 to (72 kyr)−1), in comparison to the background over this range, there is an enhancement of about 80 % due to the strong peaks (the enhancement is about 100 % for the 7th to the 12th frequencies). This means that, although the (94±9 kyr)−1 peak represents 31 % of the total variability over the range from (800 kyr)−1 to (25 years)−1, it is only about 15 % above the “background” (note that only 5 % of the total variance is between (25 years)−1 and (1 kyr)−1). We did not do the corresponding analysis for the (41 kyr)−1 obliquity frequency since Fig. 4 shows visually that it is barely discernable above the background.

The overall conclusion is that the background represents between 85 % and 96 % of the total variance.

## 3.3 Haar fluctuation analysis

Figure 5 shows the Haar fluctuations comparing their statistics for both dimensional and non-dimensional cycles as well as for the mean and rms fluctuations (bottom and top set of curves respectively). To start, let us consider the direct interpretations of the fluctuations in terms of the variability of the dust flux. Recall that when the fluctuations increase with scale, they represent typical differences, whereas when then decrease with scale, they represent typical anomalies (deviations from long-term mean values). For example, typical variations over a glacial–interglacial cycle (half cycle ≈50 kyr) are about ±3 mg m−2 yr−1 (i.e. a range of 6 mg m−2 yr−1, the dashed horizontal line), whereas typical variations at the 250-year minimum are $\approx ±\mathrm{0.5}$ mg m−2 yr−1.

The macroweather, climate, and macroclimate regimes noted in Fig. 4 are also clearly visible. In Fig. 5, we can clearly see the short regime with H<0 (up to about 250 years), a scaling regime with H>0 (up to glacial–interglacial periods ≈50 kyr), and finally a long-time (possibly scaling) decrease in variability. The spectral and real-space statistics are linked via the relation $\mathit{\beta }=\mathrm{1}+\mathit{\xi }\left(\mathrm{2}\right)$ (see Eqs. 4, 5). Starting with the high-frequency macroweather regime, the exponents $H=-\mathrm{0.05}$, K(2)≈0.10 correspond to β=0.8 (Fig. 4) and the real-space macroweather–climate transition scale (τc≈250 years) is close to the spectral transition scale ($\mathrm{1}/{\mathit{\omega }}_{\mathrm{c}}\approx \mathrm{300}$ years, Fig. 4). In the middle (climate) regime, the top (rms) curves (slope 0.33) implies ξ(2)=0.66, β=1.66, which is close to the corresponding exponent in Fig. 4 (βh=1.7). Finally, at the longest (macroclimate) scales, the low-frequency part of the spectrum in Fig. 4 (${\mathit{\beta }}_{\mathrm{l}}=-\mathrm{2.5}$) implies that the fluctuation exponent $H\approx \left({\mathit{\beta }}_{\mathrm{l}}-\mathrm{1}\right)/\mathrm{2}=-\mathrm{1.75}$. However, this is less than the minimum detectable by Haar analysis ($H=-\mathrm{1}$); therefore, we expect the far-right slope to equal −1 (as shown by a reference line). To correctly estimate this steep slope, one must use other definitions of fluctuations. We could also note that the climate–macroclimate transition timescale is broad and a little shorter than the value spectral value 1∕ωc estimated in Fig. 4.

Beyond confirming the results of the spectral analysis and allowing for direct interpretations of the fluctuation values in terms of typical fluxes, Haar analysis also quantifies the intermittency from the convergence of the rms and mean statistics at larger and larger timescales (see the clear difference in slopes shown in the climate regime: 0.38 versus 0.33). This underlines the limitation of spectral analysis discussed earlier: the fact that it is a second-order statistic that is only a partial characterization of the variability. Finally, the figure also shows that regardless of whether the cycles are defined in dimensional or in non-dimensional time, statistical characterizations (including the exponents) are virtually unaffected.

Figure 6a shows the fluctuation probabilities of the entire 800 kyr series at a 25-year resolution (here the fluctuations are simply taken as absolute differences at a 25-year resolution). We see that the large fluctuations (the tail) part of the distribution is indeed quite linear on a log–log plot, with exponents qD≈2.75 and 2.98 in time and depth respectively (both from fits to the extreme 0.1 % of the distributions). To get an idea of how extreme these distributions are, consider the depth distribution with qD=2.98. With this exponent, dust flux fluctuations 10 times larger than typical fluctuations occur only 102.98≈1000 times less frequently. In comparison, for a Gaussian, they would be ≈1023 times less likely; they would never be observed.

Figure 6(a) The probability distribution PrF>s) of random changes in dust flux (ΔF) exceeding a fixed threshold s in time at a 25-year resolution (brown; 32 000 points) and in depth at 1 cm resolution (black; 251 075 points corresponding to the last 400 kyr). The frequency scales on the right give the number (N) of jumps in each of the series that exceeds the threshold s. The straight lines indicate power law probability tails with exponents qD indicated. Also shown (parabolas) are the Gaussians with the same mean and standard deviations. In time, the maximum change in flux corresponds to about 28 standard deviations (i.e. to a Gauss probability $\approx {\mathrm{10}}^{-\mathrm{91}}$); in depth, it corresponds to 51 standard deviations (i.e. to $p\approx {\mathrm{10}}^{-\mathrm{455}}$). On the right, we provide axes giving the actual number of flux increments that exceed s (brown for the fluctuations in time; black for those in depth). (b) Same as (a) except for the increments of the log of the dust flux (brown is in time, 25-year resolution; black is in depth, 1 cm resolution). The curves are the closest-fitting (log-) Gaussians. The threshold S is dimensionless, and the numerical values are correct if F is measured in units of milligrams per square metre per year.

While the dust fluxes are always positive and so cannot be Gaussians, the increments analysed here could easily be approximately so. Nevertheless, a common way of trying to tame the spikes is by making a log transformation of fluxes. Figure 4 already showed that this did not alter the spectrum very much; here it similarly has only a marginal effect. For example, Fig. 6b shows that the extreme tails on the log dust flux distribution has qD=3.60 in time (25 years) and 4.59 in depth (at 1 cm resolution). The log-transformed variable still displays huge extremes with the extreme log flux corresponding to a log-Gaussian probability of 10−30 and 10−50 (time and depth respectively). Whether or not taking logarithms yields a more climate-relevant parameter, it does not significantly change the problem of intermittency or of the extremes.

We must mention the problem of estimating the uncertainties in the exponents. In the familiar case, we test a deterministic model and then uncertainty estimates are based on a stochastic model of the errors which are often assumed to be independent Gaussian random variables. In our case, the basic model is a stochastic one, and therefore one needs a stochastic model of the underlying process from which one can draw random time series. While our paper aims to provide a basis for the formulation of such a model, it is beyond our present scope. In order to obtain robust conclusions, we instead rely primarily on cycle-to-cycle comparisons, two different definitions of time (dimensional and non-dimensional) as well as a diversity of analysis techniques (spectral, fluctuation analysis, probability distributions). We should also mention that the use of fluxes (product of 1 cm concentrations and 55 cm accumulation rate) introduces an additional source of uncertainty due to the different time ranges contained in these sections at various depths. However, we prefer using the fluxes because they are more directly representative of climatic changes than concentrations.

However there are some results that are worth mentioning. For example, Lovejoy and Schertzer (2012) performed a numerical analysis of the uncertainties in first- and second-order exponent estimates obtained from Haar fluctuations of a universal multifractal model with C1=0.1 and a range of values of H (close to the value found here; see Figs. 9, 12). When the scaling ranges covered factors of about 1000, they found only a small bias (≈0.02) in estimates of H and a comparable uncertainty. However, in practice – such as the estimates here – the main source of uncertainty is the subjective choice of the scaling range itself: Fig. 9 shows that values of slopes depend on the region over which trends are fit, hence the straight reference lines.

Finally, for the problem of estimating probability tail exponents (qD; Fig. 6a, b), Clauset et al. (2009) found that the maximum likelihood method is optimal. However, they assumed that the range over which the power tail was valid was pre-determined. The real difficulty in Fig. 6a and b is that one must make an initial subjective choice about the exact range over which the exponent is estimated; using sophisticated estimators does not seem warranted.

## 3.4 Phases

Scaling is a statistical symmetry, a consequence of a time and space scaling symmetry of the underlying dynamics. Being statistical means that on average the statistics at small, medium and large scales are the same in some way (more precisely, it holds over a statistical ensemble). The difficulty is that on a single realization – such as that available here, i.e. a single core from a single planet earth – the symmetry will necessarily be broken. For example, in the spectrum in Fig. 4, in each of the proposed scaling regimes, scaling only predicts that the actual spectrum from this single core will vary about the indicated straight lines that represent the ensemble behaviour. Since this variability is strong, we made the potential scaling regimes more obvious by either averaging the spectrum over frequency bins (the red and blue spectra) or by breaking the series into shorter parts and averaging the spectra over all the parts, effectively treating each segment as a separate realization of a single process (green). In any event, all that any empirical analysis can show is that the data are consistent with the scaling hypothesis.

This already illustrates the general problem: in order to obtain robust statistics we need to average over numerous realizations, and since here we have a single series, the best we can do is to break the series into disjointed segments and average the statistics over them, assuming that the major underlying processes were constant over the last 800 000 years. Yet at the same time, in order to see the wide-range scaling picture (which also helps to more accurately estimate the scaling properties or exponents), we need segments that are as long as possible. The compromise that we chose between numerous short segments and a small number of long ones was to break the series into eight glacial–interglacial cycles and each cycle into eight successive phases. As a first approximation, we defined eight successive 100 kyr periods (hereafter called “segments”; Fig. 7a), corresponding fairly closely to the main periodicity of the series. As we discussed, the spectral peak is broad implying that the duration of each cycle is variable – the cycles are only “quasi-periodic”. It is therefore of interest to consider an additional somewhat flexible definition of cycles as the period from one interglacial to the next (hereafter called “cycle”; Fig. 7b). The break points were taken at interglacial optima: 0.4, 128.5, 243.5, 336, 407.5, 490, 614, 700, and 789 kyr BP, i.e. 96.9±18.7 kyr per cycle. Using the latter definition, the cycles were non-dimensionalized so that non-dimensional time was defined as the fraction of the cycle, effectively stretching or compressing the cycles by ±19 %.

Figure 7Panel (a): successive segments of theoretical 100 kyr long glacial cycles using usual (dimensional) time (present to past: bottom to top, the segment number is at the far right) with the 12.5 kyr phases indicated by vertical dashed lines. The short red lines indicate the interglacial dust minima. Each glacial–interglacial cycle is shifted by 25 units in the vertical for clarity. The red markers in (a) are mapped to the first dashed blue line in (b). Panel (b): successive cycles using non-dimensional time (interglacial to interglacial) and then shifted by one phase to better line up with the usual time segments (the left-most phase of the bottom line of (b) is zeroed). The average (nominal) resolution is 25 years. The interglacial dust minima were taken as 128.5, 243.5, 336, 407.5, 490, 614, 700, and 789 kyr BP, and the data start at 373 yr BP. Each cycle is shifted by 25 units in the vertical for clarity. The data older than 789 kyr BP were not used in these non-dimensional cycles.

With either of these definitions, we have eight segments or cycles, each with eight phases. Note that in our nomenclature, phases 1 and 8 are the youngest and oldest phases respectively and that time flows from phase 8 to phase 1. Figure 8 shows the phase-by-phase information summarized by the average flux over each cycle including the dispersion of each cycle about the mean (for the segments in panel a and the cycles in panel b). We see that the variability is highest in the middle of a cycle and lowest at the ends.

Figure 8Panel (a): averaging over the eight cycles at a 25-year resolution, we get the above picture: the mean is brown and the 1 standard deviation cycle-to-cycle variability is shown in red. The dashed vertical lines give a further division into 8×12.5 kyr segments, the eight “phases” of the cycle. Panel (b): the same but for the non-dimensional time. The relative position of the interglacial minimum at the first dashed line is indicated.

The spectra showed that there were wide-scale ranges that are on average scale–invariant power laws, and Fig. 4 quantifies the glacial–interglacial cycle. We are thus interested in characterizing the scaling properties over the different phases of the cycle; for this we turn to real-space statistics. In Fig. 9 we compare the statistics averaged over cycles and the statistics averaged over phases. The figure shows that the phase to phase differences are much more important than the cycle to cycle differences, in particular for the average fluctuations 〈(ΔFt))〉 (Fig. 9, lower left). We could also note that since the different cycles had quite similar statistics (panels b and d), this implies that there is no bias in the flux estimates with depth of the core.

Figure 9Panels (a, b) show the intermittency function Gt) (whose slope on the log–log plot is C1), and panels (c, d) show the mean absolute Haar fluctuation S1t) (whose slope on the log–log plot is H); (a, c) show the result for each phase after averaging over the eight cycles with the numbers next to each line indicate the phase number (each colour corresponds to the same number); (b, d) show the result for each cycle after averaging over the phases. Here, the same colours and numbers correspond to the cycle number; shown are only cycles 1, 4, and 8 to avoid clutter. Whereas each cycle is fairly similar to every other cycle (the b, d), each phase is quite different (a, c). We see that the most significant difference is the fluctuation amplitude as a function of phase (c).

From the global statistics (e.g. Figs. 4, 5), it is clear that in each glacial–interglacial cycle there are two regimes, so that before characterizing the structure functions by their exponents (e.g. H=ξ(1) for the mean fluctuations), we have to determine the macroweather–climate transition timescale τc whose average (from Figs. 4, 5) is 250–300 years.

One way of estimating the transition scale τc is to make a bilinear fit of log 10S1t) (i.e. Haar with q=1, the mean absolute fluctuation) with the mean slopes −0.05 (small Δt) and slope +0.35 (large Δt; the values were chosen because they are roughly the H estimates from the average over all the cycles) (Fig. 9). The hypothesis here was that there were two regimes, each characterized by a different exponent, each of which was estimated from the ensemble statistics. Therefore, the analysis only needed to estimate the scale at which the low-frequency process exceeded the high-frequency one. Bilinear fits were made for each phase of each segment (blue) as well as for each phase of each cycle (black). For each phase there were thus eight transition scales, which were used to calculate the mean and its standard deviation (shown here as representative black arrows). From the figure we see that at first (phases 8–3) the transition scale is relatively short (250–400 years) but that it rapidly moves to longer (1–2 kyr) scales for the final phases 2 and 1. The average transition scale over all phases is around 300 years.

The figure shows that our results are robust since the results are not very different using dimensional and non-dimensional time (segments and cycles). Comparing the blue and black curves, we see that in all cases the late phases have much larger τc than the early and middle phases. Also shown in Fig. 10 (dashed) is a plot of the break points estimated by a more subjective method that attempts to visually determine a break point on log S1log Δt plots. Again, we reach the same conclusion with quantitatively very similar results: a transition of millennia for phases 1 and 2 and a few centuries in the middle of the cycle. The cycle average value (τc≈300 years) is therefore not representative of the latest phases where τc is many times larger (glacial maxima and interglacials). The Holocene has an even larger transition scale (τc=7.9 kyr, marked by an X in Fig. 10), but it lies just outside the standard deviation of the first non-dimensional phases (red arrows in Fig. 10). Although the Holocene value of τc is the largest in phase 1, it corresponds to 1.55 standard deviations above the mean with (assuming a Gaussian variability) a corresponding p value of 0.12, roughly the expected extreme of a sample of eight; it is therefore not a statistical outlier.

Figure 10The transition scale τc estimated in two ways for each of the eight phases and from two definitions of the phases. The first method (solid lines) used a bilinear fit to the (logarithm) of the Haar q=1 structure function (i.e. mean absolute fluctuation) as a function of log time lag Δt. To obtain robust results, a small Δt region with the slope −0.05 and a large Δt slope +0.25 was imposed with the transition point (τc) determined by regression. This was done for each segment and cycle. For each phase there were thus eight transition scales, which were used to calculate the mean of the logarithm of τc and its standard deviation. Results are shown for dimensional (segments, blue) and non-dimensional time (cycles, black). The second method used to estimate τc was graphical and relied on a somewhat subjective fitting of scaling regimes and transitions, but without imposing small and large Δt slopes (exponents H). The results are shown in dashed lines; they are quite similar, although we can note some differences for the first phase (dimensional, blue) and the middle phases (non-dimensional, black). There is also considerable cycle-to-cycle spread that was quantified by the standard deviations. In order to avoid clutter, typical spreads are shown by the double headed black arrows. Dashed horizontal lines show the ensemble mean transition scale (about 250 years) as well as ensemble mean for phases 1 and 2 (around 2 kyr), which stands out compared to the rest of the phases. The red arrow shows 1 standard deviation for the non-dimensional first phases, while the X marks the value of the Holocene τc (7.9 kyr) just outside the 1σ limit.

Alternatively, rather than fixing a phase and determining the variation in the mean fluctuation and intermittency function (Fig. 9), we can consider the variation in the Haar fluctuations at fixed timescales and see how they vary from phase to phase (Fig. 11). The figure shows the phase-to-phase variation in Haar fluctuations at 50, 100, 200, 400, 800, 3300, and 7000 years scales (bottom to top; the dashed and solid lines alternate to demarcate the different curves; they are not uncertainties). Over the macroweather regime (up to about 400–800 years), the fluctuations tend to cancel so that the variability is nearly independent of timescale. In contrast, once we reach the longer scales in the climate regime (up to 7000 years), the fluctuations increase noticeably as the time interval Δt is increased. For every timescale, there is a clear cyclicity (left to right), with fluctuation amplitudes largest in the middle phases. We note that the cycle-to-cycle variability is fairly large; about a factor of 2 (for clarity the error bars indicating this cycle-to-cycle spread were not shown).

Figure 11Using non-dimensional time, the amplitude of the Haar fluctuations are averaged over all the cycles The curves from bottom to top are for timescales of Δt=50, 100, 200, 400, 800, 1600, 3500, and 7000 years, alternating solid and dashed lines (for clarity, only some of the Δt's are marked). The cycle-to-cycle variability (the dispersion around each line) is about a factor of 2 (it is not shown to avoid clutter).

Finally, we describe for each phase the drift tendency (H) and the intermittency C1 as well as fluctuation amplitude and extremeness of the data. In Fig. 12 we show the result on the non-dimensional phases over the range 500 years $<\mathrm{\Delta }t<\mathrm{3000}$ years (panels a and b; the range was chosen to be mostly in the climate regime, i.e. with Δt>τc, and it was fixed so as to avoid any uncertainty associated with the algorithm used to estimate τc). Recall that the fluctuation exponent H>0 quantifies the rate at which the average fluctuations increase with timescale. Similarly, the exponent C1 characterizes the rate at which the spikiness near the mean (the intermittency exponent) increases with scale. We see (panel a) that H is fairly high in the early phases with H reaching small value in the later phases (with H actually a little bit negative on average in phase 1 due to the large τc value). C1, on the other hand (panel b), decreases a bit in the middle the phases. The error bars show that there is quite a lot of cycle-to-cycle variability.

Figure 12The fluctuation and intermittency exponents H and C1 (a, b) are estimated over the range 500–3000 years, as a function phase with the standard deviations from the cycle-to-cycle variability (all using non-dimensional time). Panel (a) (H) shows low drift in phases 1 and 2 but becomes driftier in the middle and older phases. The intermittency (C1, b) is moderate at the beginning and end of the cycles and a little weaker in the middle. Panel (c) shows the amplitude of the fluctuations at 25 years determined by the standard deviation of the dust flux (units: milligrams per square metre per year). We see that the flux has low amplitude fluctuations at the beginning and end of the cycles and 3–4 times higher amplitude fluctuations in the middle. Panel (d) shows the probability exponent qD estimated from the 25-year resolution data for each phase; the extreme 5 % of the flux changes were used to determine the exponent in each phase; the cycle-to-cycle spread is indicated by the error bars (overall average over the phases: ${q}_{\mathrm{D}}=\mathrm{2.62}±\mathrm{0.42}$).

If H quantifies the “drift” and C1 the “spikiness”, then Fig. 12 shows that the early phases have high drift and medium spikiness, and the middle phases have high drift and lower spikiness, while phases 1–2 have low drift but medium spikiness. To understand this better, consider the transition timescales in Fig. 10. The youngest two phases with the low drift and spikiness are also the phase with the longest transition scales. This means that the rate at which the variability builds up is small and that it only builds up over a short range of scales (from τc to roughly Δt=50 kyr – the half cycle duration; this can be checked in Fig. 9, which shows the phase-by-phase structure functions and intermittency functions). Conversely, phases 3 and 4 with high drift and high intermittency also have a smaller τc so that both the fluctuations and spikiness build up faster (Fig. 11) and over a wider range of scales (Fig. 10).

Another useful characterization of the phases is to directly consider the flux variability at a fixed reference scale, taken here as the 25-year resolution; quantifying the amplitude of the variability of each segment by its standard deviation A at 25-year timescale (Fig. 12c). This is not the difference between neighbouring values or fluctuations (as in Fig. 11), it is rather the variability of the series itself at a 25-year resolution. For each of the phases, we have eight estimates (one from each cycle); these are used to calculate the mean (central solid line) and standard deviation shown by the error bars showing the cycle-to-cycle dispersion of the values. We can see that the amplitude of the 25-year-scale fluctuations is about 4 times higher in the middle of the ice age (phase 4) than at the interglacial (phase 1). The figure clearly shows the strong change in variability across the cycle.

Whereas C1 characterizes the intermittency near the mean, we have seen that the probability exponent qD characterizes the extreme spikiness. An extreme (low) exponent qD phase implies that most of the time the changes in flux are small, but occasionally there are huge transitions. Conversely, a high (less extreme) qD implies that there is a wider range of different flux changes so that most of the changes tend to be in a restricted range. Figure 12d compares qD phase by phase. If the value of qD is smaller, the extreme fluctuations are more and more extreme relative to the typical ones. Therefore, from the figure, we see that the extremes are stronger in the beginning and end of the cycle and somewhat less pronounced in the middle phases of the cycle (note that the overall mean is 2.62±0.42; this can be compared to the value qD=3.60 for the overall log-transformed data; Fig. 6b). Notice that for phase 8, qD=2.03; this is close to the value qD=2, below which the extremes are so strong that the variance (and hence spectrum) does not converge. Summarizing, we can now categorize the phase-by-phase spikiness as strong extremes and medium spikiness (phases 1, 2, 8) and intermediate extremes and low spikiness (phases 3–7). For the cycle-to-cycle estimates (not shown), the value ${q}_{\mathrm{D}}=\mathrm{2.75}±\mathrm{0.41}$ seems to be fairly representative of all the cycles, although there is a slight tendency for qD to decrease for the older cycles implying that they may have been a bit more extreme than the recent ones.

4 Discussion

An attractive aspect of dust fluxes is that they are paleo-indicators with unparalleled resolutions over huge ranges of temporal scales. However, they come with two difficulties. First, their dynamical interpretation is not unambiguous: they depend on temperature, wind, and precipitation; dust flux variability is hard to attribute to a specific process, and it is a holistic climate indicator. Second, their appearance as a sequence of strong spikes is unlike that of any of the familiar proxies. Indeed, we argue that their highly spiky (intermittent) nature (i.e. with C1>0) is outside the purview of conventional statistical frameworks including autoregressive, moving average, or more generally quasi-Gaussian or even quasi-log-Gaussian processes.

Due to the dominance of the continuum (spectral background) variability, physical interpretations must be based on an understanding of climate variability as a function of scale. We first consider overall analyses over the whole dust flux series and then focus on the phases. The spectral analysis (Fig. 4) is the most familiar, and for the dust fluxes, it is qualitatively similar to previous results obtained with temperature data, although temperature spectra with anything approaching the resolution of Fig. 4 are only possible over the most recent glacial cycle. The most striking spectral feature is the peak over the background at 100 kyr periodicity. The broadness of this peak already indicates the irregularity of the Earth system response to the eccentricity-forced orbital cycles. The (near-) absence of obliquity frequencies at 41 kyr is notable and is consistent with the corresponding analysis of paleotemperatures. Although there is definitely power in that frequency range, it is barely larger than the background continuum, suggesting a low response to that forcing. Finally, our high-resolution data allow us to discern two different power law regimes: one at low frequencies with an exponent β=1.7 and one at high frequencies with exponent β=0.8, with the transition between the two at around 300 years.

In Sect. 2.3, we discussed some of the difficulties inherent in interpreting spectra and showed that the exponent of the integrated spectrum β−1 is more directly relevant than β (ignoring intermittency, this is the same as the wandering or cancelling criterion H>0 or H<0). Applying this understanding to the dust exponents, we see that in macroweather, there is a weak high-frequency dominance ($\mathrm{1}-\mathit{\beta }\approx \mathrm{0.2}>\mathrm{0}$), whereas the climate regime is dominated by low frequencies ($\mathrm{1}-\mathit{\beta }\approx -\mathrm{0.7}<\mathrm{0}$). A plausible physical explanation is that over long periods of time (at climate regime scales), the amount of dust in the SH (Southern Hemisphere) atmosphere is driven by changes in glacier and vegetation coverage, which is itself forced by SH temperature change. There is therefore a very strong correlation between dust and temperature at climatic scales (Lambert et al., 2008). At higher (macroweather) frequencies, temperature oscillations are too fast to overcome the inertia of ice sheet and vegetation responses; dust and temperature correlations are very low. Instead, dust deposition in Antarctica will be more sensitive to temporary atmospheric disturbances in the winds and the hydrological cycle.

To interpret the analysis by the phase of the dust record (Fig. 12), one must understand the significance of A and of the exponents H, C1, and qD in the context of dust deposition. The H exponent and the amplitude A are directly linked to mean fluctuations and values, A being the standard deviation (〈F21∕2) of the dust flux variability at a fixed (here, 25-year) timescale, whereas H determines the rate at which the flux fluctuations ((〈ΔFt)21∕2)) change with timescale Δt. We saw that a positive H exponent signifies a tendency to drift, whereas when H<0, the dust fluctuations tend to cancel each other out and the record will cluster around a mean value. In contrast, H>0 indicates that the dust fluxes will not cluster around a mean value; in essence, the process wanders or drifts and does not stay constant; it appears to be unstable. The low H numbers during phases 1 and 2 (interglacial and glacial maxima) indicate a very constant, stable climatic state, with Patagonian dust production being either very low during interglacials (low glacier activity, large vegetation cover) or very high (Patagonian ice cap fully grown, large outwash plains on the Argentinian side). In contrast, the high H and amplitude A values during the mid-glacial may have been due to strong variability in glacier extent during that time (García et al., 2018; Sugden et al., 2009) and therefore a very variable dust supply (see also Fig. 11 that shows how the amplitude of the fluctuations at different timescales varies with the phase). The glacial inception (phases 7 and 8) features low A but a high H exponent. This implies that the mean dust level was highly variable, but the dust supply was still low, thus not allowing for large amplitude fluctuations. The higher amplitudes in phases 6 and 7 indicate that dust supply became abundant then. Since the Argentinian continental shelf was still submerged at that moment and the outwash plains not yet fully extended, the higher dust emissions may have been due to a transformation in vegetation cover about 30 kyr after glacial inception, possibly accompanied by changes in glacial and periglacial processes in the Andes.

The exponents C1 and qD are associated with the intermittency or spikiness of the data. C1 is a measure of the sparseness (or degree of clustering) of the mean-level spikes (i.e. whose amplitudes contribute most to the mean spike level). It is equal to one minus the fractal dimension of the set of spikes that exceed the mean level (${D}_{\mathrm{1}}=\mathrm{1}-{C}_{\mathrm{1}}$). qD characterizes how extreme the most extreme spike values are. The dust flux record is generally more intermittent (with sparser, more clustered spikes, larger C1) in phases 8, 1, and 2 (glacial inception, interglacial, glacial maximum) than in the mid-glacial, with also more extreme spike values (lower qD). These power law fluctuations implied by the low values of qD are so large that according to the classical assumptions, they would be outliers. While Gaussians are mathematically convenient and can be justified when dealing with measurement errors, in atmospheric science thanks to scaling, non-linear dynamics, very few processes are Gaussian. This has important applications in tipping point analysis, where noise-induced tipping points are generally studied using well-behaved white or Gaussian noise (e.g. Dakos et al., 2012).

The exponents characterize the variability of the dust signal over a wide range of scales. To understand the two scaling regimes, it may be helpful to recall that the ice core dust signal depends on both the variability of the dust source and that of the overall climate system. For example, a spike in the dust source and a fast change in the system state (e.g. Dansgaard–Oeschger – DO – events in the NH) could both produce a similar signal. However, fast changes in system state – such as the DO events in the NH – apparently do not occur in the SH where the corresponding signals are more triangular and gradual in shape. High-frequency variations in dust deposition (at scales in the macroweather regime) are thus likely to be dominated by dust source dynamics rather than ice sheet changes that have generally larger reaction times. One hypothesis is that the transition timescale τc is the scale at which the source variability that decreases with scale (H<0) becomes less than the system variability that increases with scale (H>0). The macroweather variability is therefore likely dominated by vegetation and/or atmospheric changes. Large-scale natural fires could alter the landscape in a very short time, allowing for more dust uptake by the winds and a sudden rise in atmospheric dust. The recuperation of vegetation cover would be more gradual, though, resulting in a sawtooth shape of the dust spike that we do not observe in the data. Similarly, it has been suggested that rapid climate change in the Northern Hemisphere (e.g. DO events) would have synchronously changed the Southern Hemisphere atmospheric circulation and wind belts (Buizert et al., 2018; Markle et al., 2017). This could again have quickly changed the source or transport conditions but would again have resulted in a sawtooth-shaped peak, either by steady regrowth of vegetation in the dust source areas or as climate conditions in the north Atlantic gradually return to stadial (Pedro et al., 2018).

Finally, we could mention volcanoes. Volcano eruptions usually saturated the dust-measuring device and were mostly cut from the record. Using the sulfate record to identify eruptions is tricky because many large sulfate peaks do not have a corresponding dust peak. This means that even if you do have matching dust and sulfate peaks, it could be an eruption or a coincidence. Therefore, the influence of volcanic variability on the results cannot be completely eliminated, although our key results are fairly robust with respect to the phase of the cycle and are therefore unlikely to be influenced by volcanic eruptions.

Although the spikes occur at all scales (see Fig. 3), the most likely explanation for the (shorter) macroweather-scale dust spikes is disturbances in the atmosphere, involving either the winds or the hydrological cycle (or both at the same time). The obvious candidate for a perturbation that would lead to increased dust in the atmosphere is drought. We will therefore interpret macroweather dust spikes as multiannual to multidecadal or multicentennial drought events in southern South America. With this interpretation, we can conclude that glacial maxima, interglacials, and glacial inceptions were characterized by more frequent and more severe drought events than during the mid-glacial. During glacial maxima, such extreme dust events could have contributed to Southern Hemisphere deglaciation by significantly lowering ice sheet albedo at the beginning of the termination (Ganopolski and Calov, 2011). In contrast, more frequent dust events could have contributed to glacial inception through negative radiative forcing of the atmosphere.

5 Conclusions

Until now, a systematic comparison of the different glacial–interglacial cycles has been hindered by a limitation of the most common paleoclimate indicators – the low resolution of Pleistocene temperature reconstructions from ice or marine sediment cores. Due to this intrinsic characteristic, the older cycles are poorly discerned; we gave the example of EPICA paleotemperatures whose resolution in the most recent cycle was 25 times higher than the resolution in the oldest one. In this paper, we therefore took advantage of the unique EPICA Dome C dust flux dataset with 1 cm resolution measuring 320 000 cm, whose worst time resolution over the whole core is 25 years.

Dust fluxes are challenging not only because of their high resolutions, but also because of their unusually high spikiness (intermittency) and their extreme transitions that occur over huge ranges of timescales. Standard statistical methodologies are inappropriate for analysing such data. They typically assume exponential decorrelations (e.g. autoregressive or moving average processes) that have variability confined to narrow ranges of scale. In addition, they assume that the variability is quasi-Gaussian or at least that it can be reduced to quasi-Gaussian through a simple transformation of variables (e.g. by taking logarithms). In this paper, using standard spectral and probability distribution analysis, we show that both the spectral and the probability tails were power laws, not exponential and requiring nonstandard approaches.

The high resolution of the data allowed us to not only quantitatively compare glacial–interglacial cycles with each other, but also to subdivide each cycle into eight successive phases that could also be compared to one another. One of the key findings was that there was a great deal of statistical similarity between the different cycles and that within each cycle there were systematic variations in the statistical properties with phase. These conclusions would not have been possible with the corresponding much lower-resolution temperature proxy data.

Our variability analysis using real-space (Haar) fluctuations confirmed that the majority of the variability was in the macroweather and climate scaling regime backgrounds with an average transition scale τc of about 300 years. In the climate regime (timescales above τc), dust variability is more affected by long-term hemispheric-wide climate changes affecting slow-response subsystems like glaciers and vegetation, which explains the high correlation of dust and temperature at these scales. In contrast, dust variability in the macroweather regime (timescales below τc) would have been more influenced by short-term atmospheric perturbations such as droughts and precipitation minima.

Using various techniques, τc was found to be systematically larger in the youngest two phases than in the middle and oldest phases; about 2 kyr but with nearly a factor of 4 cycle-to-cycle spread and equal to 300 years (with a factor of 2 spread) for the six remaining phases. For the Holocene, τc was found to be 7.9 kyr, which makes it an exceptionally stable interglacial, but not a statistical outlier compared to other interglacials. Similarly, the typical (rms) variation in flux amplitude was smaller in the early phase increases by (on average) a factor of 4 from ±0.13 to about ±0.5 mg m−2 yr−1 in the middle and later phases. The Holocene (with an amplitude of ±0.08 mg m−2 yr−1) was again particularly stable with respect to the phase 1 of other cycles, but it was not an outlier.

We addressed the task of statistically characterizing the cycles by primarily characterizing the phases' variability exponents H, C1, qD, and amplitude A. We show that the atmosphere was relatively stable during glacial maxima and interglacials, but highly variable during glacial inception and mid-glacial. However, the low amplitude of dust variability during glacial inceptions indicates that vegetation cover and dust production processes did not significantly change until ∼30 kyr after glacial inception.

We interpret the intermittency indicators as suggesting a higher frequency of drought events and more severe droughts during glacial inception, interglacials, and glacial maxima than during mid-glacial conditions. These short-term spikes in atmospheric dust could have helped trigger Southern Hemisphere deglaciation through albedo feedback of ice sheet surfaces or glacial inception through negative radiative forcing.

The results presented in this paper are largely empirical characterizations of a relatively less known source of climate data: dust fluxes. Dust flux statistics defy standard models: they require new analysis techniques and better physical models for their explanation. These reasons explain why our results may appear to be rough and approximate. Readers may nevertheless wonder why we did not provide standard uncertainty estimates. But meaningful uncertainties can only be made with respect to a theory and we have become used to theories that are deterministic, whose uncertainty is parametric and that arises from measurement error. The present case is quite different: our basic theoretical framework is rather a stochastic one; it implicitly involves a stochastic “earth process” that produces an infinite number of statistically identical planet earths of which we only have access to a single ensemble member. Unfortunately, we do not yet have a good stochastic process model from which we can infer sampling errors and uncertainties. In addition, from this single realization, we neglected measurement errors and estimated various exponents that characterized the statistical variability over wide ranges of timescale, realizing that the exponents themselves are statistically variable from one realization to the next. In place of an uncertainty analysis, we therefore quantified the spread of the exponents (which themselves quantify variability). In the absence of a precise stochastic model we cannot do much better.

This paper is an early attempt to understand this unique very high-resolution dataset. In future work, we will extend our methodology to the EPICA paleotemperatures and to the scale-by-scale statistical relationship between the latter and the dust fluxes.

Data availability
Data availability.

The dust flux data are available here: https://doi.org/10.1594/PANGAEA.779311 (Lambert et al., 2012b).

Author contributions
Author contributions.

Both authors analysed the data and contributed to the writing.

Competing interests
Competing interests.

The authors declare that they have no conflict of interest.

Acknowledgements
Acknowledgements.

Shaun Lovejoy's contribution to this fundamental research was unfunded. Fabrice Lambert acknowledges support by CONICYT projects Fondap 15110009, ACT1410, Fondecyt 1171773, and 1191223 and the Millennium Nucleus Paleoclimate. Millennium Nucleus Paleoclimate is supported by the Millennium Scientific Initiative of the Ministry of Economy, Development and Tourism (Chile).

Financial support
Financial support.

This research has been supported by the CONICYT (FONDAP, project nos. 15110009 and ACT1410), and FONDECYT (project nos. 1171773 and 1191223), the Millennium Nucleus Paleoclimate, supported by the Millennium Scientific Initiative of the Ministry of Economy, Development and Tourism (Chile).

Review statement
Review statement.

This paper was edited by Carlo Barbante and reviewed by Michel Crucifix and two anonymous referees.

References

Buizert, C., Sigl, M., Severi, M., Markle, B. R., Wettstein, J. J., McConnell, J. R., Pedro, J. B., Sodemann, H., Goto-Azuma, K., Kawamura, K., Fujita, S., Motoyama, H., Hirabayashi, M., Uemura, R., Stenni, B., Parrenin, F., He, F., Fudge, T. J., and Steig, E. J.: Abrupt ice-age shifts in southern westerly winds and Antarctic climate forced from the north, Nature, 563, 681–685, https://doi.org/10.1038/s41586-018-0727-5, 2018.

Clauset, A., Shalizi, C. R., and Newman, M. E. J.: Power-Law Distributions in Empirical Data, SIAM Rev., 51, 661–703, https://doi.org/10.1137/070710111, 2009.

Dakos, V., Carpenter, S. R., Brock, W. A., Ellison, A. M., Guttal, V., Ives, A. R., Kéfi, S., Livina, V., Seekell, D. A., van Nes, E. H., and Scheffer, M.: Methods for detecting early warnings of critical transitions in time series illustrated using simulated ecological data, PLoS One, 7, e41010, https://doi.org/10.1371/journal.pone.0041010, 2012.

Delmonte, B., Andersson, P. S., Hansson, M., Schöberg, H., Petit, J. R., Basile-Doelsch, I., and Maggi, V.: Aeolian dust in East Antarctica (EPICA-Dome C and Vostok): Provenance during glacial ages over the last 800 kyr, Geophys. Res. Lett., 35, L07703, https://doi.org/10.1029/2008GL033382, 2008.

Ditlevsen, P. D., Svensmark, H., and Johnsen, S.: Contrasting atmospheric and climate dynamics of the last-glacial and Holocene periods, Nature, 379, 810–812, https://doi.org/10.1038/379810a0, 1996.

Ganopolski, A. and Calov, R.: The role of orbital forcing, carbon dioxide and regolith in 100 kyr glacial cycles, Clim. Past, 7, 1415–1425, https://doi.org/10.5194/cp-7-1415-2011, 2011.

García, J. L., Hein, A. S., Binnie, S. A., Gómez, G. A., González, M. A., and Dunai, T. J.: The MIS 3 maximum of the Torres del Paine and Última Esperanza ice lobes in Patagonia and the pacing of southern mountain glaciation, Quaternary Sci. Rev., 185, 9–26, https://doi.org/10.1016/j.quascirev.2018.01.013, 2018.

Huybers, P. and Curry, W.: Links between annual, Milankovitch and continuum temperature variability, Nature, 441, 329–332, https://doi.org/10.1038/nature04745, 2006.

Jouzel, J., Masson-Delmotte, V., Cattani, O., Dreyfus, G., Falourd, S., Hoffmann, G., Minster, B., Nouet, J., Barnola, J. M., Chappellaz, J., Fischer, H., Gallet, J. C., Johnsen, S., Leuenberger, M., Loulergue, L., Luethi, D., Oerter, H., Parrenin, F., Raisbeck, G., Raynaud, D., Schilt, A., Schwander, J., Selmo, E., Souchez, R., Spahni, R., Stauffer, B., Steffensen, J. P., Stenni, B., Stocker, T. F., Tison, J. L., Werner, M., and Wolff, E. W.: Orbital and millennial Antarctic climate variability over the past 800,000 years, Science, 317, 793–796, https://doi.org/10.1126/science.1141038, 2007.

Kolmogorov, A. N.: A refinement of previous hypotheses concerning the local structure of turbulence in a viscous incompressible fluid at high Reynolds number, J. Fluid Mech., 13, 82–85, https://doi.org/10.1017/S0022112062000518, 1962.

Lambert, F., Delmonte, B., Petit, J., Bigler, M., Kaufmann, P., Hutterli, M., Stocker, T., Ruth, U., Steffensen, J., and Maggi, V.: Dust-climate couplings over the past 800,000 years from the EPICA Dome C ice core, Nature, 452, 616–619, https://doi.org/10.1038/nature06763, 2008.

Lambert, F., Bigler, M., Steffensen, J. P., Hutterli, M., and Fischer, H.: Centennial mineral dust variability in high-resolution ice core data from Dome C, Antarctica, Clim. Past, 8, 609–623, https://doi.org/10.5194/cp-8-609-2012, 2012a.

Lambert, F., Bigler, M., Steffensen, J. P., Hutterli, M., and Fischer, H.: Dust and calcium record in calculated for ice core EPICA Dome C, PANGAEA, https://doi.org/10.1594/PANGAEA.779311, 2012b.

Lovejoy, S.: A voyage through scales, a missing quadrillion and why the climate is not what you expect, Clim. Dynam., 44, 3187–3210, https://doi.org/10.1007/s00382-014-2324-0, 2015.

Lovejoy, S.: How scaling fluctuation analysis transforms our view of the climate, Past Global Changes Magazine, 25, 136–137, https://doi.org/10.22498/pages.25.3.136, 2017.

Lovejoy, S.: The spectra, intermittency and extremes of weather, macroweather and climate, Scientific Reports, 8, 12697, https://doi.org/10.1038/s41598-018-30829-4, 2018.

Lovejoy, S. and Schertzer, D.: Scale invariance in climatological temperatures and the local spectral plateau, Ann. Geophys., 4, 401–410, 1986.

Lovejoy, S. and Schertzer, D.: Haar wavelets, fluctuations and structure functions: convenient choices for geophysics, Nonlin. Processes Geophys., 19, 513–527, https://doi.org/10.5194/npg-19-513-2012, 2012.

Lovejoy, S. and Schertzer, D.: The Weather and Climate: Emergent Laws and Multifractal Cascades, Cambridge University Press, Cambridge, 2013.

Lovejoy, S., Pinel, J., and Schertzer, D.: The global space–time cascade structure of precipitation: Satellites, gridded gauges and reanalyses, Adv. Water Resour., 45, 37–50, https://doi.org/10.1016/J.ADVWATRES.2012.03.024, 2012.

Maher, B. A., Prospero, J. M., Mackie, D., Gaiero, D., Hesse, P. P., and Balkanski, Y.: Global connections between aeolian dust, climate and ocean biogeochemistry at the present day and at the last glacial maximum, Earth-Sci. Rev., 99, 61–97, https://doi.org/10.1016/j.earscirev.2009.12.001, 2010.

Mandelbrot, B. B.: Intermittent turbulence in self-similar cascades: divergence of high moments and dimension of the carrier, J. Fluid Mech., 62, 331–358, https://doi.org/10.1017/S0022112074000711, 1974.

Markle, B. R., Steig, E. J., Buizert, C., Schoenemann, S. W., Bitz, C. M., Fudge, T. J., Pedro, J. B., Ding, Q., Jones, T. R., White, J. W. C., and Sowers, T.: Global atmospheric teleconnections during Dansgaard-Oeschger events, Nat. Geosci., 10, 36–40, https://doi.org/10.1038/ngeo2848, 2017.

Markle, B. R., Steig, E. J., Roe, G. H., Winckler, G., and McConnell, J. R.: Concomitant variability in high-latitude aerosols, water isotopes and the hydrologic cycle, Nat. Geosci., 11, 853–859, https://doi.org/10.1038/s41561-018-0210-9, 2018.

Nilsen, T., Rypdal, K., and Fredriksen, H.-B.: Are there multiple scaling regimes in Holocene temperature records?, Earth Syst. Dynam., 7, 419–439, https://doi.org/10.5194/esd-7-419-2016, 2016.

Pedro, J. B., Jochum, M., Buizert, C., He, F., Barker, S., and Rasmussen, S. O.: Beyond the bipolar seesaw: Toward a process understanding of interhemispheric coupling, Quaternary Sci. Rev., 192, 27–46, https://doi.org/10.1016/j.quascirev.2018.05.005, 2018.

Rehfeld, K., Münch, T., Ho, S. L., and Laepple, T.: Global patterns of declining temperature variability from the Last Glacial Maximum to the Holocene, Nature, 554, 356–359, https://doi.org/10.1038/nature25454, 2018.

Ridgwell, A. J.: Implications of the glacial CO2 “iron hypothesis” for Quaternary climate change, Geochem. Geophy. Geosy., 4, 1076, https://doi.org/10.1029/2003GC000563, 2003.

Schertzer, D. and Lovejoy, S.: Physical modeling and analysis of rain and clouds by anisotropic scaling multiplicative processes, J. Geophys. Res., 92, 9693–9714, https://doi.org/10.1029/JD092iD08p09693, 1987.

Schüpbach, S., Fischer, H., Bigler, M., Erhardt, T., Gfeller, G., Leuenberger, D., Mini, O., Mulvaney, R., Abram, N. J., Fleet, L., Frey, M. M., Thomas, E., Svensson, A., Dahl-Jensen, D., Kettner, E., Kjaer, H., Seierstad, I., Steffensen, J. P., Rasmussen, S. O., Vallelonga, P., Winstrup, M., Wegner, A., Twarloh, B., Wolff, K., Schmidt, K., Goto-Azuma, K., Kuramoto, T., Hirabayashi, M., Uetake, J., Zheng, J., Bourgeois, J., Fisher, D., Zhiheng, D., Xiao, C., Legrand, M., Spolaor, A., Gabrieli, J., Barbante, C., Kang, J. H., Hur, S. D., Hong, S. B., Hwang, H. J., Hong, S., Hansson, M., Iizuka, Y., Oyabu, I., Muscheler, R., Adolphi, F., Maselli, O., McConnell, J., and Wolff, E. W.: Greenland records of aerosol source and atmospheric lifetime changes from the Eemian to the Holocene, Nat. Commun., 9, 1476, https://doi.org/10.1038/s41467-018-03924-3, 2018.

Sugden, D. E., McCulloch, R. D., Bory, A. J.-M., and Hein, A. S.: Influence of Patagonian glaciers on Antarctic dust deposition during the last glacial period, Nat. Geosci., 2, 281–285, https://doi.org/10.1038/ngeo474, 2009.

Veizer, J., Ala, D., Azmy, K., Bruckschen, P., Buhl, D., Bruhn, F., Carden, G. A. F., Diener, A., Ebneth, S., Godderis, Y., Jasper, T., Korte, C., Pawellek, F., Podlaha, O. G., and Strauss, H.: 87Sr/86Sr, δ13C and δ18O evolution of Phanerozoic seawater, Chem. Geol., 161, 59–88, https://doi.org/10.1016/S0009-2541(99)00081-9, 1999.

Zachos, J., Pagani, M., Sloan, L., Thomas, E., and Billups, K.: Trends, rhythms, and aberrations in global climate 65 Ma to present, Science, 292, 686–693, https://doi.org/10.1126/science.1059412, 2001.