Skip to main content

Advertisement

Log in

An evaluation of experimental decadal predictions using CCSM4

  • Published:
Climate Dynamics Aims and scope Submit manuscript

Abstract

This study assesses retrospective decadal prediction skill of sea surface temperature (SST) variability in initialized climate prediction experiments with the Community Climate System Model version 4 (CCSM4). Ensemble forecasts initialized with two different historical ocean and sea-ice states are evaluated and compared to an ensemble of uninitialized coupled simulations. Both experiments are subject to identical twentieth century historical radiative forcings. Each forecast consists of a 10-member ensemble integrated over a 10-year period. One set of historical ocean and sea-ice conditions used for initialization comes from a forced ocean-ice simulation driven by the Coordinated Ocean-ice Reference Experiments interannually varying atmospheric dataset. Following the Coordinated Model Intercomparison Project version 5 (CMIP5) protocol, these forecasts are initialized every 5 years from 1961 to 1996, and every year from 2000 to 2006. A second set of initial conditions comes from historical ocean state estimates obtained through the assimilation of in-situ temperature and salinity data into the CCSM4 ocean model. These forecasts are only available over a limited subset of the CMIP5 recommended start dates. Both methods result in retrospective SST prediction skill over broad regions of the Indian Ocean, western Pacific Ocean and North Atlantic Ocean that are significantly better than reference skill levels from a spatio-temporal auto-regressive statistical model of SST. However the subpolar gyre region of the North Atlantic stands out as the only region where the CCSM4 initialized predictions outperform uninitialized simulations. Some features of the ocean state estimates used for initialization and their impact on the forecasts are discussed.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Notes

  1. This idea has been challenged in recent years by the Deser et al. (2012a) study demonstrating that internal climate variability can influence long-term trends on regional spatial scales.

  2. The CMIP5 protocol calls for predictions to be “initialized toward the end of 1960, 1965 ...” (Taylor et al. 2012). In this regard, the HDInit experiments follow the spirit of the protocol, if not the exact specifications.

  3. The 1-year offset for the first five DAInit experiments was done to maximize the chance of detecting the mid-1970’s climate shift in the Pacific.

  4. This statement follows from the well-known statistical property of normally distributed variables: Given k independent samples from a normally distributed random variable with variance \(\sigma ^2\), the variance of the sample ensemble mean is \(\frac{\sigma ^2}{k}.\)

  5. Variability on timescales less than 4 years is excluded from this analysis because it is standard practice in the decadal prediction community to evaluate skill in terms of 4-year means (Goddard 2013).

  6. The mid-1970’s climate shift was characterized in the observations by a shift in the long-term Pacific SST toward warmer tropical SST and cooler mid-latitude SST.

  7. For experiments evaluated over the 1975–2006 period this is six 4-year averages at five year intervals followed by six 4-year averages at one year interval.

References

  • Anderson J (2001) An ensemble adjustment Kalman filter for data assimilation. Mon Wea Rev 129:2884–2902

    Article  Google Scholar 

  • Anderson J, Hoar T, Raeder K, Liu H, Collins N, Torn R, Avellano A (2009) The data assimilation research testbed. Bull Am Meteorol Soc 90:1283–1296

    Article  Google Scholar 

  • Bowman A, Azzalini A (1997) Applied smoothing techniques for data analysis. Oxford University Press, New York

    Google Scholar 

  • Branstator G, Teng H (2010) Two limits of initial-value decadal predictability in a CGCM. J Clim 23:6292–6311

    Article  Google Scholar 

  • Cane M (2010) Decadal predictions in demand. Nature Geosci 3:231–232

    Article  Google Scholar 

  • Chikamoto Y et al (2013) An overview of decadal climate predictability in a multimodel ensemble by climate model MIROC. Clim Dyn 40:1201–1222

    Article  Google Scholar 

  • CLIVAR (2011) Data and bias correction for decadal climate predictions. CLIVAR publication series 150, International CLIVAR Project Office

  • Danabasoglu G et al (2014) North Atlantic simulations in coordinated ocean-ice reference experiments phase II (CORE-II). Part I: mean states. Ocean Model 73:76–107

    Article  Google Scholar 

  • Danabasoglu G, Bates S, Briegleb B, Jayne SR, Jochum M, Large W, Peacock S, Yeager S (2012) The CCSM4 ocean component. J Clim 25:1361–1389

    Article  Google Scholar 

  • Deser C, Knutti R, Solomon S, Phillips A (2012a) Communication of the role of natural variability in future North American climate. Nature Clim Change 2:775–779

    Article  Google Scholar 

  • Deser C, Phillips A, Bourdette V, Teng H (2012b) Uncertainty in climate change projections: the role of internal variability. Clim Dyn 38:527–546

    Article  Google Scholar 

  • Doblas-Reyes F et al (2013) Initialized near-term regional climate change prediction. Nature Commun 1–9. doi:10.1038/ncomms2704

  • Efron B, Tibshirani R (1993) An introduction to the bootstrap. Chapman & Hall, London

    Book  Google Scholar 

  • Gangsto R, Weigel A, Liniger M, Appenzeller C (2013) Methodological aspects of the validation decadal predictions. Clim Res 55:181–2000

    Article  Google Scholar 

  • Garcia-Serrano J, Doblas-Reyes F (2012) On the assessment of near-surface global temperature and North Atlantic multi-decadal variability in the ENSEMBLES decadal hind cast. Clim Dyn 39:2025–2040

    Article  Google Scholar 

  • Gent P et al (2011) The community climate system model version 4. J Clim 24:4973–4991

    Article  Google Scholar 

  • Goddard L et al (2013) A verification framework for interannual-to-decadal predictions experiments. Clim Dyn 40:245–272

    Google Scholar 

  • Hawkins E, Sutton R (2009) The potential to narrow uncertainty in regional climate predictions. Bull Am Meteorol Soc 90:1095–1107

    Article  Google Scholar 

  • Hazeleger W, Wouters B, van Oldenborgh G, Corti S, Palmer T, Smith D, Dunstone N, Kroger J, Pohlmann H, von Storch J (2013) Predicting multiyear North Atlantic Ocean variability. J Geophys Res Oceans 118:1087–1098

    Article  Google Scholar 

  • Jin E et al (2008) Current status of ENSO prediction skill in coupled ocean-atmosphere models. Clim Dyn 31:647–664

    Article  Google Scholar 

  • Karspeck A et al (2013) An ensemble adjustment Kalman Filter for the CCSM4 ocean component. J Clim 26:7392–7413

    Article  Google Scholar 

  • Keenlyside N, Latif M, Kornblueh L, Roeckner E (2008) Advancing decadal-scale climate prediction in the North Atlantic sector. Nature 453:84–88

    Article  Google Scholar 

  • Kirtman B, Shaklee J, Balmaseda M, Graham N, Penland C, Xue Y, Zebiak X (2001) Current status of ENSO forecast skill: a report to the CLIVAR working group on seasonal to interannual prediction. CLIVAR publication series 56, International CLIVAR Project Office

  • Knight J, Fold C, Scaife A (2006) Climate impacts of the Atlantic multidecadal oscillation. Geophys Res Lett 33(L17):706

    Google Scholar 

  • Kröger J, Müller W, vonStorch J (2012) Impact of different ocean reanalyses on decadal climate prediction. Clim Dyn 39:795–810

    Article  Google Scholar 

  • Large W, Yeager S (2009) The global climatology of an interannually varying air-sea flux data set. Clim Dyn 33:341–364. doi:10.1007/s00382-008-0441-3

    Article  Google Scholar 

  • Liu Z (2012) Dynamics of interdecadal climate variabiltiy: a historical perspective. J Clim 25:1963–1995

    Article  Google Scholar 

  • Mantua N, Hare S, Wallace YZJ (1997) A Pacific interdecadal climate oscillation with impact on salmon production. Bull Am Meteorol Soc 78:1069–1079

    Article  Google Scholar 

  • Matei D, Pohnmann H, Jungclas J, Müller W, Haak H, Marotzke J (2012) Two tales of initializing decadal climate prediction experiments wit the echam5/mpi-om model. J Clim 25:8502–8522

    Article  Google Scholar 

  • McCabe G, Palecki M, Betancourt J (2004) Pacific and Atlantic Ocean influences on multidecadal drought frequency in the United States. Proc Natl Acad Sci 101(12):4136–4141

    Article  Google Scholar 

  • Meehl G, Teng H (2012) Case studies for the initialized decadal hindcasts and predictions for the pacific regions. Geophys Res Lett 39(L22):705. doi:10.1029/2012GL053,423

    Google Scholar 

  • Meehl G, Hu A, Arblaster J, Fasullo J, Trenberth K (2013) Externally forced and internally generated decadal climate variability associated with the interdecadal Pacific oscillation. J Clim. doi:10.1175/JCLI-D-12-00,548.1

  • Mochizuki T, Chikamoto Y, Kimoto M, Ishii M, Tatebe H, Komuro Y, Sakamoto TT, Watanabe M, Mori M (2012) Decadal prediction using a recent series of MIROC global climate models. J Meteorol Soc Jpn 90A:373–383

  • Moss R et al (2010) The next generation of scenarios for climate change research and assessment. Nature 463:747–756

    Article  Google Scholar 

  • Munoz E, Kirtman B, Weijer W (2011) Varied representation of the atlantic meridional overturning across multidecadal ocean reanalysis. Deep Sea Res II 58:1848–1857

    Article  Google Scholar 

  • Neale R, Richter J, Park S, Lauritzen R, Vavrus S, Rasch P, Zhang M (2013) The mean climate of the Community Atmosphere Model (CAM4) in forced SST and fully coupled experiments. J Clim 26:5150–5168

    Article  Google Scholar 

  • Newman M (2013) An emperical benchmark for decadal forecasts of global surface temperature anomalies. J Clim 26:5260–5269

    Article  Google Scholar 

  • Penland C, Sardeshmukh P (1995) The optimal growth of tropical sea surface temperature anomalies. J Clim 8:1999–2024

    Article  Google Scholar 

  • Pohlmann H, Jungclaus J, Köho A, Stammer D, Marotzke J (2009) Initializing decadal climate predictions with the GECCO oceanic synthesis: effects on the North Atlantic. J Clim 22:3926–3938

    Article  Google Scholar 

  • Pulwarty R (2003) Climate and water in the West: science, information, and decision making. Water Resour 124:4–12

    Google Scholar 

  • Rayner N, Parker D, Horton E, Folland C, Alexander L, Rowell D, Kent E, Kaplan A (2003) Globally complete analyses of sea surface temperature, sea ice and night marine air temperature, 1871–2000. J Geophys Res 108(D14):4407. doi:10.1029/2002JD002670

    Article  Google Scholar 

  • Robson J, Sutton R, Lohmann K, Smith D (2012) Causes of the rapid warming of the North Atlantic Ocean in the mid-1990’s. J Clim 25:4116–4134

    Article  Google Scholar 

  • Ross S (2007) Introduction to probability theory. Elsevier, Amsterdam

    Google Scholar 

  • Seager R, Kushnir Y, Herweijer C, Naik N, Velez J (2005) Modeling of tropical forcing of persistent droughts and pluvials over western North America:1856–2000. J Clim 18:4065–4088

    Article  Google Scholar 

  • Smith D, Cusack S, Colman A, Holland C, Harris G, Murphy J (2007) Improved surface temperature prediction for the coming decade from a global climate model. Science 317:796–799

    Article  Google Scholar 

  • Smith D, Eade R, Dunstone N, Fereday D, Murphy J, Pohlmann H, Scaife A (2010) Skilful multi-year predictions of Atlantic hurricane frequency. Nature Geosci 3:846–849

    Article  Google Scholar 

  • Smith T, Reynolds R, Peterson T, Lawrimore J (2008) Improvements to NOAA’s historical merged land-ocean surface temperature analysis (1880–2006). J Clim 21:1601–1612

    Google Scholar 

  • Stan C, Kirtman B (2008) The influence of atmospheric noise and uncertainty in ocean initial conditions on the limit of predictability. J Clim 21:3487–3503

    Article  Google Scholar 

  • Sutton R, Hodson D (2005) Atlantic Ocean forcing of North American and European summer climate. Science 309:115–118

    Article  Google Scholar 

  • Taylor K, Stouffer R, Meehl G (2012) An overview of the CMIP5 and the experiment design. Bull Am Meteorol Soc 93. doi:10.1175/BAMS-D-11-00,094.1

  • Ting M, Kushnir Y, Seager R, Li C (2009) Forced and internal twentieth-century sst trends in the north Atlantic. J Clim 22:1469–1481

    Article  Google Scholar 

  • Troccoli A, Palmer T (2007) Ensemble decadal predictions from analyzed initial conditions. Phil Trans R Soc 365A:2179–2191

    Article  Google Scholar 

  • van Oldenborgh G, Doblas-Reyes F, Wouters B, Hazeleger W (2012) Skill in the trend and internal variability in a multi-model decadal prediction ensemble. Clim Dyn 38(7):1263–1280

    Article  Google Scholar 

  • Yasunaka S, Ishii M, Kimoto M, Mochizuki T, Shiogama H (2011) Influence of XBT temperature bias on decadal climate prediction with a coupled climate model. J Clim 24:5303–5308

    Article  Google Scholar 

  • Yeager S, Karspeck A, Danabasoglu G, Tribbia J, Teng H (2012) A decadal prediction case study: late 20th century North Atlantic Ocean heat content. J Clim 25:5173–5189

    Article  Google Scholar 

  • Zhang R, Delworth T (2006) Impact of Atlantic multidecadal oscillations on India/Sahel rainfall and Atlantic hurricanes. Geophys Res Lett 33(L17):712

    Google Scholar 

Download references

Acknowledgments

We acknowledge the hard work and dedication of all the scientists and software engineers who contributed to the development of the CCSM4. Special thanks are extended to Jeffrey Anderson, Tim Hoar, Nancy Collins, and Kevin Raeder of the National Center for Atmospheric Research (NCAR) Data Assimilation Research Section, who developed DART and provided ongoing support for the ocean assimilation system. Thanks also to Marianna Vertenstein for her guidance in developing the CCSM-DART interface and to Joe Tribbia, Peter Gent, Jerry Meehl and Grant Branstator for useful discussions. We are grateful to Tony Rosati for providing generous support during the initial phase of decadal prediction efforts at NCAR. This work was funded in part by the NOAA Climate Program Office under the Climate Variability and Predictability Program grants NA09OAR4310163 and NA13OAR4310138, and by the NSF Collaborative Research EaSM2 grant OCE-1243015. NCAR is sponsored by the National Science Foundation (NSF) and the CCSM project is supported by NSF and the Office of Science Biological and Environmental Research of the U.S. Department of Energy (DOE). Computing resources were provided by the Climate Simulation Laboratory at NCAR’s Computational and Informational Systems Laboratory, sponsored by the NSF and other agencies, and by the Oak Ridge Leadership Computing Facility at the Oak Ridge National Laboratory, which is supported by the Office of Science of the U.S. DOE under contract DE-AC05-00OR22725.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to A. Karspeck.

Appendices

Appendix 1: Details of the two initialization methods

The ocean–sea-ice states used to initialize the HDInit coupled model forecasts came from an integration of the nominal 1\(^\circ\) resolution POP2 ocean model forced with the Coordinated Ocean-Ice Reference Experiment sinterannually varying atmospheric data set (CORE-II; Large and Yeager 2009; Danabasoglu 2014) from 1948 to 2007. Because this is a single-realization ocean and sea-ice state estimate, ensembles for the HDInit experiment were generated by initializing the atmosphere from time-staggered atmospheric states from the NoInit twentieth century integration. This set of prediction experiments is described in detail in Yeager et al. (2012).

For the DAInit experiment, the ocean was initialized using an ensemble of historical ocean states obtained from the assimilation of in-situ temperature and salinity observations into the POP2 ocean model. Details of the data assimilation method, data sources, and atmospheric forcing used to generate the ocean state estimate can be found in Karspeck (2013). The ensemble assimilation system was initialized on Jan 1 1998, and hydrographic observations were assimilated daily using an ensemble adjustment Kalman filter (Anderson 2001). The Data Assimilation Research Testbed (DART; Anderson et al. 2009) software was used to implement the ensemble filter. The assimilation was run continuously until Jan 1 2006, supplying initial states for Jan 1 of years 2000, 2001, 2002, 2003, 2004, 2005, and 2006. For each of the start dates Jan 1 1975, 1980, 1985, 1990, and 1995 the filter was reinitialized one year preceding each of the forecast start dates, such that one year of in situ observations were assimilated prior to the ocean state estimate being used for forecast initialization. For the DAInit experiment a single atmospheric state taken from a CAM4 solution forced with observed SST and sea-ice was used. The sea-ice state for the DAInit predictions was from a CORE-II experiment where strong restoring to climatological salinity was employed. Sea-ice states for the two experiments were chosen for best compatibility with their respective ocean states. Similarly, while the initial atmospheric state is thought to be unimportant for prediction on this time-scale, all atmospheric initial states were taken at approximately coordinated start-dates for best agreement with the ocean state.

Note that all of the uncertainty in the initial-conditions in the DAInit experiment comes from the ensemble of ocean states, while the HDInit experiment relies on the initial uncertainty in the atmosphere to generate forecast spread.

Appendix 2: Drift correction method

We adopt the cross-validated method of drift correction recommended for CMIP5 experiments outlined in CLIVAR (2011). At each point on the ocean grid, the model drift \(d\), corresponding to the \(n\) start years \(j\) at forecast lead year \(t\) is computed as

$$\begin{aligned} d_{j,t} = \frac{1}{n-1}\sum ^n_{i=1} (\langle X_{i,t}\rangle - O_{i,t}),\;\;i \ne j \end{aligned}$$

where \(\langle X_{i,t} \rangle\) is the annual and ensemble mean SST forecasts initialized on Jan 1 of year \(i\) and \(O_{i,t}\) is the HADISST from the corresponding year. Drift-corrected forecasts are then computed as the difference between the raw forecast members and the model drift, i.e. \(X_{j,t}-d_{j,t}\).

As formulated above, the forecast at any particular start date does not contribute to the drift estimate that is subsequently applied. While cross-validation is intended to reduce the likelihood of overestimating skill, Gangsto et al. (2013) demonstrate that when the sample size is low cross-validation can sometimes result in a negatively biased skill score.

Appendix 3: A bootstrapped single-ensemble member correlation coefficient and an approximate formula for its expected value

In this appendix we describe a bootstrapping procedure for building a distribution of single-member correlation coefficients that accounts for the spread in the ensemble forecasts. This distribution is used as the skill-score in the body of the paper. We also give an approximate formula for the expected value of the distribution. Note that the specifics of this bootstrap were developed for this assessment. The reader is referred to Efron and Tibshirani (1993) for a general description of bootstrapping and its strengths and limitations.

Bootstrapped samples of correlation coefficients associated with an initialized forecast experiment are generated in the following way: For each of the \(n\) start dates, one of the 10 forecast ensemble members is chosen at random. The SST averaged over a box-region (or simply taken at a grid-point) is then computed for the 2–4 or 6–9 years lead times. Using this set of \(n\) single-member forecasts, a correlation coefficient is computed with (1). These ensemble members are then returned to the pool of ensembles associated with their start date, so that they may be randomly chosen again. This process is repeated 2,000 times, providing samples that can be used to build a bootstrapped distribution of \(r\) associated with the initialized prediction experiment.

Here we derive an approximate formula for the expected value of the distribution of the single-member anomaly correlation coefficient, \(r\). We can treat \(X_{j,\tau }\) from (1) as a Gaussian random variable, with mean assumed to be the forecast ensemble mean, and variance assumed to be the forecast ensemble variance (denoted \(\sigma ^2_{j,\tau }\)). This implies that \(r\), which is a function of \(X_{j,\tau }\), is also a random variable. Of course, the sampling distribution of \(r\) is not normally distributed, but we can use standard definitions of the statistical expectance of functions of normally distributed variables (Ross 2007) to derive the expected value of the distribution of \(r\).

We make the simplifying assumption that \(s^2_X\) in (1), which is a random variable by virtue of its functional dependence on \(X_{j,\tau }\), can be replaced by a constant \(\tilde{s^2_X}\) that is defined as the expected value, i.e.,

$$\begin{aligned} \tilde{s^2_X} \equiv E(s^2_X) = s^2_{\langle X \rangle } + \frac{1}{n}\sum _{j=1}^n \sigma ^2_{i,\tau } = s^2_{\langle X \rangle }+ \overline{\sigma ^2_{\tau }}. \end{aligned}$$

Here \(\langle \cdot \rangle\) denotes an ensemble average and the overbar is an average over start dates. Using \(\tilde{s^2_X}\), we can approximate the expected value of \(r\) as:

$$\begin{aligned} E(r) \approx \frac{1}{n-1}\sum _{j=1}^n \frac{(\; \langle X_{j,\tau }\rangle -\overline{\langle X_{\tau }\rangle }\;) \;\; (O_{j,\tau }-\overline{O_{\tau }})}{s_O \sqrt{s^2_{\langle X \rangle }+ \overline{\sigma ^2_{\tau }}}}. \end{aligned}$$
(2)

We have expressed the expected value of \(r\) as a function of the forecast ensemble statistics to highlight the relationship between this single-member based correlation metric and the correlation score as it is commonly computed using the ensemble mean. Due to the forecast ensemble spread, the expected value of \(r\) will be smaller than what would be computed if \(X_{j,\tau }\) were an ensemble mean forecast at each start date. In the limit that the average ensemble spread (\(\overline{\sigma ^2_{\tau }}\)) is much smaller than the time-variance of the ensemble mean (\(s^2_{\langle X \rangle }\)), the expected value of \(r\) will be the correlation coefficient as it is commonly computed using the ensemble mean. The approximate relationship in (2), can be in error due to the assumption that \(s^2_X\) can be replaced by \(\tilde{s^2_X}\), and also because bootstrapping with replacement yields values of \(r\) that are not independent samples. In a separate test of the utility of (2), we found that for our experimental setup (2) underestimates the bootstrapped mean value of \(r\) by about 3 %. It is thus a reasonable approximate formula.

Appendix 4: Reference statistical forecasts

To create a benchmark for what are meaningful levels of skill, we generate two reference forecasts. Both reference forecasts used in this study are based on representing global SST variability as a first order auto-regressive (AR1) process that includes both spatial and temporal auto-correlations in the statistical model. Modeling the auto-correlation in both time and space is important because we are evaluating SST in regionally and temporally averaged boxes and because there is long-term autocorrelation in the observations and the forecasts.

The multivariate spatial patterns are computed as the leading 15 Empirical Orthogonal Functions (EOF) of the annually averaged HADISST anomaly relative to the 1870–1960 time mean. The leading 15 modes can be used to account for about 88 % of the annual, coarse-resolution global SST variance from 1870 to 1960. We only use the data up until 1960 to avoid overlap with our forecast evaluation period. The multivariate AR1 process for the vector \(v\) of time-varying Principal Components (PC) associated with the spatial EOF patterns can be expressed:

$$\begin{aligned} v_{t+1} = Gv_t + \epsilon , \;\;\;\; \epsilon \sim \mathcal {N}(0,\Sigma ^2). \end{aligned}$$
(3)

The vector random variable \(\epsilon\) is normally distributed with mean zero and variance \(\Sigma ^2\). It is time-uncorrelated and uncorrelated across modes (i.e. \(\Sigma ^2\) is a diagonal matrix). The least squares estimate of \(G\) can be computed as \(G = C_1C_0^{-1}\), where the one year lag covariance is \(C_1 = \langle v_{t+1} v_t^T\rangle\) and the lag-zero covariance is similarly \(C_0 = \langle v_{t} v_t^T \rangle\). Here \(\langle \cdot \rangle\) refers to an average over all years \(t\) from 1870 to 1960. The error variance associated with each EOF/PC mode (the elements of \(\Sigma ^2\)), can be computed directly from the residual (i.e., \(v_{t+1}-Gv_t\)). Statistical realizations of SST variability that share the spatio-temporal statistical properties of the observed SST anomaly can be generated by integrating (3) forward in time to form PC time series that are then multiplied by the associated EOF spatial patterns.

Samples of correlation coefficients associated with the null hypothesis of no correlation are generated in the following way: two independent 50-year realizations of (3) are generated with randomly generated initial conditions at \(t_0\). The SST averaged over a box-region (or at a grid-point) is then computed. Four year averages are computed at intervals set by the prediction experiment. For consistency with these experiments, this are nine 4-year averages at five year intervals (associated with the 1961–2000 start dates) followed by six 4-year averages at one year interval (associated with the 2001–2006 start-dates).Footnote 7 A sample correlation coefficient is then computed with these averaged and subsampled time-series. This process is repeated 2,000 times, providing samples that can be used to build an empirical distribution of \(r\) associated with ‘no-skill.’

The AR1 process described above can also be used to make initialized statistical predictions. When used in this way, an AR1 model is similar to the linear inverse model (or LIM) benchmark forecasts outlined by Newman (2013). To make LIM forecasts, the HADISST from the start years of the HDInit and DAInit experiments are projected onto the leading 15 EOFs, and a single ensemble member is forecasted from the resulting PC using (3). These ensemble members have identical initial conditions and are different only because of the random variations in \(\epsilon\). Just as in the HDInit and DAInit experiments, correlation coefficients are computed between the statistical forecast and the corresponding observations of HADISST. This process is repeated 2,000 times. Sample correlation coefficients from these LIM predictions are used just as for the HDInit and DAInit coupled model predictions to build an empirical distribution of \(r\) associated with LIM.

Appendix 5: Computing the probability that one set of forecasts outperforms another

The probability that one forecast outperforms another, which are tabulated in Tables 1 and 2, are computed in the following way. A normal kernel density function (we use the Matlab ‘ksdensity’ function; Bowman and Azzalini 1997) is used to generate smooth (but discrete) probability density functions from the bootstrapped correlation samples. We then perform a set of discrete integrals to compute the probability \(p\) that correlation coefficients from distribution \(f\) (e.g., from the HDInit and DAInit experiments) will exceed correlation coefficients from distribution \(g\) (e.g, from the ‘no-skill’ or LIM statistical forecasts). This can be written:

$$\begin{aligned} p=\sum _{r=-1}^{1}g(r)\sum _{r'=r}^{1} f(r') \Delta r' \Delta r\;\;\;\;\; (r,r') \in [-1:\Delta r:1] \end{aligned}$$

where we have made use of the fact that correlation coefficients can only range from \(-1\) to \(1, \Delta r\) and \(\Delta r'\) are (identical) bin-widths that define the discrete values of \(r\) and \(r'\), and \(f\) and \(g\) are expressed as discrete probability density functions.

In the case of the NoInit experiments, where there are only five samples, a distribution is not computed. Instead, we assume that each ensemble member is equally likely to occur with a probability of 20 %. The probably of outperforming the NoInit experiment can then be written:

$$\begin{aligned} p=\sum _{r} 0.20\sum _{r'=r}^{1} f(r') \Delta r',\;\; r \in [r_1, r_2 ... r_5], \;\;r' \in [-1:\Delta r:1] \end{aligned}$$

where \(r_1\) through \(r_5\) are the correlation coefficients computed for each of the NoInit ensemble members.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Karspeck, A., Yeager, S., Danabasoglu, G. et al. An evaluation of experimental decadal predictions using CCSM4. Clim Dyn 44, 907–923 (2015). https://doi.org/10.1007/s00382-014-2212-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00382-014-2212-7

Keywords

Navigation