Further Evidence on the Usefulness of Real-Time Datasets for Economic Forecasting

In this paper, we assess the relevance of real-time datasets for forecasting. We construct a variety of real-time prediction models and evaluate their performance in a series of ex-ante prediction experiments that are designed to mimic forecasting approaches used when constructing forecasts in real-time for output, prices and money. We assess the models within univariate and multivariate frameworks by including revision errors as regressors, allowing us to examine the marginal predictive content of the revision process. In another multivariate application for output we add money, thus examining the real-time predictive content of money for income. The most important result we obtain is that the choice of which release of data to predict seems not to have an impact on which releases of data should be used in estimation and prediction construction but that differences in how to utilize realtime datasets do arise when the variable being modelled and predicted changes. Overall our findings point to the importance of making real-time datasets available to forecasters, as the revision process has marginal predictive content, and because predictive accuracy increases when multiple releases of data are used when specifying and estimating prediction models. This underscores the importance of collecting and maintaining such real-time datasets.


Introduction
Recent empirical research has presented strong evidence in favor of the usefulness of making realtime datasets available to economists.Aruoba (2008) finds that for most U.S. macroeconomic time series, revision errors have a positive bias and are highly predictable.These finding are based on the use of information available at the time of the first release.In Aruoba, Diebold, and Scotti (2008), the findings of Aruoba (2008) are used as one of the main motivating factors for the construction of a real-time business conditions measurement index.In summary, both of these papers suggest that there is much to be gained by using multiple vintages of data in the construction of predictions and prediction models.For example, one might fruitfully choose to estimate prediction models that employ all vintages (releases) of available variables, say using the Kalman filter or some other filtering procedure (see, e.g., Mariano and Tanizaki (1995)).However, the literature has said little about which release of data to predict, and whether it is preferable to use mixed releases of data when forming predictions and prediction models (as is done when "latest available release" data are used).In this paper, we attempt to shed new light on the above issues.Key precedents to our research on the informational content of realtime datasets include: Diebold and Rudebusch (1991), Hamilton and Perez-Quiros (1996), Bernanke and Boivin (2003), and the papers cited therein.
Our approach is to construct a variety of different real-time prediction models and to evaluate their performance in a series of ex-ante prediction experiments that are designed to mimic forecasting approaches used when constructing forecasts in real-time, for the purpose of policy setting and generic real-time decision making.For this purpose we use real-time datasets on output, prices, and money.Our prediction models include, among others, one that uses only first release data and one that utilizes only the latest available data (i.e., uses a mixture of most recent first release data and more distant later release data).Some of our models include revision errors as regressors, hence allowing us to examine the marginal predictive content of the revision process.In addition, our experiments are designed to find out whether it matters which release of data one chooses to forecast, which release(s) of data should be used when estimating prediction models, and how definitional change in real-time variables affects our findings regarding which release of data to predict.In another implementation of our experimental setup within a multivariate framework, we carry out an empirical analysis in which we examine the real-time predictive content of money for income, building on the work of Stock and Watson (1989), Amato and Swanson (2001), and others.Finally, our experiments are used to form simple rationality tests that are based solely on the examination of ex-ante predictions, rather than being based on in-sample regression analysis, as are many tests in the extant literature (see Corradi, Fernandez, and Swanson (2009) for further discussion).
The results of our prediction experiments point clearly to the need for making real-time datasets available to empirical researchers.In almost all cases that we consider, multiple releases of a variable are useful for constructing MSFE-best predictions.More important, we present evidence concerning whether or not one should aim to predict the "first release" or "final" release of a variable, and which data are most useful for model estimation and prediction construction.We find that regardless of which release of prices one specifies as the "target" variable to be predicted, using only "first release" data in model estimation and prediction construction yields MSFE-best predictions.On the other hand, models estimated and implemented using "latest available release" data are MSFE-best for predicting all releases of money.Thus, perhaps surprisingly, in our empirical analysis we find that the choice of which release of data to predict seems not to have an impact on which releases of data should be used in estimation and prediction construction.However, differences in how to utilize real-time datasets do arise when the variable being modelled and predicted changes.As for our multivariate forecasting experiment involving the real-time predictive content of money for income, while we find little marginal predictive content in money, we note that vector autoregressions with money do not perform significantly worse than autoregressions, when predicting output in the past 20 years.This is somewhat surprising because models with irrelevant variables should be less efficiently estimated, leading in many cases to worsened predictive performance.Finally, we also find new evidence that early releases of money are rational, whereas prices and output are irrational.
The rest of the paper is organized as follows.In Section 2, we outline our notation, discuss the empirical methodology used by presenting the variety of real-time prediction models that we analyze, and we discuss the real-time datasets used.Section 3 contains our empirical findings.Concluding remarks are given in Section 4. Tables and figures are collected at the end of the paper.

Setup
Let t+k X t denote a variable (reported as an annualized growth rate) for which real-time data are available, where the subscript t denotes the time period to which the datum pertains, and the subscript t + k denotes the time period during which the datum becomes available.In this setup, if we assume a one-month reporting lag, then first release or "preliminary" data are denoted by t+1 X t .In addition, we denote fully revised or "final" data, which is obtained as k → ∞, by f X t .Data are grouped into releases and vintages.The first release is preliminary data, the second release is 2 nd available data, and so on.In regard to vintages, the 2000:1 vintage is the time series of latest release data available in 2000:1, and the 2000:2 vintage is the time series of latest release data available in 2000:2, and so on.Regression models parameterized using the latest available release of data at each point in time use the most recently available vintage.Such models correspond to those usually used in practice.Regression models parameterized using only a single release of data use time series constructed by taking a single observation from each vintage of data.To further set notation, let t+2 u t+1 t = t+2 X t − t+1 X t , thus, t+2 u t+1 t and t+1 u t t−1 denote the errors between the second and the first releases at time t + 2 and at time t + 1, respectively.

Prediction
In this subsection, we discuss the prediction models that will be used for addressing the questions outlined above.In particular, we consider the issue of prediction using various variable/vintage combinations as defined in the following set of models.
In the above models, the time subscripts on the model coefficients are meant to indicate that the parameters are estimated using particular calendar and release dated observations from our real-time datasets and correspond to the final calendar date and release combination in the dataset used to estimate the models.Our analysis is carried out by recursively estimating the above models and constructing sequences of ex-ante 1-step ahead predictions, for various values of k.Notice that when k = 2, we are assuming that the "target variable" to be forecast is the first release.
Model A has explanatory variables that are formed using only first available data.Thus, the first model corresponds to the approach of simply using first available data and ignoring all later releases, regardless of which release of data is being forecasted.This model should be expected to perform well if data revisions are "news" and/or if type of definitional changes do not contaminate the data.

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25 Model B is specified using explanatory variables that are available k − 1 months ahead, corresponding to (k − 1) st available data.Thus, this model uses data that have been revised k − 1 times in order to predict data that likewise have been revised k − 1 times.In this sense, Model B is included only as a "reality check", as the model uses stale information in all instances other than the case in which k = 2 (in which case Models A and B are equivalent).However, when k > 2, the calendar date of information used to predict one-step ahead is at least two periods prior to the prediction period.Therefore, the "cost" of using Model B is the inclusion of "stale" data, and hence the model should be expected to perform poorly.
In Model C, the latest release of each observation is used in prediction, so that the dataset is fully updated prior to each new prediction being made.We refer to this model as our "latest available data" model because policy makers and others who construct new predictions each period, after updating their datasets and re-estimating their models, generally use this type of model.If useful information accrues via the revision process, then one might expect that using latest available data (Model C) would yield a better predictor of t+k X t than when only "stale" first release data are used (Model A), for example.Of course, the last statement has a caveat.Namely, it is possible that first release data are best predicted using only first release regressors, second release using second release regressors, etc.This might arise if the use of real-time data as in Model C results in an "informational mix-up" due to the fact that every observation used to estimate the model is a different release, and only one of these releases corresponds to the release being predicted, at any point in time.For further discussion of real-time forecasting using models such as Model C, the reader is referred to Swanson and van Dijk (2006) and Faust and Wright (2009).
Given the above considerations, one natural approach is to compare the following prediction equations using least squares estimators.For example, the cases with only the intercept and first order slope parameter in each model are: In our prediction experiments, we set: (i) p = 1; (ii) p = S IC; (iii) p = AIC; (iv) p = 0 (random walk with drift model).* Additionally, we set k ={2, 3, 4, 6, 12, 24}.
Notice that the estimators used in the three prediction models are indeed quite different.Moreover, analogous least squares estimators for all other parameters in the prediction equations follow immediately (by simply setting k = 2 in the original models).Given the above formulation, it is also clear that the straw-man random walk with drift prediction model that we also consider in our experiments has intercept parameter that differs across the three models.Importantly, note that, in order to carry out true real-time prediction, we must assume that we have observations for calendar dates only up until period t, using only vintages t + 1 and earlier.
All experiments are based on the examination of the MSFEs associated with 1-step ahead predictions constructed using recursively estimated models, where R observations are used in our first estimation, R + 1 observations are used in our second estimation, etc.We thus construct sequences of P − k ex-ante predictions and prediction errors, where T = R + P is the sample size.We set R to be 1969:4, so that our first prediction is calendar date 1970:1.The start calendar date of our dataset is 1959:4, and we have vintages of data from 1965:4.Another set of predictions are constructed using recursively estimated models with ex-ante prediction periods beginning in 1983:1 or 1990:1.
In the multivariate version of the forecasting experiment that we undertake we include: (i) money, income, prices, and interest rates; and (ii) income, prices, and interest rates.In these models, it is assumed that the target variable of interest is output growth.Thus, we are examining, in real-time, the marginal predictive content of money for output, using various data vintages, various revision errors, and for a target variable that corresponds to various releases of output growth.Other recent papers examining the usefulness of real-time data for prediction include Robertson and Tallman (1998), Gallo and Marcellino (1999), and the papers cited therein.
MSFEs are examined via the use of Diebold and Mariano (DM: 1995) and Clark and McCracken (2001) predictive accuracy test (see also Clark and McCracken (2005), Clark and McCracken (2009)).† The test has a null hypothesis of equal predictive accuracy and is defined as follows: ) is a random variable defined to be the difference between the prediction errors of two models that are being compared, when transformed according to a given loss function, l, d = 1 , and the denominator is a heteroskedasticity and autocorrelation consistent covariance estimator, such as the Newey-West estimator.The limiting distribution of the DM statistic is given in Theorems 3.1 and 3.2 in Clark and McCracken (2005) under quadratic loss, so that 2,t,k , and is N(0, 1) in cases in which the prediction models are nonnested and parameter estimation error vanishes (or the in-sample and out-of-sample loss functions are the same -see also Corradi and Swanson (2006)).In the sequel, we consider only quadratic loss and hence report mean square forecast errors (MSFEs) as well as DM test statistics based on quadratic loss.
In addition to examining MSFE performance using DM statistics, we also formally examine the rationality of early releases.Following the literature ‡ , we test for the rationality of t+k X t , by finding † Clark and McCracken (2009) reconsider tests for comparing nonnested as well as nested forecasting models, when forecasts are produced using real-time data.They show that, under the news hypothesis, data revisions do not affect the limiting distributions of tests for predictive evaluation.On the other hand, the use of real-time data plays a crucial role whenever revisions are noisy and effects are different, depending on whether we are comparing nonnested or nested models.
‡ See Mankiw, Runkle, and Shapiro (1984), Mankiw and Shapiro (1986), Kavajecz and Collins (1995), Mork (1987), Keane and out whether the conditioning information in a vector t+1 W t , available to the data issuing agency at the time of first release, has been efficiently used.Our implementation involves including t+1 W t in the regression of each of the three models suggested above with various different revision errors.Formally, we consider the following alternatives: , where the notation used in these regressors is defined at the end of the previous subsection.Thus, we directly use the regression models to construct sequences of ex-ante predictions.Then, the accuracy of these predictions is assessed using the DM test.This is therefore a truly outof-sample rationality test and is in the spirit of those suggested in Ashley, Granger, and Schmalensee (1980), Chao, Corradi, and Swanson (2001), and Corradi and Swanson (2002).

Data
Our real-time dataset includes real GDP (seasonally adjusted), the GDP chain-weighted price index (seasonally adjusted), the money stock (measured as M1, seasonally adjusted) and the interest rate (measured as the rate on the three-month Treasury bill).All series have a quarterly frequency and our real-time dataset for each of the four variables was obtained from the Federal Reserve Bank of Philadelphia's real-time dataset for Macroeconomists (RTDSM).The RTDSM can be accessed online at http://www.phil.frb.org/econ/forecast/readow.html.The series were obtained from the "by-variable" files of the "core variables/quarterly observations/quarterly vintages" dataset, and are discussed in detail in Croushore andStark (2001, 2003) and Croushore (2006).Note also that interest rates are not revised, and hence our interest rate dataset is a vector rather than a matrix (see Swanson, Ghysels and Callan (1999) and Ghysels, Swanson, and Callan (2002) for a detailed discussion of the calendar date/vintage structure of real-time datasets).
The first vintage in our sample is 1965:4, for which the first calendar observation is 1959:3.This means that the first observation in our dataset is the observation that was available to researchers in the fourth quarter of 1965, corresponding to calendar-dated data for the third quarter of 1953.The datasets range up to the 2006:4 vintage and the corresponding 2006:3 calendar date, allowing us to keep track of the exact data that were available at each vintage for every possible calendar-dated observation up to one quarter before the vintage date.This makes it possible to trace the entire series of revisions for each observation across vintages.
Various summary information about the datasets is depicted in the first six plots in Figures 1-3.We use log-differences throughout our analysis (except for interest rates); and various releases of the logdifferences of all variables, except the interest rate, are depicted in the plots.Also included are plots of the first and second revision errors measured as the difference between the first vintage (e.g.first available) of a calendar observation and the second and third vintages, respectively, and cumulative revision errors for various releases.As can readily be seen on inspection of the distributions of the revision errors as well as via examination of the summary statistics reported in Table 1, the first revision (i.e., the difference between the first and second vintages) is fairly close to normally distributed.On the other hand, the distribution of the second revision errors is mostly concentrated near zero, implying that much of the revision process has already taken place in the first revision.Indeed, the distributional shape of revision errors beyond the first revision is very much the same as that reported for the second revision in these plots, with the exception of revision errors associated with definitional and/or other Runkle (1990) , and Rathjens and Robins (1995) for further details.A summary of this sort of test is given in Swanson and van Dijk (2006).

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25 structural issues associated with the variables.This is one of the reasons why much of our analysis focuses only on the impact of first and second revision errors -later revision errors appear to offer little useful information other than signalling the presence of definitional and related structural issues associated with the variables.This feature of the data is illustrated in plots with title "Calendar data across vintages" in Figures 1-3, where we have plotted early calendar dates (e.g., 1959:1; 1962:1; and 1965:3) across all available vintages in our sample.Data for a particular calendar date sometimes vary significantly across vintages.For instance, looking at the 1959:Q4 calendar observation for output across all vintages, one can observe several discrete movements driving the value of that particular observation from a monthly growth of 1% for the earlier vintages to 0.5% for the later vintages.It seems reasonable to argue that most (if not all) of the discrete variations in that particular calendar observation are not due to "pure revisions", but are solely a consequence of definitional breaks in the measurement of output.Similar breaks are observed for prices and money in Figures 2 and 3. Additionally, y is the mean of the series, σ y is the standard error of the series, σ y is the standard error of y, skew is skewness, kurt is kurtosis, LB is the Ljung-Box statistic, JB is the Jarques-Bera statistic, and ADF is the augmented Dickey-Fuller statistic, where lag augmentations are selected via use of the Schwarz information criterion.See Section 2 for further details.

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25 ( * ) Notes: The upper six panels describe the main properties of the real-time dataset for the growth rate of output (X) as follows: (i) "Releases" plots the time series for t+k X t , k = 1, 2, 6, 24; "Cumulative Revision Errors" plots the time series for t+k X t − t+1 X t , k = 1, 2, 6, 24; (iii) "Calendar data across vintages" plots the time series for three calendar dates across all available vintages, where the calendar dates are 1959:4+k X 1959:4 , 1962:1+k X 1962:1 , 1965:3+k X 1965:3 , for all k; (iv) "Growth Rates, First Releases" plots the distribution of t+1 X t across the entire sample; and (v) "First" and "Second Revision Error" plots the distribution of t+2 X t − t+1 X t and t+3 X t − t+2 X t , respectively.The lower six panels describe the main results from the recursive estimation of Models A and C as follows: (i) "Intercept", "First order Slope", and "Intercept in Random Walk Model" plot the recursive estimates of, respectively, a M t+1,t , β M 1,t+1,t , a RW M t+1,t , for M = A, C, where a RW M t+1,t is the slope of the random walk model associated to model M (see notes to Table 2); (iii) "Lags Selected using SIC" reports the number of lags selected using the SIC across all recursively estimated models; and (iv) "Forecast Errors, k=2" and "k=6" report, respectively, t+k X t − t+k X M t k = 2, 6; M = A, C.

Prediction Experiments
As discussed in Section 3, we carried out three types of autoregressive prediction experiments, in which the objective was to forecast output, prices, and money.The methods involved fitting regression Models A, B, and C. Recall that Model B is our "strawman" model, and should be expected to perform increasingly poorly as k increases.Moreover, Model A involves constructing predictions using only

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25 first release data and hence might be expected to perform poorly for predicting k th releases, when k is large, assuming that either our data are irrational, or definitional changes result in "contamination" of the earliest first release calendar-dated observations used in the construction of our prediction models.
On the other hand, Model C uses a mixture of releases both in parameter estimation and in prediction construction.Thus, even if there are efficiency problems associated with using first release data in Model A, these may be outweighed by the cost of using mixed releases of data in Model C, and hence Model A might still be MSFE-best.Tables 2-4 report our results on the predictive accuracy of Models A-C using simple autoregressive models.Entries in the tables are MSFEs and DM test statistics that are calculated with Model A as the benchmark.In Tables 2-4, entries in bold denote lowest MSFEs for a particular value of k, across all three models.Each Table is subdivided into different panels according to the date of the first prediction (either 1970:1, 1983:1, or 1990:1).All predictions are constructed using recursively estimated models.In all cases, Model A is set as the "benchmark" model, so that a negative statistic means that Model A is "MSFE-better" than the particular model against which it is being compared.All estimated models are either pure autoregressions or autoregressions with revision error(s) included as additional explanatory variables.Lags are selected using the Schwarz Information Criterion.The pure autoregression models are: Model A (First Available Data), Model B (k th Available Data) , and Model C (Latest Available Data) .In the models, X denotes the growth rate of either output, prices, or money.Also, RWD is the random walk with drift model in log levels, u C1 = t+1 u t t−k , k = 1; u C2 = t+1 u t t−k , k = 1, 2; and u C3 = t+1 u t+2−k t+1−k , k = 3.Further details are contained in Section 2. A number of clear-cut conclusions emerge when the results reported in these tables are examined.In Table 2, the MSFE-best output predictions result when using Model A for low values of k, and using Model C for high values of k.However, for price predictions (see Table 3), Model A is always MSFE-best, while for money, Model C is always MSFE-best, regardless of release being predicted and data subsample (see Table 4).As expected, Model B performs poorly and is particularly ineffective for larger values of k for all three variables.Additionally, the real-time random walk with drift models that we estimate never yields predictions as accurate as those based on our autoregressive type models.
We have clear evidence that whether or not one should use the latest or earlier releases of data in model parameter estimation and prediction construction is not just dependent on the release of data to be predicted, but is also dependent on what the target variable of interest is.When the target variable is money, Model C is preferred (i.e., use the latest available data).However, when the target variable

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25 is prices, Model A is preferred.Most importantly, precisely which model is preferred is independent of the release, k, being predicted, for both prices and money.This suggests that the "target release" to be predicted is actually not very important for these variables, which is quite surprising as one might expect the cumulative effect that the combination of inefficiency, measurement error, and definitional change has on model choice will result in different models being chosen when the target release to be predicted increases from preliminary to final data, as is the case with output.Finally, note that MSFEs associated with the "best" models for money (i.e., Model C) largely decrease as k increases.This is consistent with the view that using "latest available" data that have been revised as much as possible, when forming model coefficient estimates, leads to estimates that are more accurate when the objective is the prediction of later release or even "final" data.This in turn suggests that later releases of data should be predicted more accurately using Model C (as is indeed the case for money, where Model C is actually the MSFE-best model), but not necessarily when using other models.Indeed, notice that for prices, where Model A wins, the MSFEs actually increase as one increases k from 2 to 3 to 4, before beginning to decrease.The same sort of mixed pattern of increasing and decreasing MSFEs characterizes output.
Tables 5-7 repeat the above experiments adding revision errors into the mix, hence allowing us to assess rationality from a different perspective.If revision errors are useful in ex-ante predictions of the sort that we report on in our tables, then we have direct evidence of inefficiency.To aid in the presentation of our results, in Tables 5-7 bold numbers highlight experiments in which the inclusion of revision errors lowers the MSFE compared their counterparts in Tables 2-4.Starred entries in the tables denote cases in which a different model than the one found in Tables 2-4 obtains a lower MSFE.Consider the case of prices first (Table 6).Point MSFEs reported for Model A in this case are often lower than the comparable MSFEs reported in Table 3.This interesting result suggests that early price releases are actually irrational and that the reason Model A "wins" in Table 3 is that the use of mixed release data associated with the estimation and implementation of Model C is simply "too costly" relative to the predictive accuracy losses associated with using mildly irrational first release data.This finding is consistent with the finding that prices are irrational when extant tests in the literature are carried out (see, e.g., Corradi, Fernandez, and Swanson (2009)).It should be noticed that, however, although point MSFEs are lower in virtually every case (when considering Model A) when u C 1 and u C 2 are included, the absolute magnitude of the difference in MSFEs is rather small, suggesting that the difference is likely insignificant.Moreover, examination of Tables 5 and 7 suggests that there is little information in the revision processes for the other two variables.In particular, for the case of output, notice that in Table 5, MSFEs are lower than those reported in Table 2 for only a small number of cases that correspond to the longest forecasting period starting in 1970:1, and where Model C is now preferred to Model A. Likewise, for the case of money, there is no clear evidence to indicate that the revision process is useful when predicting money, particularly when the two longest forecasting periods are used.This is consistent with our earlier findings and those reported in Corradi, Fernandez, and Swanson (2009) that money is rational.

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25 , and u C4 = ( t+1 u t t−1 , t+1 u t t−2 ).Bold numbers highlight cases in which the inclusion of revision errors lowers the MSFE compared to models without revision errors.A star denotes cases in which a different model than the one found in Tables 2-4 reaches a lower MSFE.Further details are contained in Section 2.

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25  Taken together, these results constitute strong evidence that real-time datasets are indeed useful as failure to use them will result in sub-optimal predictions, when the objective is to minimize MS-FEs.There are many noteworthy empirical analyses in the literature that present evidence concerning the empirical usefulness of real-time datasets.A key early paper that underscores the importance of revisions for predicting industrial production is Diebold and Rudebusch (1991).Hamilton and Perez-Quiros expand on results in the Diebold and Rudebusch paper by asking the question: "What do the leading indicators lead?"These authors find that simple linear models that include leading indicators are useful for predicting GDP.Further empirical evidence on the usefulness of real-time data is discussed in Bernanke and Boivin (2003), Gilbert (2011), Franses (2013), Franses and Segers (2010), and the references cited therein.Our results clearly complement the findings in this literature.

Real-Time Marginal Predictive Content of Money for Output
In this section we implement our experimental setup within a multivariate framework that examines the real-time predictive content of money for income, building on the work of Stock and Watson (1989), Amato and Swanson (2001), Garratt, Koop, Mise, and Vahey (2009), and others.To this end we implement vector versions of Models A and C to examine whether money and revision errors from money and other variables have marginal predictive content for output.Results are gathered in Table 8 and correspond to those reported in Table 2, except that vector autoregressions are estimated rather than autoregressions and the target variable to be predicted is output.Note that models with and without money (and revision errors of money) are estimated.The number of lags in the regression models for output, money, prices, and interest rates is selected using the SIC.
Following the notation used in the previous Section, the models that we examine are: Model A (output equation from associated vector autoregression): Model C:(output equation from associated vector autoregression): , corresponding to the two cases considered (the two cases are denoted u C 1 and u C 2 in Table 8, respectively), t+1 u t 8 are MSFEs, and starred entries denote rejection of the Diebold-Mariano null hypothesis of equal predictive accuracy at a 10% level, using standard normal critical values, and assuming that the benchmark model is the simple autoregression given as Model A in Table 2. Entries in bold denote the lowest MSFEs across all models reported on in the table, for a given value of k.Finally, entries in italics are MSFE-best across all models that include money, for a given value of k.
On inspection of Table 8, it is clear that it is always the case that (regardless of sample period, model, and vintage) the models with money yield higher MSFEs than the models without money (entries that are in bold denote MSFE-best models).This result holds for Models A and C, regardless of whether

Quantitative Finance and Economics
Volume 1, Issue 1, 2-25 or not revision errors are included.Thus, at least based on the comparison of point MSFEs, there is evidence that money does not contain any marginal predictive content for output.This result should be viewed with caution, however, unless the sole purpose of the modeler is to predict output as accurately as possible.In particular, when carrying out policy analysis, for example, one often aims to include control variables that the government can manipulate.Simply specifying an autoregressive model has little use in such cases.For this reason, a better measure of the usefulness of money might be whether these money can be added to the prediction model without worsening predictive performance.If such is the case, then one has evidence that increased parameter and model uncertainty associated with including extra explanatory variables does not worsen predictive performance, hence suggesting that the "bigger" model is "adequately" specified.In light of this argument, note that the lack of starred entries associated with the MSFE-best models in Table 8 (i.e., see entries in bold in the table) for values of k greater than 3 suggests the "adequacy" of vector autoregression models for predicting later release data.This is because the failure of the DM test to reject the null of equal predictive accuracy implies that nothing is lost by moving from a simple autoregression to a vector autoregression.However, this result still tells us nothing about the "adequacy" of models with money.For this reason, we examine the "adequacy" of our models with money by italicizing the MSFE-best models that include money for each release.Interestingly, for our longer prediction periods beginning in 1971 and 1983, models with money do not appear to be "adequate", as the autoregression models yield significantly more accurate predictions, as indicated by the fact that all italicized entries are starred (indicating that the simpler autoregressive model is preferred).However, for the shortest forecast period from 1990, the MSFEbest models with money are "adequate" for all releases except first release data, since in these cases the DM test does not find evidence that simple autoregressive models without models yield more accurate predictions (see the second row of entries in the third panel of the table ).Also interesting is the fact that when considering VAR models, output is always best predicted using varieties of Model A, regardless of release being predicted, placing this variable together with prices as being variables for which use of preliminary data yields the most accurate predictions.Needless to say, the findings of this illustration suggest that there is much to be learned via analysis of real-time datasets, again underscoring the importance of building and maintaining such datasets.

Concluding Remarks
While recent empirical research has presented strong evidence in favor of the usefulness of making real-time datasets available to economists, the literature has not yet carefully assessed the relevance of these datasets for macroeconomic forecasting.In this paper, we attempt to shed new light on these issues by constructing a variety of different real-time prediction models and evaluating their performance in a series of ex-ante prediction experiments that are designed to mimic forecasting approaches used when constructing forecasts in real-time for three macroeconomic variable: Output, prices and money.The prediction models we use include, among others, autoregressive processes that uses only first release data and others that utilize only the latest available data.We also assess the models in a multivariate framework by including revision errors as regressors, hence allowing us to examine the marginal predictive content of the revision process and forming simple rationality tests that are based solely on the examination of ex-ante predictions.In another multivariate application we examine the real-time predictive content of money for income.
Perhaps the most important result we obtain is that the choice of which release of data to predict seems not to have an impact on which releases of data should be used in estimation and prediction construction but that differences in how to utilize real-time datasets do arise when the variable being modelled and predicted changes.For example, we find that regardless of which release of prices one specifies as the "target" variable to be predicted, using only "first release" data in model estimation and prediction construction yields MSFE-best predictions, and that models estimated and implemented using "latest available release" data are MSFE-best for predicting all releases of money.Our experiments that include revision errors as regressors point that early releases of money are rational, whereas prices and output are irrational.As for our multivariate forecasting experiment involving the real-time predictive content of money for income, while we find little marginal predictive content in money, we note that vector autoregressions with money do not perform significantly worse than autoregressions, when predicting output in the past 20 years.Taking together we view our results as providing strong evidence that real-time datasets are indeed useful as failure to use them will result in sub-optimal predictions, when the objective is to minimize MSFEs.Clearly, many of our conclusions would not have been possible without the availability of real-time datasets, underscoring the importance of collecting and maintaining such datasets.
Many issues in this literature remain unresolved.For example, from an empirical perspective it remains to extend the analysis that we carry out to more releases of data and to other variables in order to further examine the relevance of real-time datasets for forecasting.Another topic that we briefly mention is the importance of definitional change, particularly when considering forecasting final releases.From a theoretical perspective, it remains to examine the properties of various predictive accuracy tests in the recursive and real-time framework employed in this paper.

Figure 1 .
Figure 1.Output -Historical Data and Prediction Results.

Figure 2 .
Figure 2. Prices -Historical Data and Prediction Results.

Figure 3 .
Figure 3. Money -Historical Data and Prediction Results.

Table 1 .
Growth Rate and Revision Error Summary Statistics -Output, Prices, and Money ( * ) .
( * ) Summary statistics are reported for a generic variable denoted by y, where y = output, price, and money growth rates (see table rows where vint = 1,2, corresponding to first and second available data), as well where y = the revision error associated with these variables (see table rows where R-Err = 1,2, corresponding to revision errors associated with second and third available datai.e., t+2 u t+1 t and t+3 u t+2 t ).Statistics are reported for samples beginning in 1970:1, 1983:1, and 1990:1.All samples end in 2006:4.

Table 2 .
MSFEs Calculated Based on Simple Real-Time Autoregressions Without Revision Errors for Output ( * ) .* ) In Panel A, forecast mean square errors (MSFEs) are reported based on predictions constructed using recursively estimated models with estimation period beginning in 1965:4 and ex-ante prediction periods beginning in 1970:1, 1983:1, or 1990:1.Corresponding Diebold-Mariano predictive accuracy test statistics are reported in Panel B. (

Table 3 .
MSFEs Calculated Based on Simple Real-Time Autoregressions Without Revision Errors for Prices ( * ) .

Table 4 .
MSFEs Calculated Based on Simple Real-Time Autoregressions Without Revision Errors for Money ( * ). .

Table 5 .
MSFEs Calculated Based on Simple Real-Time Autoregressions With Revision Errors for Output ( * ) .See notes to Table 2. Revision errors included as additional regressors in the prediction equations reported on in Table 2 are:

Table 6 .
MSFEs Calculated Based on Simple Real-Time Autoregressions With Revision Errors for Prices( * ).See notes to Table2 and Table 5.

Table 7 .
MSFEs Calculated Based on Simple Real-Time Autoregressions With Revision Errors for Money( * ).See notes to Table2 and Table 5.

Table 8 .
MSFEs Calculated Based on Real-Time Vector Autoregressions With and Without Money and Revision Errors( * ).See notes to Table2.Vector autoregressions with and without money are used to predict real-time output.Entries are MSFEs, and starred entries denote rejection of the Diebold-Mariano null hypothesis of equal predictive accuracy at a 10% level, using standard normal critical values, and assuming that the benchmark model is Model A from Table2.Entries in bold are the lowest MSFEs across all models reported on in the table, for a given value of k.Finally, entries in italics are MSFE-best across all models that include money, for a given value of k.See Section 2 of the paper for complete details.