Skip to main content
Log in

Measuring public inflation perceptions and expectations in the UK

  • Published:
Empirical Economics Aims and scope Submit manuscript

Abstract

The Bank of England Inflation Attitudes Survey asks individuals about their inflation perceptions and expectations in eight intervals including an indifference limen. This paper studies fitting a mixture normal distribution to such interval data, allowing for multiple modes. Bayesian analysis is useful since ML estimation may fail. A hierarchical prior helps to obtain a weakly informative prior. The No-U-Turn Sampler speeds up posterior simulation. Permutation invariant parameters are free from the label switching problem. The paper estimates the distributions of public inflation perceptions and expectations in the UK during 2001Q1–2017Q4. The estimated means are useful for measuring information rigidity.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

Notes

  1. Interval data lose some quantitative information, however. If the first or last interval has an open end, then one cannot draw a histogram nor compute its mean. Moreover, one cannot find the median if it lies in the first or last interval with an open end. To solve the problem, one can assume the minimum or maximum of the population distribution, or fit a parametric distribution to the frequency distribution.

  2. If intervals represent durations, then we apply survival analysis.

  3. Since May 2011, the survey asks further questions with more intervals to those who have chosen either category 1 or 8.

  4. Nonparametric kernel density estimation for numerical data does not apply directly to interval data.

  5. Since these panels are not histograms but bar charts, the multiple modes may be spurious. Without knowing the width of the indifference limen, one cannot draw a histogram.

  6. If such estimates seem unreasonable, then one can impose a subjective prior on the lower bound of the indifference limen, which will change the results further.

  7. Fitting a mixture of four normal distributions gives similar results, though error bands are wider.

  8. One may think that numerical data are easier to analyze than interval data. This is not the case if one considers the rounding problem seriously, since some respondents round to integers but others round to multiples of 5 or 10; see Manski and Molinari (2010). To account for such rounding in numerical data on inflation expectations from the Opinion Survey on the General Public’s Views and Behavior conducted by the Bank of Japan, Kamada et al. (2015) introduce point masses at multiples of 5. Binder (2017) interprets heterogeneous rounding as a measure of uncertainty. The rounding problem is irrelevant to our interval data.

  9. A generalized Gibbs sampler by Liu and Sabatti (2000) does not work well in our context.

  10. With covariates, our model extends an ordered probit model by allowing for some known cutpoints and a mixture normal distribution. Lahiri and Zhao (2015) use a hierarchical ordered probit (HOPIT) model, which reduces to an ordered probit model if no covariate is available.

  11. The result seems robust to other priors on \(\beta _0\); e.g., a half-Cauchy prior with a large-scale parameter.

  12. There were five trial surveys quarterly from November 1999 to November 2000.

  13. We set an extremely high target rate to eliminate divergent transitions as much as possible in all 136 quarters. In practice, it can be much lower in most quarters, and the results are almost identical as long as there is no divergent transition.

  14. This increases the maximum number of leapfrog steps from \(2^{11}-1=2047\) to \(2^{13}-1=8191\).

  15. Normal mixture models with \(K \ge 3\) are not identifiable from our data.

  16. The nor1mix package for R is useful for calculating quantiles of a mixture normal distribution.

  17. A noisy information model gives the same regression model if \(\{y_t\}\) is AR(1); see Coibion and Gorodnichenko (2015, sec. I.B).

  18. For \(K=1\), we use the flat prior on \((\mu ,\sigma )\) and the default tuning parameters. For \(K=2\), the estimates are the same as those in Figs. 6 and 7.

  19. The full sample periods are 1987Q4–2017Q4 for the change in the log oil price and 1971Q2–2017Q4 for the change in the log exchange rate.

  20. Coibion and Gorodnichenko (2015, pp. 2664–2665) obtain similar results using data on forecasts of various macroeconomic variables from the Survey of Professional Forecasters.

  21. For US consumers, Coibion and Gorodnichenko (2015, p. 2662) report that \(\beta =.705\) for inflation expectations relative to the CPI, which implies \(\lambda \approx .413\). Thus, information rigidities in inflation expectations among individuals in the UK and USA seem not too different, though the two results may not be directly comparable.

  22. Since the leapfrog method preserves volume, its Jacobian of transformation is 1; see Neal (2011, pp. 117–122). To justify the Metropolis step, the proposal should be \((\varvec{\theta }',-\varvec{z}')\) rather than \((\varvec{\theta }',\varvec{z}')\), but this is unnecessary in practice since \(H(\varvec{\theta }',\varvec{z}')=H(\varvec{\theta }',-\varvec{z}')\) and we discard \(\varvec{z}'\) anyway; see Neal (2011, p. 124). Since the Hamiltonian is approximately constant during the leapfrog method, the acceptance probability is close to 1 if \(\epsilon \) is small.

References

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yasutomo Murasawa.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

I thank three referees and Naohito Abe for useful comments. This work was supported by JSPS KAKENHI Grant Number JP16K03605.

Appendix: HMC methods and the NUTS

Appendix: HMC methods and the NUTS

Understanding an HMC method requires some knowledge of physics. The Boltzmann (Gibbs, canonical) distribution describes the distribution of possible states in a system. Given an energy function E(.) and the absolute temperature T, the pdf of a Boltzmann distribution is \(\forall x\),

$$\begin{aligned} f(x) \propto \exp \left( -\frac{E(x)}{kT}\right) \end{aligned}$$
(A.1)

where k is the Boltzmann constant. One can think of any pdf f(.) as the pdf of a Boltzmann distribution with \(E(x):=-\ln f(x)\) and \(kT=1\).

The Hamiltonian is an energy function that sums the potential and kinetic energies of a state. An HMC method treats \(-\ln p(\varvec{\theta }|\varvec{y})\) as the potential energy at position \(\varvec{\theta }\) and introduces an auxiliary momentum \(\varvec{z}\) drawn randomly from \({\mathrm {N}}(\varvec{0},\varvec{\varSigma })\), whose kinetic energy is \(-\ln p(\varvec{z}) \propto \varvec{z}'\varvec{\varSigma }^{-1}\varvec{z}/2\) with mass matrix \(\varvec{\varSigma }\). The resulting Hamiltonian is

$$\begin{aligned} H(\varvec{\theta },\varvec{z})&:=-\ln p(\varvec{\theta }|\varvec{y})-\ln p(\varvec{z}) \nonumber \\&=-\ln p(\varvec{\theta },\varvec{z}|\varvec{y}) \end{aligned}$$
(A.2)

An HMC method draws from \(p(\varvec{\theta },\varvec{z}|\varvec{y})\), which is a Boltzmann distribution with energy function \(H(\varvec{\theta },\varvec{z})\).

The Hamiltonian is constant over (fictitious) time t by the law of conservation of mechanical energy, i.e., \(\forall t \in \mathbb {R}\),

$$\begin{aligned} \dot{H}(\varvec{\theta }(t),\varvec{z}(t))=0 \end{aligned}$$
(A.3)

or

$$\begin{aligned} \dot{\varvec{\theta }}(t)H_{\varvec{\theta }}(\varvec{\theta }(t),\varvec{z}(t)) +\dot{\varvec{z}}(t)H_{\varvec{z}}(\varvec{\theta }(t),\varvec{z}(t)) =\varvec{0}\end{aligned}$$
(A.4)

Thus, Hamilton’s equation of motion is \(\forall t \in \mathbb {R}\),

$$\begin{aligned} \dot{\varvec{\theta }}(t)&=H_{\varvec{z}}(\varvec{\theta }(t),\varvec{z}(t)) \end{aligned}$$
(A.5)
$$\begin{aligned} \dot{\varvec{z}}(t)&=-H_{\varvec{\theta }}(\varvec{\theta }(t),\varvec{z}(t)) \end{aligned}$$
(A.6)

The Hamiltonian dynamics says that \((\varvec{\theta },\varvec{z})\) moves on a contour of \(H(\varvec{\theta },\varvec{z})\), i.e., \(p(\varvec{\theta },\varvec{z}|\varvec{y})\).

Conceptually, given \(\varvec{\varSigma }\) and an initial value for \(\varvec{\theta }\), an HMC method proceeds as follows:

  1. 1.

    Draw \(\varvec{z}\sim {\mathrm {N}}(\varvec{0},\varvec{\varSigma })\) independently from \(\varvec{\theta }\).

  2. 2.

    Start from \((\varvec{\theta },\varvec{z})\) and apply Hamilton’s equations of motion for a certain length of (fictitious) time to obtain \((\varvec{\theta }',\varvec{z}')\), whose joint probability density equals that of \((\varvec{\theta },\varvec{z})\).

  3. 3.

    Discard \(\varvec{z}\) and \(\varvec{z}'\), and repeat.

This gives a reversible Markov chain on \((\varvec{\theta },\varvec{z})\), since the Hamiltonian dynamics is reversible; see Neal (2011, p. 116). The degree of serial dependence or speed of convergence depends on the choice of \(\varvec{\varSigma }\) and the length of (fictitious) time in the second step. The latter can be fixed or random, but cannot be adaptive if it breaks reversibility.

In practice, an HMC method approximates Hamilton’s equations of motion in discrete steps using the leapfrog method. This requires choosing a step size \(\varepsilon \) and the number of steps L. Because of approximation, the Hamiltonian is no longer constant during the leapfrog method, but adding a Metropolis step after the leapfrog method keeps reversibility. Thus, given \(\varvec{\varSigma }\) and an initial value for \(\varvec{\theta }\), an HMC method proceeds as follows:

  1. 1.

    Draw \(\varvec{z}\sim {\mathrm {N}}(\varvec{0},\varvec{\varSigma })\) independently from \(\varvec{\theta }\).

  2. 2.

    Start from \((\varvec{\theta },\varvec{z})\) and apply Hamilton’s equations of motion approximately by the leapfrog method to obtain \((\varvec{\theta }',\varvec{z}')\).

  3. 3.

    Accept \((\varvec{\theta }',\varvec{z}')\) with probability \(\min \{\exp (-H(\varvec{\theta }',\varvec{z}'))/\exp (-H(\varvec{\theta },\varvec{z})),1\}\).Footnote 22

  4. 4.

    Discard \(\varvec{z}\) and \(\varvec{z}'\), and repeat.

The degree of serial dependence or speed of convergence depends on the choice of \(\varvec{\varSigma }\), \(\varepsilon \), and L. Moreover, the computational cost of each iteration depends on the choice of \(\varepsilon \) and L.

The NUTS developed by Hoffman and Gelman (2014) adaptively chooses L while keeping reversibility. Though the algorithm of the NUTS is complicated, it is easy to use with Stan, a modeling language for Bayesian computation with the NUTS (and other methods). Stan tunes \(\varvec{\varSigma }\) and \(\varepsilon \) adaptively during warm-up. The user only specifies the data, model, and prior. One can call Stan from other popular languages and softwares such as R, Python, MATLAB, Julia, Stata, and Mathematica. The NUTS often has better convergence properties than other popular MCMC methods such as the Gibbs sampler and M–H algorithm.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Murasawa, Y. Measuring public inflation perceptions and expectations in the UK. Empir Econ 59, 315–344 (2020). https://doi.org/10.1007/s00181-019-01675-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00181-019-01675-8

Keywords

JEL Classification

Navigation