Abstract
Markov models are commonly used for decision-making studies in many application domains; however, there are no widely adopted methods for performing sensitivity analysis on such models with uncertain transition probability matrices (TPMs). This article describes two simulation-based approaches for conducting probabilistic sensitivity analysis on a given discrete-time, finite-horizon, finite-state Markov model using TPMs that are sampled over a specified uncertainty set according to a relevant probability distribution. The first approach assumes no prior knowledge of the probability distribution, and each row of a TPM is independently sampled from the uniform distribution on the row’s uncertainty set. The second approach involves random sampling from the (truncated) multivariate normal distribution of the TPM’s maximum likelihood estimators for its rows subject to the condition that each row has nonnegative elements and sums to one. The two sampling methods are easily implemented and have reasonable computation times. A case study illustrates the application of these methods to a medical decision-making problem involving the evaluation of treatment guidelines for glycemic control of patients with type 2 diabetes, where natural variation in a patient’s glycated hemoglobin (HbA1c) is modeled as a Markov chain, and the associated TPMs are subject to uncertainty.
Similar content being viewed by others
References
American Diabetes Association (2015) Standards of medical care in diabetes–2015: summary of revisions. Diab Care 38:S4
Anderson T, Goodman L (1957) Statistical inference about Markov chains. Ann Math Stat 28:89–110
Arias E (2007) 2011 United States life tables. Natl Vital Stat Rep 59
Bennett WL, Maruthur NM, Singh S, Segal JB, Wilson LM, Chatterjee R, Marinopoulos SS, Puhan MA, Ranasinghe P, Block L, Nicholson WK, Hutfless S, Bass EB, Bolen S (2011) Comparative effectiveness and safety of medications for type 2 diabetes: an update including new drugs and 2-drug combinations. Ann Intern Med 154:602–13
Benson RV (1966) Euclidean geomtery and convexity. McGraw-Hill, New York
Bickel PJ, Doksum KA (2007) Mathematical statistics: Basic ideas and selected topics, 2nd edn. Pearson Prentice Hall, Upper Saddle River, NJ
Billingsley P (1995) Probability and measure, 3rd edn. Wiley, New York
Briggs AH, Ades AE, Price MJ (2003) Probabilistic sensitivity analysis for decision trees with multiple branches: Use of the Dirichlet distribution in a Bayesian framework. Med Decis Making 23:341–50
Brillinger DR (1981) Time series: Data analysis and theory. Holden-Day, San Francisco. Expanded edition
Centers for Disease Control and Prevention (2013) Age at diagnosis of diabetes among adult incident cases aged 18–79 years, http://www.cdc.gov/diabetes/statistics/incidence_national.htm
Chen Q, Ayer T, Chhatwal J (2017) Sensitivity analysis in sequential decision models. Med Decis Making 37:243–252
Clarke PM, Gray AM, Briggs A, Farmer AJ, Fenn P, Stevens RJ, Matthews DR, Stratton IM, Holman RR (2004) A model to estimate the lifetime health outcomes of patients with type 2 diabetes: the United Kingdom prospective diabetes study (UKPDS) outcomes model (UKPDS No. 68). Diabetologia 47:1747–1759
Craig A, Sendi PP (2002) Estimation of the transition matrix of a discrete-time Markov chain. Health Econ 11:33–42
Devroye L (1986) Non-uniform random variate generation. Springer-Verlag, New York
Dugundji J (1966) Topology. Allyn and Bacon, Boston
Geyer CJ (1992) Practical Markov chain Monte Carlo. Stat Sci 7:473–483
Geyer CJ (2011) Introduction to Markov chain Monte Carlo. CRC Press, Boca Raton
Goh J, Bayati M, Zenios SA, Singh S, Moore D (2015) Data uncertainty in markov chains: application to cost-effectiveness analyses of medical innovations. Working paper, http://www.hbs.edu/faculty/Publication%20Files/uncertain_markov_chain_319c847c-29b9-485e-a7cd-cd11e745511f.pdf
Gold MR, Siegel JE, Russell LB, Weinstein MC (1996) Cost-effectiveness in health and medicine. Oxford University Press, New York
Gold MR, Stevenson D, Fryback DG (2002) HALYs and QALYs and DALYs, Oh My: Similarities An differences in summary measures in population health. Ann Rev Public Health 23:115–134
Gold RZ (1963) Tests auxiliary to χ 2 tests in a Markov chain. Ann Math Stat 34:56–74
Hörmann W., Leydold J, Derflinger G (2004) Automatic nonuniform random variate generation. Springer-Verlag, Berlin
Horn RA, Johnson CR (2013) Matrix analysis, 2nd edn. Cambridge University Press, New York
Karlin S, Taylor H (1975) A first course in stochastic processes, 2nd edn. Academic Press, New York
Kramer H, Cao G, Dugas L, Luke A, Cooper R, Durazo-Arvizu R (2010) Increasing BMI and waist circumference and prevalence of obesity among adults with type 2 diabetes: the national health and nutrition examination surveys. J Diab Compl 24:368–74
Lada EK, Steiger NM, Wilson JR (2006) Performance evaluation of recent procedures for steady-state simulation analysis. IIE Trans 38:711–727
Lada EK, Wilson JR (2006) A wavelet-based spectral procedure for steady-state simulation analysis. Eur J Oper Res 174:1769–1801
Lada EK, Wilson JR (2008) Sbatch: a spaced batch means procedure for steady-state simulation analysis. J Simul 2:170–185
Lada EK, Wilson JR, Steiger NM, Joines JA (2007) Performance of a wavelet-based spectral procedure for steady-state simulation analysis. INFORMS J Comput 19:150–160
Mandelblatt JS, Stout NK, Schechter CB et al (2016) Collaborative modeling of the benefits and harms associated with different U.S. breast cancer screening strategies. Ann Intern Med 164:215–225
Mannor S, Simester D, Sun P, Tsitsiklis JN (2007) Bias and variance approximation in value function estimates. Manag Sci 53:308–322
Nathan DM, Buse JM, Davidson MB (2009) Medical management of hyperglycemia in type 2 diabetes: A consensus algorithm for the initiation and adjustment of therapy: A consensus statement of the American diabetes association and the European association of the study of diabetes. Diab Care 32:193–203
Neuts MF (1973) Probability. Allyn and Bacon, Boston
Program NCE (2002) Third report of the national cholesterol education program (NCEP) expert panel on detection, evaluation, and treatment of high blood cholesterol in adults (adult treatment panel III) final report. Circulation 106:3143–3421
Puterman ML (2005) Markov decision processes: discrete stochastic dynamic programming. Wiley, New York
Rockafellar RT (1970) Convex analysis. Princeton University Press, Princeton, NJ
Royden HL, Fitzpatrick PM (2010) Real analysis, 4th edn. Prentice Hall, New York
Shiryaev A (1996) Probability, 2nd edn. Springer, New York
Smith RL (1984) Efficient Monte-Carlo procedures for generating points uniformly distributed over bounded regions. Oper Res 32:1296–1308
Steiger NM, Lada EK, Wilson JR, Joines JA, Alexopoulos C, Goldsman D (2005) ASAP3: a batch means procedure for steady-state simulation output analysis. ACM Trans Model Comput Simul 15:39–73
Steiger NM, Wilson JR (2002) An improved batch means procedure for simulation output analysis. Manag Sci 48:1569–1586
Tafazzoli A, Steiger NM, Wilson JR (2011) N-Skart: a nonsequential Skewness- and autoregression-adjusted batch-means procedure for simulation analysis. IEEE Trans Autom Control 56:254–264
Tafazzoli A, Wilson J (2011) Skart: a skewness- and autoregression-adjusted batch-means procedure for simulation analysis. IIE Trans 43:110–128
Tafazzoli A, Wilson JR, Lada EK, Steiger NM (2011) Performance of Skart: a skewness- and autoregression-adjusted batch-means procedure for simulation analysis. INFORMS J Comput 23:297–314
The MathWorks Inc. (2013) MATLAB and Simulink version R2013a
Wright JC, Weinstein MC (1998) Gains in life expectancy from medical interventions — standardizing data on outcomes. England J Med 339:380–386
Yeaw J, Lee WC, Aagren M, Christensen T (2012) Cost of self-monitoring of blood glucose in the United States among patients on an insulin regimen for diabetes. J Manag Care Pharm 18:21–32
Zhang Y, McCoy R, Mason JE, Smith SA, Shah ND, Denton BT (2014) Second-line agents for glycemic control for type 2 diabetes: Are newer agents better? Diab Care 37:1338–1345
Acknowledgments
This material is based upon work supported in part by the National Science Foundation through Grant Number CMMI 1462060. Any opinion, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendix 1: Convergence properties of algorithm 1 when using algorithm 2 or 3
To describe the overall framework for the following analysis, we build on the setup for Section 2.3 in which (a) qi has the uncertainty set \(\mathcal { Q}_{i}\) contained in the hyperplane \(\mathcal { H}\subset \mathbb {R}^{n}\), and \(\mathcal { H}\) has the relative topology inherited from \(\mathbb {R}^{n}\); and (b) \(\mathcal { Q}_{i}\) is an open, bounded, convex subset of \(\mathcal { H}\). Corresponding to the true TPMs in \(\mathcal { P}^{\,{\boldsymbol {\pi }}}\), we let \(\mathcal { U}^{{\boldsymbol {\pi }}}\) denote the uncertainty set in \(\mathbb {R}^{\tau }\) (where Eq. 4 implies that \(\tau \leq n^{2}(T-1)|\mathcal { A}|\)); and we let \(\mathcal { H}^{\boldsymbol {\pi }} \subset \mathbb {R}^{\tau }\) denote the subspace in which Algorithms 2 and 3 operate, where \(\mathcal { U}^{{\boldsymbol {\pi }}} \subset \mathcal { H}^{\boldsymbol {\pi }}\) and \(\mathcal { H}^{\boldsymbol {\pi }}\) has the relative topology inherited from \(\mathbb {R}^{\tau }\). Hence \(\mathcal { U}^{{\boldsymbol {\pi }}}\) is an open, bounded, convex subset of \(\mathcal { H}^{\boldsymbol {\pi }}\); and in \(\mathbb {R}^{\tau }\), we see that \(\mathcal { H}^{\boldsymbol {\pi }}\) is the Cartesian product of at most \(n(T-1)|\mathcal { A}|\) copies of \(\mathcal { H}\) [15, pp. 98–100].
Each point \(\boldsymbol {u}\in \mathcal { U}^{{\boldsymbol {\pi }}}\) represents an assignment of values to the elements of the TPMs used by the MDP (1) such that Eq. 5 is satisfied for each TPM; and we let \(\mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u})\) denote the resulting expected total reward (3) based on policy π. Let \(\bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\) denote the closure of \(\mathcal { U}^{{\boldsymbol {\pi }}}\) in \(\mathcal { H}^{\boldsymbol {\pi }}\) so that \(\bar {\mathcal {U}}^{{\boldsymbol {\pi }}}\) includes the boundary points of \(\mathcal { U}^{{\boldsymbol {\pi }}}\) [15, pp. 69–72]. Thus \(\bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\) is a closed, bounded, convex set in \(\mathcal { H}^{\boldsymbol {\pi }}\) so that \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\) is also compact in \(\mathcal { H}^{\boldsymbol {\pi }}\) [15, p. 233]. Equations 2 and 3 imply that \(\mathbb {V}_{\boldsymbol {\pi }}(\cdot )\) is a polynomial function of its arguments, so it is continuous on \(\mathbb {R}^{\tau }\); thus \(\mathbb {V}_{\boldsymbol {\pi }}(\cdot )\) is also continuous when its domain is restricted to \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\) [15, p. 79]. Let \(\underline {v}^{{\boldsymbol {\pi }}}\) ≡\(\min \!\left \{\, \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u}) \, |\, \boldsymbol {u} \in \bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\, \right \},\) the minimum value of Eq. 3 over \(\bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\) [15, p. 227]. Similarly let \(\overset {\sim }{\smash {{\underline {v}}}}_{M}\) ≡\(\min \! \left \{\, \mathbb {V}_{\boldsymbol {\pi }}\left (\boldsymbol {U}_{j} \right )\, |\, j = 1,\ldots , M \,\right \}\), the minimum observed value of Eq. 3 over the random vectors \(\{ \boldsymbol {U}_{j} \in \mathcal { U}^{{\boldsymbol {\pi }}} \,|\, j = 1, \ldots , M\}\) corresponding to the sample \(\left \{\, \overset {\sim }{\smash {\mathcal { P}}}_{j}^{{\boldsymbol {\pi }}}\, |\, j = 1,\ldots , M \,\right \}\).
Proof of Property A 1
When Algorithm 1 uses Algorithm 2 to generate {Uj | j = 1,…,M},we show that with probability one \(\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} \rightarrow \underline {v}^{{\boldsymbol {\pi }}}\)as M →∞.Given any ε > 0,let \(g_{M} \equiv \Pr \left \{\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} > \underline {v}^{{\boldsymbol {\pi }}} + \varepsilon \,\right \}\)for M ≥ 1.Because \(\mathbb {K} \equiv \mathbb {V}_{\boldsymbol {\pi }}^{-1}\left \{\,\left (\,\underline {v}^{{\boldsymbol {\pi }}},\, \underline {v}^{{\boldsymbol {\pi }}}+\varepsilon \,\right )\,\right \}\)is an open subset of \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)in \(\mathcal { H}^{\boldsymbol {\pi }}\)and because \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)is compact in \(\mathcal { H}^{\boldsymbol {\pi }}\),the sampling scheme of Algorithm 2 ensures that for each ℓ ∈{1,…,M − 1}and \(\boldsymbol {u} \in \bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\),the function \(h(\boldsymbol {u}) \equiv \Pr \left \{ \boldsymbol {U}_{\ell + 1}\;\, \not \in \mathbb {K} \, | \, \boldsymbol {U}_{\ell } =\boldsymbol {u} \, \right \}\)has the following properties: (a) it depends on u but not on ℓ;(b) it is continuous at u;and (c) it attains a lower limit η > 0and an upper limit δ < 1on \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)[15, p. 227]. For \(\boldsymbol {u} \in \mathbb {R}^{\tau }\)and ℓ = 0,…,M,we let F(ℓ)(u)denote thec.d.f.of Uℓ,where \(\boldsymbol {U}_{\,0} \in \mathcal { U}^{{\boldsymbol {\pi }}}\)is the user-specified (fixed) initial point of Algorithm 2 so that thec.d.f. F(0)(u)is degenerate [7, p. 193]. Note that \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)is the support of Uℓfor ℓ ≥ 0[7, p. 161]. Property (c) of h(⋅)and the law of total probability [7, Eq. (33.8)] imply that
Since \(g_{j} = \Pr \{\boldsymbol {U}_{\ell } \not \in \mathbb {K} \text {\ for\ } \ell = 1, \ldots , j \}\), we proceed byshowing the intermediate result gj ≤ δj− 1for j = 1,…,M using inductionon j. Clearly g1 ≤ 1.Then for M ≥ 2and ℓ = 1,…,M − 1, thelaw of total probability implies that
and by Eq. A.1, we have
From Eq. A.2 for ℓ = 1 and M = 2, we see that \(g_{2} = g_{1}\Pr \!\left \{ \,\boldsymbol {U}_{\,2}\not \in \mathbb {K}\, | \, \boldsymbol {U}_{1}\not \in \mathbb {K}\, \right \} \leq {\delta }\).If M ≥ 3 and j = 2,…, M − 1, we assume byinduction that gj ≤ δj− 1.To continue, we observe that the sampling scheme of Algorithm 2 ensures the process {Uj | j = 1,…, M} has theMarkov property. We have
where (A.3) follows by the law of total probability, Eq. A.4 follows from the Markov property[38, p. 112], and Eq. A.5 holds by the induction hypothesis. Therefore we have gM ≤ δM− 1 → 0for M →∞ so we can apply the same line of reasoning as in Remark 2. Hence with probability one there exists M∗such that\(\overset {\sim }{\smash {{\underline {v}}}}_{M} \leq \underline {v}^{\boldsymbol {\pi }} + \varepsilon \)for all M ≥ M∗. Since ε is arbitrary, it follows thatwith probability one \(\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} \rightarrow \underline {v}^{{\boldsymbol {\pi }}}\)as M →∞. Similarly we can prove convergencewith probability one to \({\bar {v}}^{\,{\boldsymbol {\pi }}} = \max \!\left \{\, \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u}) \, |\, \boldsymbol {u} \in \bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\, \right \}\) for thecorresponding sample statistic based on {Uj | j = 1,…,M}.
Next we prove the comparable convergence result for Algorithm 1 when it uses Algorithm 3 for sampling. The notation of theprevious paragraphs is reused. In this case the sampling scheme of Algorithm 3 ensures that the random vectors {Uj | j = 1,…,M} are i.i.d. Based on Eqs. 14and 15, first we show that the {Uj | j = 1,…,M} can be expressed in terms of i.i.d. samples from a truncated u-variate normal distribution for a suitable dimension\(u \leq (n-1)n(T-1)|\mathcal { A}|\). Let Z = [Z1,…,Zu] denote a standard normal random vector with probability density function (p.d.f.)\(\phi _{u}(\boldsymbol {z}) = (2\pi )^{-u/2}\exp \left (-\frac {1}{2}{\sum }_{i = 1}^{u} {z_{i}^{2}}\right )\)for\(\boldsymbol {z} \in \mathbb {R}^{u}\). Let\(\boldsymbol {\zeta }: \mathbb {R}^{u} \mapsto \mathcal { H}^{\boldsymbol {\pi }}\) denote the continuous function representing the u-dimensional generalization of Steps 5 and 6 in Algorithm 3. Since\(\mathcal { U}^{{\boldsymbol {\pi }}}\)is open in\(\mathcal { H}^{\boldsymbol {\pi }}\), we seethat \(\mathcal { F} \equiv \boldsymbol {\zeta }^{-1}(\mathcal { U}^{{\boldsymbol {\pi }}})\) is open in \(\mathbb {R}^{u}\).
By randomly sampling the { Aj | j = 1,…,M }from the u-variate truncated normal p.d.f.
we obtain the i.i.d.random vectors { ζ(Aj) | j = 1,…,M }on \(\mathcal { U}^{{\boldsymbol {\pi }}}\). By using an argument similar to Remark 2, we see that Steps 4–10 of Algorithm 3 generatei.i.d.random vectors {Uj | j = 1,…, M} having the samedistribution as the { ζ(Aj) | j = 1,…,M }.Since \(\mathbb {K}\) is openin \(\mathcal { H}^{\boldsymbol {\pi }}\), we seethat \(\mathbb {L} \equiv \boldsymbol {\zeta }^{-1}(\mathbb {K})\) is open in \(\mathbb {R}^{u}\). Because fA(a) > 0 for all \(\boldsymbol {a} \in \mathcal { F}\), we have
[37, p. 80, Eq. (10)]. Hence \(\delta \equiv \Pr \left \{\, \boldsymbol {U}_{j} \not \in \mathbb {K}\, \right \} < 1 \)for j = 1,…,M so that gM = δM → 0as M →∞. As in the previous paragraph, it follows that with probability one \(\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} \rightarrow \underline {v}^{{\boldsymbol {\pi }}}\)as M →∞. Similarly we can prove convergence with probability one to\({\bar {v}}^{\,{\boldsymbol {\pi }}} = \max \left \{\, \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u}) \, |\, \boldsymbol {u} \in \bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\, \right \}\) for the corresponding sample statistic based on {Uj | j = 1,…,M}. □
Proof of Property A 2
Finally we consider the use of Algorithm 1 to estimate the distribution of theexpected total reward (3) based on a given distribution on the uncertainty set\(\mathcal { U}^{{\boldsymbol {\pi }}}\).When Algorithm 1 uses Algorithm 3 for sampling, the random variables \(\left \{\, Y_{j} = \mathbb {V}_{\boldsymbol {\pi }}\left (\boldsymbol {U}_{j}\right ) \, |\, j = 1, \ldots , M\right \}\)are independent with a common c.d.f. F(y) for \(y \in \mathbb {R}\); and letting \( \widehat {F}_{\!\!M}(y)\)(for \(y \in \mathbb {R}\)) denote the empirical c.d.f. for this random sample of size M, we see that with probability one\( \widehat {F}_{\!\!M}(y) \rightarrow F(y)\) uniformly in y as M →∞[7, p. 269]. Moreover the estimator \(\widehat {F}_{\!\!M}(y)\) has mean F(y) and variance F(y)[1 − F(y)]/M; and for M sufficiently large, \( \widehat {F}_{\!\!M}(y)\)is approximately normal [7, p. 358, Ex. 27.1]. □
Proof of Property A 3
When Algorithm 1 uses Algorithm 2 to sample the { Uj | j = 1,…, M} from the uniform distribution on \(\mathcal { U}^{{\boldsymbol {\pi }}}\), the random variables \(\left \{\, Y_{j} = \mathbb {V}_{\boldsymbol {\pi }}\left (\, \boldsymbol {U}_{j}\,\right )\,| \, j = 1, \ldots , M\, \right \}\)are generally neither independent or identically distributed. However, Remark 2 ensures that Ujconverges in distribution to a random vector U∗ that is uniformly distributed on \(\mathcal { U}^{{\boldsymbol {\pi }}}\) as j → ∞; and because \(\mathbb {V}_{\boldsymbol {\pi }}(\cdot )\)is a continuous function on \(\mathcal { U}^{{\boldsymbol {\pi }}}\),we see that Yjconverges in distribution to the random variable \(Y^{*} \equiv \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {U}^{*})\)having thec.d.f. F∗(y)(for \(y \in \mathbb {R}\))as j →∞(see [7, pp. 327–329] and [6, Theorem B.7.2 (c)]). Let\( \widehat {F}_{\!\!M}^{*}(y)\) denote the empirical c.d.f. based on the sample { Yj | j = 1,…,M }. Hence beyond a sufficiently long warm-up period for Algorithm 2, we may regard the process { Yj | j = 1,…, M} as approximately covariance stationary [24, p. 445] with marginal c.d.f. F∗(⋅). In this situation, under certain widely applicable conditions (see [9, Eq. (5.1.3)] and [24, p. 476]), for each \(y\in \mathbb {R}\) separately we see that \( \widehat {F}_{\!\!M}^{*}(y)\) converges in mean square and hence in probability to F∗(y)as M →∞[33, pp. 258–260]. □
Appendix 2: MLEs and CI estimators of HbA1c transition probabilities
Rights and permissions
About this article
Cite this article
Zhang, Y., Wu, H., Denton, B.T. et al. Probabilistic sensitivity analysis on Markov models with uncertain transition probabilities: an application in evaluating treatment decisions for type 2 diabetes. Health Care Manag Sci 22, 34–52 (2019). https://doi.org/10.1007/s10729-017-9420-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10729-017-9420-8