Skip to main content
Log in

Probabilistic sensitivity analysis on Markov models with uncertain transition probabilities: an application in evaluating treatment decisions for type 2 diabetes

  • Published:
Health Care Management Science Aims and scope Submit manuscript

Abstract

Markov models are commonly used for decision-making studies in many application domains; however, there are no widely adopted methods for performing sensitivity analysis on such models with uncertain transition probability matrices (TPMs). This article describes two simulation-based approaches for conducting probabilistic sensitivity analysis on a given discrete-time, finite-horizon, finite-state Markov model using TPMs that are sampled over a specified uncertainty set according to a relevant probability distribution. The first approach assumes no prior knowledge of the probability distribution, and each row of a TPM is independently sampled from the uniform distribution on the row’s uncertainty set. The second approach involves random sampling from the (truncated) multivariate normal distribution of the TPM’s maximum likelihood estimators for its rows subject to the condition that each row has nonnegative elements and sums to one. The two sampling methods are easily implemented and have reasonable computation times. A case study illustrates the application of these methods to a medical decision-making problem involving the evaluation of treatment guidelines for glycemic control of patients with type 2 diabetes, where natural variation in a patient’s glycated hemoglobin (HbA1c) is modeled as a Markov chain, and the associated TPMs are subject to uncertainty.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

  1. American Diabetes Association (2015) Standards of medical care in diabetes–2015: summary of revisions. Diab Care 38:S4

    Google Scholar 

  2. Anderson T, Goodman L (1957) Statistical inference about Markov chains. Ann Math Stat 28:89–110

    Article  Google Scholar 

  3. Arias E (2007) 2011 United States life tables. Natl Vital Stat Rep 59

  4. Bennett WL, Maruthur NM, Singh S, Segal JB, Wilson LM, Chatterjee R, Marinopoulos SS, Puhan MA, Ranasinghe P, Block L, Nicholson WK, Hutfless S, Bass EB, Bolen S (2011) Comparative effectiveness and safety of medications for type 2 diabetes: an update including new drugs and 2-drug combinations. Ann Intern Med 154:602–13

    Article  Google Scholar 

  5. Benson RV (1966) Euclidean geomtery and convexity. McGraw-Hill, New York

    Google Scholar 

  6. Bickel PJ, Doksum KA (2007) Mathematical statistics: Basic ideas and selected topics, 2nd edn. Pearson Prentice Hall, Upper Saddle River, NJ

    Google Scholar 

  7. Billingsley P (1995) Probability and measure, 3rd edn. Wiley, New York

    Google Scholar 

  8. Briggs AH, Ades AE, Price MJ (2003) Probabilistic sensitivity analysis for decision trees with multiple branches: Use of the Dirichlet distribution in a Bayesian framework. Med Decis Making 23:341–50

    Article  Google Scholar 

  9. Brillinger DR (1981) Time series: Data analysis and theory. Holden-Day, San Francisco. Expanded edition

    Google Scholar 

  10. Centers for Disease Control and Prevention (2013) Age at diagnosis of diabetes among adult incident cases aged 18–79 years, http://www.cdc.gov/diabetes/statistics/incidence_national.htm

  11. Chen Q, Ayer T, Chhatwal J (2017) Sensitivity analysis in sequential decision models. Med Decis Making 37:243–252

    Article  Google Scholar 

  12. Clarke PM, Gray AM, Briggs A, Farmer AJ, Fenn P, Stevens RJ, Matthews DR, Stratton IM, Holman RR (2004) A model to estimate the lifetime health outcomes of patients with type 2 diabetes: the United Kingdom prospective diabetes study (UKPDS) outcomes model (UKPDS No. 68). Diabetologia 47:1747–1759

    Article  Google Scholar 

  13. Craig A, Sendi PP (2002) Estimation of the transition matrix of a discrete-time Markov chain. Health Econ 11:33–42

    Article  Google Scholar 

  14. Devroye L (1986) Non-uniform random variate generation. Springer-Verlag, New York

    Book  Google Scholar 

  15. Dugundji J (1966) Topology. Allyn and Bacon, Boston

    Google Scholar 

  16. Geyer CJ (1992) Practical Markov chain Monte Carlo. Stat Sci 7:473–483

    Article  Google Scholar 

  17. Geyer CJ (2011) Introduction to Markov chain Monte Carlo. CRC Press, Boca Raton

    Book  Google Scholar 

  18. Goh J, Bayati M, Zenios SA, Singh S, Moore D (2015) Data uncertainty in markov chains: application to cost-effectiveness analyses of medical innovations. Working paper, http://www.hbs.edu/faculty/Publication%20Files/uncertain_markov_chain_319c847c-29b9-485e-a7cd-cd11e745511f.pdf

  19. Gold MR, Siegel JE, Russell LB, Weinstein MC (1996) Cost-effectiveness in health and medicine. Oxford University Press, New York

    Google Scholar 

  20. Gold MR, Stevenson D, Fryback DG (2002) HALYs and QALYs and DALYs, Oh My: Similarities An differences in summary measures in population health. Ann Rev Public Health 23:115–134

    Article  Google Scholar 

  21. Gold RZ (1963) Tests auxiliary to χ 2 tests in a Markov chain. Ann Math Stat 34:56–74

    Article  Google Scholar 

  22. Hörmann W., Leydold J, Derflinger G (2004) Automatic nonuniform random variate generation. Springer-Verlag, Berlin

    Book  Google Scholar 

  23. Horn RA, Johnson CR (2013) Matrix analysis, 2nd edn. Cambridge University Press, New York

    Google Scholar 

  24. Karlin S, Taylor H (1975) A first course in stochastic processes, 2nd edn. Academic Press, New York

    Google Scholar 

  25. Kramer H, Cao G, Dugas L, Luke A, Cooper R, Durazo-Arvizu R (2010) Increasing BMI and waist circumference and prevalence of obesity among adults with type 2 diabetes: the national health and nutrition examination surveys. J Diab Compl 24:368–74

    Article  Google Scholar 

  26. Lada EK, Steiger NM, Wilson JR (2006) Performance evaluation of recent procedures for steady-state simulation analysis. IIE Trans 38:711–727

    Article  Google Scholar 

  27. Lada EK, Wilson JR (2006) A wavelet-based spectral procedure for steady-state simulation analysis. Eur J Oper Res 174:1769–1801

    Article  Google Scholar 

  28. Lada EK, Wilson JR (2008) Sbatch: a spaced batch means procedure for steady-state simulation analysis. J Simul 2:170–185

    Article  Google Scholar 

  29. Lada EK, Wilson JR, Steiger NM, Joines JA (2007) Performance of a wavelet-based spectral procedure for steady-state simulation analysis. INFORMS J Comput 19:150–160

    Article  Google Scholar 

  30. Mandelblatt JS, Stout NK, Schechter CB et al (2016) Collaborative modeling of the benefits and harms associated with different U.S. breast cancer screening strategies. Ann Intern Med 164:215–225

    Article  Google Scholar 

  31. Mannor S, Simester D, Sun P, Tsitsiklis JN (2007) Bias and variance approximation in value function estimates. Manag Sci 53:308–322

    Article  Google Scholar 

  32. Nathan DM, Buse JM, Davidson MB (2009) Medical management of hyperglycemia in type 2 diabetes: A consensus algorithm for the initiation and adjustment of therapy: A consensus statement of the American diabetes association and the European association of the study of diabetes. Diab Care 32:193–203

    Article  Google Scholar 

  33. Neuts MF (1973) Probability. Allyn and Bacon, Boston

    Google Scholar 

  34. Program NCE (2002) Third report of the national cholesterol education program (NCEP) expert panel on detection, evaluation, and treatment of high blood cholesterol in adults (adult treatment panel III) final report. Circulation 106:3143–3421

    Article  Google Scholar 

  35. Puterman ML (2005) Markov decision processes: discrete stochastic dynamic programming. Wiley, New York

    Google Scholar 

  36. Rockafellar RT (1970) Convex analysis. Princeton University Press, Princeton, NJ

    Book  Google Scholar 

  37. Royden HL, Fitzpatrick PM (2010) Real analysis, 4th edn. Prentice Hall, New York

    Google Scholar 

  38. Shiryaev A (1996) Probability, 2nd edn. Springer, New York

    Book  Google Scholar 

  39. Smith RL (1984) Efficient Monte-Carlo procedures for generating points uniformly distributed over bounded regions. Oper Res 32:1296–1308

    Article  Google Scholar 

  40. Steiger NM, Lada EK, Wilson JR, Joines JA, Alexopoulos C, Goldsman D (2005) ASAP3: a batch means procedure for steady-state simulation output analysis. ACM Trans Model Comput Simul 15:39–73

    Article  Google Scholar 

  41. Steiger NM, Wilson JR (2002) An improved batch means procedure for simulation output analysis. Manag Sci 48:1569–1586

    Article  Google Scholar 

  42. Tafazzoli A, Steiger NM, Wilson JR (2011) N-Skart: a nonsequential Skewness- and autoregression-adjusted batch-means procedure for simulation analysis. IEEE Trans Autom Control 56:254–264

    Article  Google Scholar 

  43. Tafazzoli A, Wilson J (2011) Skart: a skewness- and autoregression-adjusted batch-means procedure for simulation analysis. IIE Trans 43:110–128

    Article  Google Scholar 

  44. Tafazzoli A, Wilson JR, Lada EK, Steiger NM (2011) Performance of Skart: a skewness- and autoregression-adjusted batch-means procedure for simulation analysis. INFORMS J Comput 23:297–314

    Article  Google Scholar 

  45. The MathWorks Inc. (2013) MATLAB and Simulink version R2013a

  46. Wright JC, Weinstein MC (1998) Gains in life expectancy from medical interventions — standardizing data on outcomes. England J Med 339:380–386

    Article  Google Scholar 

  47. Yeaw J, Lee WC, Aagren M, Christensen T (2012) Cost of self-monitoring of blood glucose in the United States among patients on an insulin regimen for diabetes. J Manag Care Pharm 18:21–32

    Google Scholar 

  48. Zhang Y, McCoy R, Mason JE, Smith SA, Shah ND, Denton BT (2014) Second-line agents for glycemic control for type 2 diabetes: Are newer agents better? Diab Care 37:1338–1345

    Article  Google Scholar 

Download references

Acknowledgments

This material is based upon work supported in part by the National Science Foundation through Grant Number CMMI 1462060. Any opinion, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yuanhui Zhang.

Appendices

Appendix 1: Convergence properties of algorithm 1 when using algorithm 2 or 3

To describe the overall framework for the following analysis, we build on the setup for Section 2.3 in which (a) qi has the uncertainty set \(\mathcal { Q}_{i}\) contained in the hyperplane \(\mathcal { H}\subset \mathbb {R}^{n}\), and \(\mathcal { H}\) has the relative topology inherited from \(\mathbb {R}^{n}\); and (b) \(\mathcal { Q}_{i}\) is an open, bounded, convex subset of \(\mathcal { H}\). Corresponding to the true TPMs in \(\mathcal { P}^{\,{\boldsymbol {\pi }}}\), we let \(\mathcal { U}^{{\boldsymbol {\pi }}}\) denote the uncertainty set in \(\mathbb {R}^{\tau }\) (where Eq. 4 implies that \(\tau \leq n^{2}(T-1)|\mathcal { A}|\)); and we let \(\mathcal { H}^{\boldsymbol {\pi }} \subset \mathbb {R}^{\tau }\) denote the subspace in which Algorithms 2 and 3 operate, where \(\mathcal { U}^{{\boldsymbol {\pi }}} \subset \mathcal { H}^{\boldsymbol {\pi }}\) and \(\mathcal { H}^{\boldsymbol {\pi }}\) has the relative topology inherited from \(\mathbb {R}^{\tau }\). Hence \(\mathcal { U}^{{\boldsymbol {\pi }}}\) is an open, bounded, convex subset of \(\mathcal { H}^{\boldsymbol {\pi }}\); and in \(\mathbb {R}^{\tau }\), we see that \(\mathcal { H}^{\boldsymbol {\pi }}\) is the Cartesian product of at most \(n(T-1)|\mathcal { A}|\) copies of \(\mathcal { H}\) [15, pp. 98–100].

Each point \(\boldsymbol {u}\in \mathcal { U}^{{\boldsymbol {\pi }}}\) represents an assignment of values to the elements of the TPMs used by the MDP (1) such that Eq. 5 is satisfied for each TPM; and we let \(\mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u})\) denote the resulting expected total reward (3) based on policy π. Let \(\bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\) denote the closure of \(\mathcal { U}^{{\boldsymbol {\pi }}}\) in \(\mathcal { H}^{\boldsymbol {\pi }}\) so that \(\bar {\mathcal {U}}^{{\boldsymbol {\pi }}}\) includes the boundary points of \(\mathcal { U}^{{\boldsymbol {\pi }}}\) [15, pp. 69–72]. Thus \(\bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\) is a closed, bounded, convex set in \(\mathcal { H}^{\boldsymbol {\pi }}\) so that \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\) is also compact in \(\mathcal { H}^{\boldsymbol {\pi }}\) [15, p. 233]. Equations 2 and 3 imply that \(\mathbb {V}_{\boldsymbol {\pi }}(\cdot )\) is a polynomial function of its arguments, so it is continuous on \(\mathbb {R}^{\tau }\); thus \(\mathbb {V}_{\boldsymbol {\pi }}(\cdot )\) is also continuous when its domain is restricted to \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\) [15, p. 79]. Let \(\underline {v}^{{\boldsymbol {\pi }}}\)\(\min \!\left \{\, \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u}) \, |\, \boldsymbol {u} \in \bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\, \right \},\) the minimum value of Eq. 3 over \(\bar {\mathcal { U}}^{{\boldsymbol {\pi }}}\) [15, p. 227]. Similarly let \(\overset {\sim }{\smash {{\underline {v}}}}_{M}\)\(\min \! \left \{\, \mathbb {V}_{\boldsymbol {\pi }}\left (\boldsymbol {U}_{j} \right )\, |\, j = 1,\ldots , M \,\right \}\), the minimum observed value of Eq. 3 over the random vectors \(\{ \boldsymbol {U}_{j} \in \mathcal { U}^{{\boldsymbol {\pi }}} \,|\, j = 1, \ldots , M\}\) corresponding to the sample \(\left \{\, \overset {\sim }{\smash {\mathcal { P}}}_{j}^{{\boldsymbol {\pi }}}\, |\, j = 1,\ldots , M \,\right \}\).

Proof of Property A 1

When Algorithm 1 uses Algorithm 2 to generate {Uj | j = 1,…,M},we show that with probability one \(\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} \rightarrow \underline {v}^{{\boldsymbol {\pi }}}\)as M.Given any ε > 0,let \(g_{M} \equiv \Pr \left \{\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} > \underline {v}^{{\boldsymbol {\pi }}} + \varepsilon \,\right \}\)for M ≥ 1.Because \(\mathbb {K} \equiv \mathbb {V}_{\boldsymbol {\pi }}^{-1}\left \{\,\left (\,\underline {v}^{{\boldsymbol {\pi }}},\, \underline {v}^{{\boldsymbol {\pi }}}+\varepsilon \,\right )\,\right \}\)is an open subset of \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)in \(\mathcal { H}^{\boldsymbol {\pi }}\)and because \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)is compact in \(\mathcal { H}^{\boldsymbol {\pi }}\),the sampling scheme of Algorithm 2 ensures that for each ∈{1,…,M − 1}and \(\boldsymbol {u} \in \bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\),the function \(h(\boldsymbol {u}) \equiv \Pr \left \{ \boldsymbol {U}_{\ell + 1}\;\, \not \in \mathbb {K} \, | \, \boldsymbol {U}_{\ell } =\boldsymbol {u} \, \right \}\)has the following properties: (a) it depends on u but not on ;(b) it is continuous at u;and (c) it attains a lower limit η > 0and an upper limit δ < 1on \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)[15, p. 227]. For \(\boldsymbol {u} \in \mathbb {R}^{\tau }\)and = 0,…,M,we let F()(u)denote thec.d.f.of U,where \(\boldsymbol {U}_{\,0} \in \mathcal { U}^{{\boldsymbol {\pi }}}\)is the user-specified (fixed) initial point of Algorithm 2 so that thec.d.f. F(0)(u)is degenerate [7, p. 193]. Note that \(\bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\)is the support of Ufor ≥ 0[7, p. 161]. Property (c) of h(⋅)and the law of total probability [7, Eq. (33.8)] imply that

$$ \Pr\!\left\{\, \boldsymbol{U}_{\ell} \not \in \mathbb{K}\, \right\} = {\int}_{\bar{\mathcal{ U}}{~}^{{\boldsymbol{\pi}}}} h(\boldsymbol{u}) \, \text{d} F_{(\ell-1)}(\boldsymbol{u}) \geq {\eta} > 0 \text{\ \:for\ \:} \ell = 1, \ldots, M\, . $$
(A.1)

Since \(g_{j} = \Pr \{\boldsymbol {U}_{\ell } \not \in \mathbb {K} \text {\ for\ } \ell = 1, \ldots , j \}\), we proceed byshowing the intermediate result gjδj− 1for j = 1,…,M using inductionon j. Clearly g1 ≤ 1.Then for M ≥ 2and = 1,…,M − 1, thelaw of total probability implies that

$$ \Pr\!\left\{ \boldsymbol{U}_{\ell+ 1}\;\, \not\in \mathbb{K}\, \text{\ and\ } \boldsymbol{U}_{\ell} \not \in \mathbb{K} \right\} = {\int}_{\bar{\mathcal{ U}}{~}^{{\boldsymbol{\pi}}}\setminus\mathbb{K}} h(\boldsymbol{u})\,\text{d} F_{(\ell)}(\boldsymbol{u}) \leq {\delta} \Pr\!\left\{ \boldsymbol{U}_{\ell} \not \in \mathbb{K} \right\} \,; $$

and by Eq. A.1, we have

$$ \Pr\!\left\{\boldsymbol{U}_{\ell+ 1} \,\,\not \in \mathbb{K} \, |\, \boldsymbol{U}_{\ell} \not \in \mathbb{K} \right\} \leq {\delta} < 1 \text{\ for\ } \ell = 1, \ldots, M-1 \, . $$
(A.2)

From Eq. A.2 for = 1 and M = 2, we see that \(g_{2} = g_{1}\Pr \!\left \{ \,\boldsymbol {U}_{\,2}\not \in \mathbb {K}\, | \, \boldsymbol {U}_{1}\not \in \mathbb {K}\, \right \} \leq {\delta }\).If M ≥  3 and j =  2,…, M − 1, we assume byinduction that gjδj− 1.To continue, we observe that the sampling scheme of Algorithm 2 ensures the process {Uj | j =  1,…, M} has theMarkov property. We have

$$\begin{array}{@{}rcl@{}} g_{j + 1} &=& {\int}_{\bar{\mathcal{ U}}{~}^{{\boldsymbol{\pi}}}\setminus\mathbb{K}} \Pr\left\{\, \boldsymbol{U}_{\ell} \not\in\mathbb{K} \text{\ for\ } \ell = 1, \ldots, j-1 \text{\ and\ } \boldsymbol{U}_{j + 1}\not\in\mathbb{K}\,|\, \boldsymbol{U}_{j} = \boldsymbol{u} \,\right\} \, \text{d} F_{(j)}(\boldsymbol{u}) \end{array} $$
(A.3)
$$\begin{array}{@{}rcl@{}} &=&{\int}_{\bar{\mathcal{ U}}{~}^{{\boldsymbol{\pi}}}\setminus\mathbb{K}} \Pr\left\{\, \boldsymbol{U}_{\ell} \not\in\mathbb{K} \text{\ for\ } \ell = 1, \ldots, j-1\,|\, \boldsymbol{U}_{j} = u \,\right\} h(\boldsymbol{u}) \, \text{d} F_{(j)}(\boldsymbol{u}) \end{array} $$
(A.4)
$$\begin{array}{@{}rcl@{}} &\leq& {\delta} {\int}_{\bar{\mathcal{ U}}{~}^{{\boldsymbol{\pi}}}\setminus\mathbb{K}} \Pr\left\{\, \boldsymbol{U}_{\ell} \not\in\mathbb{K} \text{\ for\ } \ell = 1, \ldots, j-1\,|\, \boldsymbol{U}_{j} = \boldsymbol{u} \,\right\}\,\text{d} F_{(j)}(\boldsymbol{u}) \\ &=&{\delta} g_{j} \leq {\delta}^{j}\,, \end{array} $$
(A.5)

where (A.3) follows by the law of total probability, Eq. A.4 follows from the Markov property[38, p. 112], and Eq. A.5 holds by the induction hypothesis. Therefore we have gMδM− 1 → 0for M so we can apply the same line of reasoning as in Remark 2. Hence with probability one there exists Msuch that\(\overset {\sim }{\smash {{\underline {v}}}}_{M} \leq \underline {v}^{\boldsymbol {\pi }} + \varepsilon \)for all MM. Since ε is arbitrary, it follows thatwith probability one \(\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} \rightarrow \underline {v}^{{\boldsymbol {\pi }}}\)as M. Similarly we can prove convergencewith probability one to \({\bar {v}}^{\,{\boldsymbol {\pi }}} = \max \!\left \{\, \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u}) \, |\, \boldsymbol {u} \in \bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\, \right \}\) for thecorresponding sample statistic based on {Uj | j = 1,…,M}.

Next we prove the comparable convergence result for Algorithm 1 when it uses Algorithm 3 for sampling. The notation of theprevious paragraphs is reused. In this case the sampling scheme of Algorithm 3 ensures that the random vectors {Uj | j = 1,…,M} are i.i.d. Based on Eqs. 14and 15, first we show that the {Uj | j = 1,…,M} can be expressed in terms of i.i.d. samples from a truncated u-variate normal distribution for a suitable dimension\(u \leq (n-1)n(T-1)|\mathcal { A}|\). Let Z = [Z1,…,Zu] denote a standard normal random vector with probability density function (p.d.f.)\(\phi _{u}(\boldsymbol {z}) = (2\pi )^{-u/2}\exp \left (-\frac {1}{2}{\sum }_{i = 1}^{u} {z_{i}^{2}}\right )\)for\(\boldsymbol {z} \in \mathbb {R}^{u}\). Let\(\boldsymbol {\zeta }: \mathbb {R}^{u} \mapsto \mathcal { H}^{\boldsymbol {\pi }}\) denote the continuous function representing the u-dimensional generalization of Steps 5 and 6 in Algorithm 3. Since\(\mathcal { U}^{{\boldsymbol {\pi }}}\)is open in\(\mathcal { H}^{\boldsymbol {\pi }}\), we seethat \(\mathcal { F} \equiv \boldsymbol {\zeta }^{-1}(\mathcal { U}^{{\boldsymbol {\pi }}})\) is open in \(\mathbb {R}^{u}\).

By randomly sampling the { Aj | j = 1,…,M }from the u-variate truncated normal p.d.f.

$$ f_{{\boldsymbol{A}}}(\boldsymbol{a}) \equiv \frac{\phi_{u}(\boldsymbol{a})}{{\int}_{\mathcal{ F}} \phi_{u}(\boldsymbol{z}) \, \text{d} \boldsymbol{z} } \text{\ \:for\:\ } \boldsymbol{a} \in \mathcal{ F}\, , $$
(A.6)

we obtain the i.i.d.random vectors { ζ(Aj) | j = 1,…,M }on \(\mathcal { U}^{{\boldsymbol {\pi }}}\). By using an argument similar to Remark 2, we see that Steps 4–10 of Algorithm 3 generatei.i.d.random vectors {Uj | j = 1,…, M} having the samedistribution as the { ζ(Aj) | j = 1,…,M }.Since \(\mathbb {K}\) is openin \(\mathcal { H}^{\boldsymbol {\pi }}\), we seethat \(\mathbb {L} \equiv \boldsymbol {\zeta }^{-1}(\mathbb {K})\) is open in \(\mathbb {R}^{u}\). Because fA(a) >  0 for all \(\boldsymbol {a} \in \mathcal { F}\), we have

$$\begin{array}{@{}rcl@{}} \Pr\left\{\, \boldsymbol{U}_{j} \in \mathbb{K}\, \right\} &=& \Pr\left\{\, \boldsymbol{\zeta}({\boldsymbol{A}}_{j}) \in \mathbb{K}\, \right\} = \Pr\left\{\, {\boldsymbol{A}}_{j} \in \mathbb{L} \, \right\}\\ &=& {\int}_{\mathbb{L}} f_{{\boldsymbol{A}}}(\boldsymbol{a})\, \text{d} \boldsymbol{a} > 0 \end{array} $$

[37, p. 80, Eq. (10)]. Hence \(\delta \equiv \Pr \left \{\, \boldsymbol {U}_{j} \not \in \mathbb {K}\, \right \} < 1 \)for j = 1,…,M so that gM = δM → 0as M. As in the previous paragraph, it follows that with probability one \(\,\!\overset {\sim }{\smash {{\underline {v}}}}_{M} \rightarrow \underline {v}^{{\boldsymbol {\pi }}}\)as M. Similarly we can prove convergence with probability one to\({\bar {v}}^{\,{\boldsymbol {\pi }}} = \max \left \{\, \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {u}) \, |\, \boldsymbol {u} \in \bar {\mathcal { U}}{~}^{{\boldsymbol {\pi }}}\, \right \}\) for the corresponding sample statistic based on {Uj | j = 1,…,M}. □

Proof of Property A 2

Finally we consider the use of Algorithm 1 to estimate the distribution of theexpected total reward (3) based on a given distribution on the uncertainty set\(\mathcal { U}^{{\boldsymbol {\pi }}}\).When Algorithm 1 uses Algorithm 3 for sampling, the random variables \(\left \{\, Y_{j} = \mathbb {V}_{\boldsymbol {\pi }}\left (\boldsymbol {U}_{j}\right ) \, |\, j = 1, \ldots , M\right \}\)are independent with a common c.d.f. F(y) for \(y \in \mathbb {R}\); and letting \( \widehat {F}_{\!\!M}(y)\)(for \(y \in \mathbb {R}\)) denote the empirical c.d.f. for this random sample of size M, we see that with probability one\( \widehat {F}_{\!\!M}(y) \rightarrow F(y)\) uniformly in y as M[7, p. 269]. Moreover the estimator \(\widehat {F}_{\!\!M}(y)\) has mean F(y) and variance F(y)[1 − F(y)]/M; and for M sufficiently large, \( \widehat {F}_{\!\!M}(y)\)is approximately normal [7, p. 358, Ex. 27.1]. □

Proof of Property A 3

When Algorithm 1 uses Algorithm 2 to sample the { Uj | j = 1,…, M} from the uniform distribution on \(\mathcal { U}^{{\boldsymbol {\pi }}}\), the random variables \(\left \{\, Y_{j} = \mathbb {V}_{\boldsymbol {\pi }}\left (\, \boldsymbol {U}_{j}\,\right )\,| \, j = 1, \ldots , M\, \right \}\)are generally neither independent or identically distributed. However, Remark 2 ensures that Ujconverges in distribution to a random vector U that is uniformly distributed on \(\mathcal { U}^{{\boldsymbol {\pi }}}\) as j; and because \(\mathbb {V}_{\boldsymbol {\pi }}(\cdot )\)is a continuous function on \(\mathcal { U}^{{\boldsymbol {\pi }}}\),we see that Yjconverges in distribution to the random variable \(Y^{*} \equiv \mathbb {V}_{\boldsymbol {\pi }}(\boldsymbol {U}^{*})\)having thec.d.f. F(y)(for \(y \in \mathbb {R}\))as j(see [7, pp. 327–329] and [6, Theorem B.7.2 (c)]). Let\( \widehat {F}_{\!\!M}^{*}(y)\) denote the empirical c.d.f. based on the sample { Yj | j = 1,…,M }. Hence beyond a sufficiently long warm-up period for Algorithm 2, we may regard the process { Yj | j = 1,…, M} as approximately covariance stationary [24, p. 445] with marginal c.d.f. F(⋅). In this situation, under certain widely applicable conditions (see [9, Eq. (5.1.3)] and [24, p. 476]), for each \(y\in \mathbb {R}\) separately we see that \( \widehat {F}_{\!\!M}^{*}(y)\) converges in mean square and hence in probability to F(y)as M[33, pp. 258–260]. □

Appendix 2: MLEs and CI estimators of HbA1c transition probabilities

Table 4 The MLE, sample sizes, and lower/upper bounds of transition probabilities’ 99% overall confidence intervals for male patients
Table 5 The MLE, sample sizes, and lower/upper bounds of transition probabilities’ 99% overall confidence intervals for female patients

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, Y., Wu, H., Denton, B.T. et al. Probabilistic sensitivity analysis on Markov models with uncertain transition probabilities: an application in evaluating treatment decisions for type 2 diabetes. Health Care Manag Sci 22, 34–52 (2019). https://doi.org/10.1007/s10729-017-9420-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10729-017-9420-8

Keywords

Navigation