Abstract
We propose a method for summarizing the strength of association between a set of variables and a multivariate outcome. Classical summary measures are appropriate when linear relationships exist between covariates and outcomes, while our approach provides an alternative that is useful in situations where complex relationships may be present. We utilize machine learning to detect nonlinear relationships and covariate interactions and propose a measure of association that captures these relationships. A hypothesis test about the proposed associative measure can be used to test the strong null hypothesis of no association between a set of variables and a multivariate outcome. Simulations demonstrate that this hypothesis test has greater power than existing methods against alternatives where covariates have nonlinear relationships with outcomes. We additionally propose measures of variable importance for groups of variables, which summarize each groups’ association with the outcome. We demonstrate our methodology using data from a birth cohort study on childhood health and nutrition in the Philippines.
1 Introduction
In statistical analysis, we often encounter situations where a correlation between a multivariate set of covariates and a multivariate outcome is of scientific interest. A classical approach to this problem is canonical correlation analysis [1], [2]. Canonical correlation maximizes the correlation between a linear combination of the multivariate outcome and a linear combination of the covariates. Several test statistics have been proposed for significance testing of canonical correlation including Wilks’ Λ [3], the Hotelling-Lawley trace [1], [2], the Pillai-Bartlett trace [4], [5], and Roy’s largest root [6]. One potential drawback to canonical correlation analysis is that it may fail to identify associations when nonlinear relationships exist between covariates and outcomes. This has led to recent interest in nonlinear extension of canonical correlation analysis including kernel canonical correlation analysis [7], [8], [9].
Another analytic approach in settings with multivariate outcomes is latent variable analysis. Many definitions of latent variables appear in the literature, and we refer interested readers to more thorough discussions in [10], [11], [12], [13], [14]. A commonly used definition is that a latent variable is a hypothetical construct that cannot be measured by the researcher and that describes some underlying characteristic of the participant. Others have strongly rejected the use of latent variables, instead opting for empirical explanations of observed phenomenon [15]. For our purposes it suffices to say that latent variable analysis tends to entail an unsupervised grouping of the observed outcomes into a set of lower-dimensional latent features. One technique commonly employed to this end is principal components analysis (PCA). Researchers often use PCA to reduce the observed multivariate outcome to a low-dimensional (e.g., univariate) outcome and may examine the factor loadings to ascribe a-posteriori meaning to the reduced outcomes. Researchers further may use these outcomes to test for associations of covariates. However, such tests may fail to identify associations between predictors and outcomes due to the unsupervised nature of the outcomes’ construction.
In this work, we propose an alternative method for measuring association between a multivariate outcome and a set of covariates. Our approach is related to canonical correlation analysis, but rather than maximizing the correlation between linear functions of covariates and outcomes, we maximize the predictive accuracy of a flexible machine learning algorithm and a convex combination of the outcome. The method identifies the univariate outcome that is “most easily predicted” by the covariates and a summary measure of how well that “easiest-to-predict” outcome may be predicted. This approach adapts to complex relationships between covariates and outcomes and identifies associations in situations where traditional canonical correlation does not. However, in contrast to recent nonparametric canonical correlation proposals, our method provides a measure of association on a familiar scale, and also provides asymptotically justified inference, including confidence intervals and hypothesis tests. In certain situations, our method may further provide a novel means of constructing a clinically interpretable latent variable. However, we view our method as a latent variable approach only in the sense of Harman (1960) [16], who described latent variables as a convenient means of summarizing a number of variables in many fewer factors. Our approach fits in this definition by providing a single summary measure of the strength of association between covariates and multivariate outcomes.
2 Methodology
Suppose we observe n independent copies of O := (X1, …, XD, Y1, …, YJ), where X := (X1, …, Xd) is a D-dimensional vector of covariates and Y := (Y1, …, YJ) is a J-dimensional vector of real-valued outcomes. Without loss of generality, we assume that each outcome has mean zero and standard deviation one and otherwise make no assumptions about the true distribution P0 of O. We define
A measure of how well a given algorithm
which measures the average squared distance between the prediction made by
as the proportional reduction in mean squared-error when using
With these ideas in mind, we turn now to the setting where we must simultaneously (i) develop a function
First, we note that for a given ω, the maximizer of
This implies that for any ω, a prediction function for
To begin, we assume that the data are randomly partitioned into K splits of approximately equal size. For k = 1, …, K, we use Tn, k ⊂ {1, …, n} to denote the indices of observations in the k-th training sample (consisting of all the splits expecting the k-th) and Vn,k to denote the indices of observations in the k-th validation sample (i.e., the k-th split itself).
To develop a procedure for sensibly determining outcome weights, we propose an algorithm Ω that maps a given set of data indices into weights that are most promising for prediction using the selected algorithm. The weights based on the k-th training sample are selected by maximizing a (nested) cross-validated measure of accuracy of predictions made by the learner of the combined outcome. We use
Next, the k-th training set is used to develop a prediction function for
Finally, using the respective validation samples, we estimate the following data-adaptive parameter to assess association between X and Y,
where
2.1 Detailed approach
We now provide a step-by-step approach of how our procedure unfolds in a given data set. We include accompanying illustrations for the two-fold cross-validation case (Figures 1 and 2).
Step 1.
Partition the data. Randomly partition the full data set into K splits of approximately equal size. Let Tn,k⊂{1, …, n} to denote the indices of observations in the k-th training sample and Vn,k to denote the indices of observations in the k-th validation sample for k = 1, …, K.
Step 2.
Train a learner for the multivariate outcome using the training data. For each split k, use the training sample to train the chosen learner of the multivariate outcome, resulting in learner Ψ (Tn, k) for k = 1, …, K.
Step 3.
Fit outcome weights using the training data. We now detail our procedure Ω for mapping a given training set into a vector of weights. The following procedure (illustrated in Figured 2 for K′ = 2) is applied in each of the K training samples.
Step 3a.
Partition the given training data. Given a training data set consisting of observations {Oi : i ∈ Tn, k}, randomly partition the data into K′ nested splits of approximately equal size. Let Tn,k,k′ ⊂ Tn, k denote the indices of observations in the k′-th training sample nested in Tn, k and similarly use Vn, k, k′ to denote the indices of observations in the k′-th validation sample nested in Tn, k for k′ = 1, …, K′.
Step 3b.
Fit learner for each outcome in training data. For k′ = 1, …, K′, develop a prediction function using observations Tn,k,k′ of the multivariate outcome, resulting in prediction function
Compute cross-validated R2. For any ω, we can use the data in the k-th validation sample to compute an estimate of mean squared-error of
The cross-validated mean squared-error is the average over the validation folds,
as the empirical average weighted outcome among observations in Tn, k and the empirical mean squared-error for predicting the composite outcome
Maximize cross-validated R2. Compute
Step 4.
Combine learners based on outcome weights. Given weights
Step 5.
Compute cross-validated nonparametric R2. Using each validation sample, compute the mean squared-error for the combined learner,
which provides an evaluation of the performance of the composite learner
via
and define
2.2 Inference
An asymptotically justified confidence interval can be constructed for
where
A closed-form variance estimator is constructed as follows. For k = 1, …, J, we define
These equations represent cross-validated estimates of the influence function of
3 Variable importance
Often we are not only be interested in an overall summary of the association between X and Y, but also in the relative importance of each component of X. Machine learning is often criticized as a “black box” approach, in which it is difficult to understand the relative contributions of different variables [24]. To provide better understanding of the “black box,” we propose to study differences in the estimated association between X and Y when considering different subsets of variables. Our proposal is similar to variable importance measures proposed for specific machine learning algorithms, such as random forests [25], because they measure the change in predictive performance with and without each predictor variable considered. Although existing approaches may have poorly behaved statistical inference [26], the present approach yields straightforward, asymptotically justified inference.
To assess the importance of a subset S ⊂ {1, …, D} of variables
Point estimates for (3) are constructed by plugging in the estimates, and asymptotically justified confidence intervals and hypothesis tests about these estimates are constructed using influence functions as in the previous section.
4 Simulations
We evaluated the finite-sample performance of the proposed estimators in two simulations. In the first, we studied the operating characteristics of our estimator. This simulation confirms that the asymptotic theory developed in previous sections leads to reasonable finite-sample performance for the estimators. In the second simulation, we compared our proposed method to existing methods for assessing associations between X and Y. This simulation shows that our method correctly identifies associations in situations where existing methods do not.
4.1 Simulation 1: operating characteristics
We generated 1000 simulated data sets by independently sampling n copies of X1, X2, X3, X7, X8, X9 from a Uniform(0,4) distribution and n copies of X4, X5, X6 from a Bernoulli distribution with success probabilities of 0.75, 0.25, and 0.5. We sampled
The optimal R-squared for each outcome is about 0.60. The true optimal weighting scheme for the outcomes is
The estimates of predictive performance were biased downwards in smaller sample sizes (Figure 3). However, even with n = 100 the average estimated performance was only about 5% too small. The confidence interval coverage was less than nominal for small sample sizes, but had nominal coverage in larger samples.
We excluded X2 and then excluded X7 to estimate the additive difference in R-squared values (equation (3)) for these two variables. Examining the data-generating mechanism, we note that X2 was important for predicting each individual outcome and should be important for predicting the composite outcome. The optimal R-squared for predicting each individual outcome without X2 was 0.52, and the additive importance of X2 was 0.60 − 0.52 = 0.08. For the composite outcome, the optimal weightings were unchanged (1/3, 1/3, 1/3), but the optimal R-squared without X2 decreased to 0.68 and the additive importance of X2 for the composite outcome was 0.81 − 0.68 = 0.13. In contrast, X7 was important only for predicting Y1 and had no effect on predicting Y2 or Y3. Therefore, the optimal composite outcome was expected to upweight Y2 and Y3 when excluding X7, and the importance of X7 for predicting the optimal composite outcome may be minimal. The optimal weights without X7 were (0.24, 0.38, 0.38), whereas the optimal R-squared without X7 was 0.79, leading to an additive importance of 0.81 − 0.79 = 0.02.
The importance measures for X2 exhibited substantial bias (25% truth) when n = 100, but both measures were unbiased with sample sizes >1000 (Figure 4). The nominal 95% confidence interval coverage was less than nominal in small samples, but the coverage was >90% for sample sizes >500.
4.2 Simulation 2: power of associative hypothesis tests
In this simulation, X was simulated as above. The outcome Y was simulated as a 10-dimensional multivariate normal variate with mean
We studied the power of level 0.05 tests of the hypothesis of no association between X and Y. Power was estimated as 100 times the proportion of the 1,000 total simulations where the null hypothesis was rejected. We tested this hypothesis with the proposed one-sided hypothesis test with all layers of cross validation set to five-fold. We used a super learner library that included an intercept-only regression, a main terms regression, and a generalized additive model [27]. We also tested this hypothesis using four common tests of canonical correlation: Wilks’ Λ, the Hotelling-Lawley trace, the Pillai-Bartlett trace, and Roy’s largest root. Rather than comparing these statistics to their respective approximate asymptotic distributions, we used permutation tests to compute p-values [28]. We also studied the power of a test based on a principal components approach. For this test, we constructed a composite outcome based on the first principal component of Y and subsequently fit a main-terms linear regression on X. The null hypothesis of no association was rejected whenever the p-value associated with the global F-test for this linear regression was less than 0.05.
Under the null hypothesis, each canonical correlation-based and the PCA-based test had approximately nominal type I error (Figure 5, left panel). However, our proposed test was found to be conservative, falsely rejecting the null hypothesis less than 1% of the time. In the second scenario, the canonical correlation-based tests had greater power to reject the null hypothesis than our proposed test at the two smallest sample sizes (middle panel). The PCA-based test had poor power, as the first principal component fails to adequately reflect the component of Y for which a relationship with X exists. In the third scenario, only our proposed test had power to correctly reject the null hypothesis (right panel).
4.3 Additional simulations
In the web supplementary material, we include several additional simulations that explore various settings. In particular, we repeat simulation one using a more aggressive super learner library and using a more complex data generating process. The results in terms of bias and confidence interval coverage are similar as with the results presented here. We also repeated simulation 2 using covariates that each have a Normal distribution, a setting classically associated with canonical correlation analysis. Our results are again largely similar to the results of simulation 2 presented here.
5 Data analysis
Neurocognitive impairment may affect 250 million children under five years globally [30]. The Healthy birth, growth, and development knowledge integration initiative was established, in part, to inform global public health programs toward optimizing neurocognitive development [31]. Neurocognitive development is often studied through studies that enroll pregnant women and follow their children through early childhood and adolescence. Covariate information about the child’s parents, environment, and somatic growth is recorded at regular intervals and in early adolescence, children complete tests that measure diverse domains of neurocognitive development such as motor, mathematics, and language skills. Researchers are often interested in assessing the correlation between covariate information and neurocognitive development. Such an assessment may be useful for developing effective prediction algorithms that identify children at high risk for neurocognitive deficits [32].
The ongoing Cebu Longitudinal Health and Nutrition Study (CLHNS) enrolled Filipino women who gave birth in 1983–1984, and the children born to these women have been followed prospectively [33], [34]. The long-term follow-up with these children has enabled researchers to quantify the long-term effects of prenatal and early childhood nutrition and health on outcomes (adolescent and adult health, economics, and development) [35], [36]. We are interested in assessing the strength of correlation between prenatal and early childhood data and later schooling achievement. We are also interested in understanding the extent to which somatic growth (height and body weight) early in life associates with later neurocognitive outcomes [37], [38]. In 1994, achievement tests were administered to 2166 children in three subjects: mathematics, and English and Cebuano languages.
We applied the present methods to assess association of the three test scores with variables collected from birth to age two years (Table 1). For variables that had missing values, we created an indicator of missingness, and missing values of the original variable were set = 0. The library of candidate super learner algorithms included a random forest, gradient boosted machines, and elastic net regression algorithms. The tuning parameters for each algorithm were chosen via nested five-fold cross-validation. We used 10-fold cross-validation for each layer of cross-validation.
Group | Variables |
---|---|
Health care | Health care access, use of preventive health care |
Household | Child: adult ratio, child dependency ratio, crowding index, urban score |
Socioeconomic status | Total income, socioeconomic status |
Water and sanitation | Sanitation, access to clean water |
Parental | Mother age, father age, mother height, mother education (y), |
Father education (y), marital status, mother age first child, parity | |
Growth | Weight-for-age Z-score, height-for-age z-score [29] (0, 6, 12, 18, 24 mo) |
Other | Mother smoked during pregnancy, child’s sex, gestational age at birth |
The estimated cross-validated R-squared for predicting test scores was 0.24 (95% CI: 0.21, 0.27) for mathematics, 0.31 (95% CI: 0.28, 0.34) for English, and 0.23 (95% CI: 0.20, 0.26) for Cebuano. The estimated association with the composite outcome was 0.32 (95% CI: 0.28, 0.35). The estimated association with the optimally weighted outcome was only slightly higher than predicting an equally weighted outcome (0.30, 95% CI: 0.27, 0.34), as well as an outcome weight based on a linear combination based on the first principal component of the test scores (0.28, 95% CI: 0.25, 0.31).
We computed variable importance measures by repeating the procedure, eliminating groups of variables and estimating the additive change in performance. We found that the child’s sex and parental information were responsible for the largest proportion of the association with achievement test score performance (Figure 6). However, the somatic growth variables, as a group, modestly increased the association, with an estimated change in association of 0.01 (95% CI: 0.00, 0.02; p-value = 0.02).
6 Discussion
Our proposed method provides a new means of summarizing associations with multivariate outcomes and can provide powerful tests for association in large samples when complex and nonlinear relationships are present. We found that existing tests provide greater power in small samples against alternatives with linear relationships between covariates and outcomes. This is unsurprising as our approach relies on several layers of cross-validation, which may stretch small samples too thin. Thus, we suggest that in small samples, it may be preferable to adopt traditional approaches to assessing associations in multivariate settings.
We also found that our test was conservative under the null hypothesis. This can be explained by the fact that the true value of our data-adaptive target parameter was often less than zero, while our test was based on testing a value of zero for this parameter. We suggest that a permutation test could be used to construct a hypothesis test with better operating characteristics; however, this may often prove to be computationally intractable in practice. In such cases, we may instead opt for the more conservative, but less computationally intensive test.
A possible criticism of our approach is the data-adaptive nature of the parameter
An interesting extension suggested by a review is to use cross-validation to select between our proposed measure of association and traditional canonical correlation analysis (CCA). This would require developing an objective criteria that could be used in a given training sample to determine whether there was sufficient evidence to suggest that CCA provides an adequate summary of the association between X and Y. For example, if a linear model for each component of Y has the highest (nested) cross-validated risk in a given training sample, then we may believe that CCA will appropriately describe the association between X and Y. Then, we could use the corresponding validation sample to compute the canonical correlation. Such a modification of our procedure may yield increased power in situations where linear relationships truly are present, while providing an adaptive way to switch to a nonlinear approach. We leave to future work the theoretical and practical study of such an approach. Future extensions may consider nonlinear combinations of the outcome, perhaps via alternating conditional expectation [40], which would bring our proposal closer in scope to that of kernel CCA methods.
Though our proposal is computationally intensive, the vast majority of the computational burden lies in fitting the candidate learners, which can be implemented in a parallelized framework. The computational burden of the procedure may be further reduced by clever choices of number of cross-validation folds. For example, in the context of using super learning to separately develop a prediction function of each of the J outcomes, if K = K′ and we also build a super learner based on K-fold cross-validation, then the procedure requires fitting
It may also be of interest to consider extensions for estimating covariate-adjusted effects of variables on data-adaptively-combined outcome; estimation of these effects and associated inference may be facilitated via cross-validated targeted maximum likelihood [17], [41].
Funding source: National Heart, Lung, and Blood Institute
Award Identifier / Grant number: 1R01HL137808-01A1
Funding source: Bill and Melinda Gates Foundation
Award Identifier / Grant number: OPP1147962
Author contribution: All the authors have accepted responsibility for the entire content of this submitted manuscript and approved submission.
Research funding: This research was funded by National Heart, Lung, and Blood Institute (grant no. 1R01HL137808-01A1) and Bill and Melinda Gates Foundation (grant no. OPP1147962).
Conflict of interest statement: The authors declare no conflicts of interest regarding this article.
An appealing approach for developing a prediction function is regression stacking [19], [20], also known as super learning [21]. Here we describe how to develop a super learner separately for each of the J outcomes, though in theory, this approach could be extended to simultaneous prediction of the vector Y. We first construct a library of Mj candidate learners for each of the J outcomes. The library of estimators can include parametric model-based estimators as well as machine learning algorithms. Parametric model-based methods could include linear regression of Yj on X, while machine learning-based methods could include random forests [25], gradient boosted machines [42], or deep neural networks [43]. If X is high-dimensional, learners can incorporate dimension-reduction strategies. For example, we may include a main-terms linear regression of Yj on all variables in X, while also including a linear regression using only a subset of X chosen based on their univariate association with Yj. The library of learners can also include different choices of tuning parameters for machine learning algorithms, an important consideration for methods that require proper selection of many tuning parameters (e.g., deep learning).
Because there is no way to know a-priori which of these myriad estimators will be best, a cross-validated empirical criterion is used to adaptively combine, or ensemble, the various learners. This approach is appealing in the present application in that the best learner for each outcome might be different. By allowing the data to determine the best ensemble of learners for each outcome, we expect to obtain a better fit over all outcomes and thereby obtain a more accurate summary of the association between X and Y. Indeed, the super learner is known to be optimal in the sense that, under mild assumptions, its goodness-of-fit is essentially equivalent to the (unknown) best-fitting candidate estimator [44], [45]. A full treatment of super learning [21] and the R package [46] are available. Here, we provide a brief overview of the procedure.
Super learning is often based on K*-fold cross validation, and we illustrate the method for K* = 2 in Figure 7. The super learner for a given outcome Yj may be implemented in the following steps:
Step 1:
Partition the data. Starting with a data set consisting of observations
Step 2:
Fit learners in training data. We now use the observations in the training sample to fit each of the Mj candidate learners. We use Ψj,m to denote the m-th candidate learner for the j-th outcome. The learner Ψj,m takes as input a set of indices and returns a prediction function for Yj. We use
Step 3:
Evaluate learners in validation data. Next, we use the data in the k*-th validation sample to evaluate each of the learners
where we use
Step 4:
Find ensemble weights. The cross-validation selector
Step 5:
Refit learners using full data. We next refit the learners that received non-zero weight in
Step 6:
Combine learners into super learner. The super learner is the convex combination of the Mj learners using weights
References
1. Hotelling, H. The most predictable criterion. J Educ Psychol 1935;26:139. https://doi.org/10.1037/h0058165.Search in Google Scholar
2. Hotelling, H. Relations between two sets of variates. Biometrika 1936;28:321–77. https://doi.org/10.1093/biomet/28.3-4.321.Search in Google Scholar
3. Wilks, SS. Certain generalizations in the analysis of variance. Biometrika 1932;24:471–94. https://doi.org/10.2307/2331979.Search in Google Scholar
4. Bartlett, M. The statistical significance of canonical correlations. Biometrika 1941;32:29–37. https://doi.org/10.1093/biomet/32.1.29.Search in Google Scholar
5. Pillai, KS. On the distribution of the largest or the smallest root of a matrix in multivariate analysis. Biometrika 1956;43:122–7. https://doi.org/10.2307/2333585.Search in Google Scholar
6. Roy, SN. The individual sampling distribution of the maximum, the minimum and any intermediate of the p-statistics on the null-hypothesis. Sankhyā: Indian J Stat 1945;7:133–58.Search in Google Scholar
7. Nandy, RR, Cordes, D. Novel nonparametric approach to canonical correlation analysis with applications to low cnr functional mri data. Magn Reson Med 2003;50:354–65. https://doi.org/10.1002/mrm.10537.Search in Google Scholar PubMed
8. Andrew, G, Arora, R, Bilmes, J, Livescu, K. Deep canonical correlation analysis. In: International conference on machine learning; 2013:1247–55 pp.Search in Google Scholar
9. Michaeli, T, Wang, W, Livescu, K. Nonparametric canonical correlation analysis. In: International conference on machine learning; 2016:1967–76 pp.Search in Google Scholar
10. Glymour, C, Scheines, R, Spirtes, P, Kelly, K. Discovering causal structure: artificial intelligence, philosophy of science, and statistical modeling. New York: Academic Press; 1987.10.1016/B978-0-12-286961-7.50010-XSearch in Google Scholar
11. Sobel, M. Causal inference in latent variable models. In: von Eye, A, Clogg, C, editors. Latent variables analysis: applications for developmental research. Thousand Oaks, CA, USA: Sage Publications, Inc.; 1994:3–35 pp.Search in Google Scholar
12. Edwards, JR, Bagozzi, RP. On the nature and direction of relationships between constructs and measures. Psychol Methods 2000;5:155. https://doi.org/10.1037/1082-989x.5.2.155.Search in Google Scholar PubMed
13. Hägglund, G. Milestones in the history of factor analysis. In: Structural equation modeling: present and future. Lincolnwood, IL, USA: Scientific Software Inc.; 2001.Search in Google Scholar
14. Bollen, K. Latent variables in psychology and the social sciences. Annu Rev Psychol 2002;53:605–34. https://doi.org/10.1146/annurev.psych.53.100901.135239.Search in Google Scholar PubMed
15. Skinner, BF. About behaviorism: Vintage; 1976.Search in Google Scholar
16. Harman, HH. Modern factor analysis. University of Chicago Press; 1960.Search in Google Scholar
17. Hubbard, AE, Kherad-Pajouh, S, van der Laan, MJ. Statistical inference for data adaptive target parameters. Int J Biostat 2016;12:3–19. https://doi.org/10.1515/ijb-2015-0013.Search in Google Scholar PubMed
18. Izenman, AJ. Modern multivariate statistical techniques: regression, classification and manifold learning. New York, USA: Springer; 2008.10.1007/978-0-387-78189-1Search in Google Scholar
19. Wolpert, DH. Stacked generalization. Neural Network 1992;5:241–59. https://doi.org/10.1016/s0893-6080(05)80023-1.Search in Google Scholar
20. Breiman, L. Stacked regressions. Mach Learn 1996;24:49–64. https://doi.org/10.1007/bf00117832.Search in Google Scholar
21. van der Laan, MJ, Polley, EC. Super learner. Stat Appl Genet Mol Biol 2007;6:1–23. https://doi.org/10.2202/1544-6115.1309.Search in Google Scholar
22. Ye, Y. Interior algorithms for linear, quadratic, and linearly constrained non-linear programming, Ph.D. thesis: Department of ESS, Stanford University; 1987.Search in Google Scholar
23. Ghalanos, A, Theussl, S. Rsolnp: general non-linear optimization using augmented lagrange multiplier method, r package version 1.16; 2015.Search in Google Scholar
24. Freitas, AA. Comprehensible classification models: a position paper. ACM SIGKDD Explor Newslett 2013;15:1–10. https://doi.org/10.1145/2594473.2594475.Search in Google Scholar
25. Breiman, L. Random forests. Mach Learn 2001;45:5–32. https://doi.org/10.1023/a:1010933404324.10.1023/A:1010933404324Search in Google Scholar
26. Strobl, C, Zeileis, A. Danger: high power!– exploring the statistical properties of a test for random forest variable importance: Department of Statistics: University of Munich Technical ReportsAccessed: November 17, 2016.Search in Google Scholar
27. Hastie, T, Tibshirani, R. Generalized additive models: Wiley Online Library; 1990.Search in Google Scholar
28. Menzel, U. CCP: Significance tests for canonical correlation analysis (CCA), r package version 1.1; 2012. Available from: https://CRAN.R-project.org/package=CCP.Search in Google Scholar
29. WHO Multicentre Growth Reference Study Group. WHO child growth standards: length/height-for-age, weight-for-age, weight-for-length, weight-for-height and body mass index-for-age: methods and development. Geneva: World Health Organization; 2006.Search in Google Scholar
30. Black, M, Walker, S, Fernald, L, Andersen, C, DiGirolamo, A, Lu, C, et al.. Lancet early childhood development series steering committee, early childhood development coming of age: science through the life course. Lancet 2017;389:77–90. https://doi.org/10.1016/S0140-6736(16)31389-7.Search in Google Scholar
31. Jumbe, N, Murray, JC, Kern, S. Data sharing and inductive learning – toward healthy birth, growth, and development. N Engl J Med 2016;374:2415–7. https://doi.org/10.1056/NEJMp1605441.Search in Google Scholar
32. Walker, SP, Wachs, TD, Gardner, JM, Lozoff, B, Wasserman, GA, Pollitt, E, et al.. Child development: risk factors for adverse outcomes in developing countries. Lancet 2007;369:145–57. https://doi.org/10.1016/s0140-6736(07)60076-2.Search in Google Scholar
33. Cebu Longitudinal Health and Nutrition Survey. Available from: http://www.cpc.unc.edu/projects/cebu[Accessed: November 17, 2016].Search in Google Scholar
34. Adair, LS, Popkin, BM, Akin, JS, Guilkey, DK, Gultiano, S, Borja, J, et al.. Cohort profile: the cebu longitudinal health and nutrition survey. Int J Epidemiol 2011:619–25. https://doi.org/10.1093/ije/dyq085.Search in Google Scholar PubMed PubMed Central
35. Daniels, MC, Adair, LS. Growth in young Filipino children predicts schooling trajectories through high school. J Nutr 2004;134:1439–46. https://doi.org/10.1093/jn/134.6.1439.Search in Google Scholar PubMed
36. Daniels, MC, Adair, LS. Breast-feeding influences cognitive development in Filipino children. J Nutr 2005;135:2589–95. https://doi.org/10.1093/jn/135.11.2589.Search in Google Scholar PubMed
37. Carba, DB, Tan, VL, Adair, LS. Early childhood length-for-age is associated with the work status of Filipino young adults. Econ Hum Biol 2009;7:7–17. https://doi.org/10.1016/j.ehb.2009.01.010.Search in Google Scholar PubMed PubMed Central
38. Smithers, LG, Lynch, JW, Yang, S, Dahhou, M, Kramer, MS. Impact of neonatal growth on IQ and behavior at early school age. Pediatrics 2013;132:e53–60. https://doi.org/10.1542/peds.2012-3497.Search in Google Scholar PubMed PubMed Central
39. Feng, J, Williamson, B, Simon, N, Carone, M. Nonparametric variable importance using an augmented neural network with multi-task learning. In: International conference on machine learning; 2018:1496–505 pp.Search in Google Scholar
40. Breiman, L, Friedman, JH. Estimating optimal transformations for multiple regression and correlation. J Am Stat Assoc 1985;80:580–98. https://doi.org/10.1080/01621459.1985.10478157.Search in Google Scholar
41. Zheng, W, van der Laan, MJ. Asymptotic theory for cross-validated targeted maximum likelihood estimation, UC Berkeley Division of Biostatistics Working Paper Series; 2010 Paper 273.10.2202/1557-4679.1181Search in Google Scholar PubMed PubMed Central
42. Friedman, JH. Greedy function approximation: a gradient boosting machine. Ann Stat 2001:1189–232. https://doi.org/10.1214/aos/1013203451.Search in Google Scholar
43. Ripley, BD. Pattern recognition and neural networks. Cambridge, UK: Cambridge University Press; 2007.Search in Google Scholar
44. van der Laan, MJ, Dudoit, S, Keles, S. Asymptotic optimality of likelihood-based cross-validation. Stat Appl Genet Mol Biol 2004;3:1–23. https://doi.org/10.2202/1544-6115.1036.Search in Google Scholar PubMed
45. van der Laan, MJ, Dudoit, S, van der Vaart, A. The cross-validated adaptive epsilon-net estimator. Stat Decis 2006;24:373–95. https://doi.org/10.1524/stnd.2006.24.3.373.Search in Google Scholar
46. Polley, E, LeDell, E, Kennedy, C, Lendle, S, van der Laan, M. Superlearner: super learner prediction, r package version 2.0-21; 2013. Available from: http://CRAN.R-project.org/package=SuperLearner.Search in Google Scholar
Supplementary Material
The online version of this article offers supplementary material (https://doi.org/10.1515/ijb-2019-0061).
© 2020 David Benkeser, et al., published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.