Next Article in Journal
Two-Parameter Exponentially Fitted Taylor Method for Oscillatory/Periodic Problems
Next Article in Special Issue
A Distributed Blocking Flowshop Scheduling with Setup Times Using Multi-Factory Collaboration Iterated Greedy Algorithm
Previous Article in Journal
Mathematical Approach for Mechanical Behaviour Analysis of FGM Plates on Elastic Foundation
Previous Article in Special Issue
An Improved Arithmetic Optimization Algorithm for Numerical Optimization Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of the ADMM Algorithm for a High-Dimensional Partially Linear Model

School of Mathematics and Statistics, Henan University of Science and Technology, Luoyang 471023, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2022, 10(24), 4767; https://doi.org/10.3390/math10244767
Submission received: 17 November 2022 / Revised: 13 December 2022 / Accepted: 14 December 2022 / Published: 15 December 2022
(This article belongs to the Special Issue Optimization Algorithms: Theory and Applications)

Abstract

:
This paper focuses on a high-dimensional semi-parametric regression model in which a partially linear model is used for the parametric part and the B-spline basis function approach is used to estimate the unknown function for the non-parametric part. Within the framework of this model, the constrained least squares estimation is investigated, and the alternating-direction multiplier method (ADMM) is used to solve the model. The convergence is proved under certain conditions. Finally, numerical simulations are performed and applied to workers’ wage data from CPS85. The results show that the ADMM algorithm is very effective in solving high-dimensional partially linear models.
MSC:
62J05; 90C06; 90C25; 90C30

1. Introduction

With the rapid development of modern technology, many fields have generated high-dimensional data, such as in biological information, biomedicine, meteorology, geography, econometrics, machine learning, etc. The term “high-dimensional” refers to the fact that the number of variables in data is much larger than the number of samples. In practical situations, the actual structure of a model is often unknown. If only parametric or non-parametric regression models are used for statistical inference, the results will produce large biases and erroneous conclusions. Therefore, semi-parametric regression models came into being in the 1980s, and Engle first proposed semi-parametric regression models, which contain both parametric and non-parametric components. These are more widely used than parametric or non-parametric models. A semi-parametric regression model is a statistical model in which:
Y = g ( X , β ) + m ( U ) + ε ,
where Y is a real-valued response variable, β R p is a p-dimensional unknown parameter vector, X is a d-dimensional covariate, and g ( · , · ) is a known and measurable function. U [ 0 , 1 ] is a random variable, m ( · ) is a smooth unknown function defined on [ 0 , 1 ] , and ε is a random error.
This paper focuses on high-dimensional semi-parametric regression models in which a partially linear model is used for the parametric part. A partially linear model was proposed by Engle [1] in 1986 when he studied weather and electricity problems. The response variables of the model had linear relationships with some covariates and nonparametric relationships with other covariates, so the partially linear model combined the advantages of the interpretability of linear models with the flexibility of non-parametric models. Partial linear models have been studied by many scholars, such as Heckman (1986) [2], Xu (2019) [3], Chen (2020) [4], Auerbach (2022) [5], etc., and they have achieved many results. Among them, Heckman (1986) [2] proposed a partially linear model with a smooth spline and obtained the consistency and asymptotic normality of the parameter estimation based on Bayesian estimation. Härdle (2000) [6] reviewed a series of studies on partially linear models. In the non-parametric part, the B-spline basis function method was used to estimate the unknown function. The B-spline basis function method is a global smoothing method, and its calculation accuracy and efficiency are relatively high. Numerous scholars have studied and achieved many results for high-dimensional data, such as those of Lasso [7,8,9], SCAD (smoothly clipped absolute deviation) [10,11,12], and MCP (minimax concave penalty) [13,14], and many scholars have also performed much research on high-dimensional partially linear models. Xie (2009) [10] studied SCAD-penalized regression in high-dimensional partially linear models by using polynomial regression splines to estimate the non-parametric part; Ni (2009) [15] proposed a double-penalty partially linear variable selection method that used smooth splines to estimate the non-parametric part. In the case of parameter dispersion, Chen (2012) [16] studied the variable selection problem for the contour-adaptive Elastic-Net for a partially linear model with high-dimensional covariates; Wang (2017) [17] studied constrained-contour least squares estimation based on contour Lagrange multiplier test statistics with linear constraints and gave the convergence speed and asymptotic normality of the least squares estimation.
Wang considered the following partially linear regression model (PLM):
Y = X T β + B T γ + ε ,
where Y is a univariate response variable, X = ( X 1 , , X p ) T R p , and Z R are explanatory variables. We denote Y = ( Y 1 , , Y n ) T ; X = ( X 1 , , X n ) T ; β = ( β 1 , , β p ) T is an unknown p-dimensional parameter vector; ε = ( ε 1 , , ε n ) T . B = B ( Z ) = ( B 1 ( Z ) , , B m n ( Z ) ) T is a set of B-spline basis functions of order r, and γ = ( γ 1 , , γ m n ) T is a spline coefficient vector.
Let ( Y 1 ; X 1 T ; Z 1 ) ( Y n ; X n T ; Z n ) be an independent identically distributed sample of the size of the model. We denote X i = ( X i 1 , , X i p ) T ; model (2) can be approximated by:
Y i = X i T β + B ( Z i ) T γ + ε i .
From Equation (3), we can obtain ε = Y X β B γ . By using the least squares method to estimate the parameters β and γ , minimizing the error is equivalent to:
min β , γ 1 2 Y X β B γ 2 .
In practice, the parameter estimates can also be improved by adding prior information about the regression parameters. The constraint condition is the profile Lagrange multiplier test statistic proposed by Wei and Wu (2008) [18]:
R β = d ,
where R is a given k × p matrix whose rank is k, and d is a known k-dimensional vector.
The study in this paper is equivalent to the solution of the following optimization problem:
min β , γ 1 2 Y X β B γ 2 , s . t . R β = d .
Wang (2017) [17] studied restricted profile least squares estimation, and a Lagrangian function was constructed based on linear constraints. The parameter estimation was performed by using the Lagrange multiplier method. The results showed that the algorithm was efficient when parameter information was available.
Our study considers the optimization problem in (6) by constructing an augmented Lagrangian function with linear constraints and using the alternating-direction method of multipliers (ADMM) to solve the model. The Lagrange multiplier update is a kind of ascending iteration, and its convergence can only be moderately accelerated. Therefore, the Lagrange multiplier method is more time-consuming. The augmented Lagrange multiplier method is a method that combines the Lagrange multiplier method and a penalty function method in one piece, so it is a simple and effective method. Hestenes [19] and Powell [20] first proposed the augmented Lagrangian function and multiplier method for constrained optimization in the late 1960s. The ADMM [21] is a classical algorithm for solving nonlinear problems that was proposed by Glowinski and Marroco in the 1970s. The ADMM is very suitable for convex optimization [22]. This algorithm has a large number of applications in different fields, such as regularized estimation [23], image processing [24], machine learning [25], optimal control [26], and resource allocation for wireless networks [27]. When the scale of a problem is relatively large, a distributed algorithm is faster. Considering the characteristics of the optimization problem in (6), it can be solved in blocks. This is suitable for the algorithmic framework of the ADMM, so this paper will use the ADMM to solve the high-dimensional partially linear model.

2. Introduction to the ADMM Algorithm

In this part, we summarize some useful content for the following discussion.
Firstly, we briefly review the basic knowledge of the ADMM. Our motivation is to apply the ADMM to solve the model in this paper. Let us start from a general convex minimization problem with a separable objective function and linear constraints:
min f ( x ) + g ( z ) , s . t . A x + B z = c ,
where x R m , z R n , A R p × m , B R p × n , c R p , f : R m R , and g : R n R . x and z are independent variables. The augmented Lagrangian function of the minimization problem is:
L p ( x , y , z ) = f ( x ) + g ( z ) + y T ( A x + B z c ) + p 2 A x + B z c 2 2 ,
where y is the Lagrange multiplier and p > 0 is a penalty parameter. The minimization problem can be solved with the augmented Lagrange multiplier method. With a given y 0 , the iterative scheme of the augmented Lagrangian function for the minimization problem is:
( x k + 1 , z k + 1 ) : = arg min { L p ( x , y , z k ) } , y k + 1 : = y k p ( A x k + 1 + B z k + 1 c ) .
The iterative scheme is an application of the augmented Lagrangian function method for solving the above iterations, which require the simultaneous polarization of the variables x and z in each iteration. In addition, the ADMM algorithm decomposes the above iteration into two parts [28] and continuously minimizes the variables; it is expressed as follows:
x k + 1 : = arg min { L p ( x , y k , z k ) } , z k + 1 : = arg min { L p ( x k + 1 , y , z k ) } , y k + 1 : = y k p ( A x k + 1 + B z k + 1 c ) .
The ADMM is widely used, and it is of interest that the subproblem generated by the ADMM must exist in the form of an analytical solution in each iteration.

3. Model and Algorithm

In this section, we will apply the ADMM algorithm to solve the minimization model in this paper and to derive the analytical solution form of each subproblem.

3.1. The High-Dimensional Partially Linear Model

For the optimization problem in (6), by using the augmented Lagrange multiplier method, the constrained programming problem is transformed into an unconstrained optimization problem, and the augmented Lagrangian function is:
min L ρ ( β , γ , λ ) = 1 2 Y X β B γ 2 + λ , R β d + ρ 2 R β d 2 .
Using the alternating-direction method of multipliers (ADMM), its n-step iteration starts from a given ( β n , λ n ) with:
γ n + 1 = arg min γ L ρ ( β n , γ , λ n ) , β n + 1 = arg min β L ρ ( β , γ n + 1 , λ n ) , λ n + 1 = λ n + ρ ( R β n + 1 d ) .
We get the new iteration point at ( γ n + 1 , β n + 1 , λ n + 1 ) .

3.2. Solution of the ADMM for High-Dimensional Partially Linear Models

After a simple calculation, the γ -subproblem in Equation (11) can be written as the following equation:
γ n + 1 = arg min γ { 1 2 Y X β n B γ 2 } .
According to the above method, one can find the partial derivative of γ for any given β :
L ρ ( β , γ , λ ) γ = B ( Y X β n B γ ) = 0 .
The analytical solution of γ can be obtained in the form of:
γ n + 1 = ( B T B ) 1 B T ( Y X β n ) .
For the analytical solution of β , the objective function can be solved by substituting γ into Equation (11):
β n + 1 = arg min β { 1 2 Y X β B γ n + 1 2 + λ n , R β d + ρ 2 R β d 2 } ,
with the following partial derivatives for β :
L ρ ( β , γ n + 1 , λ n ) β = ( X X T + ρ R R T ) 1 ( X T ( Y B γ n + 1 ) + R T ( ρ d λ n ) ) = 0 .
The solutions γ n + 1 , β n + 1 , and λ n + 1 are solved by calculating:
γ n + 1 = ( B T B ) 1 B T ( Y X β n ) , β n + 1 = ( X X T + ρ R R T ) 1 ( X T ( Y B γ n + 1 ) + R T ( ρ d λ n ) ) , λ n + 1 = λ n + ρ ( R β n + 1 d ) .

3.3. Algorithmic Design of ADMM for Solving High-Dimensional Partially Linear Models

In summary, the iterative algorithm for solving high-dimensional partially linear models by using the ADMM can be described as follows.
Step 1. Input the variables X, Y, and B, and given the initial variables ( β 0 , γ 0 , λ 0 ) , select the penalty parameter where ρ > 0 ;
Step 2. Input the iteration step n = 1 , 2 , , N ;
Step 3. Update the parameters γ n + 1 , β n + 1 , and λ n + 1 with Equation (18);
Step 4. Iterate through the loop, returning to step 3 until the termination conditions are met, and the algorithm is terminated;
Step 5. Output ( β N , γ N , λ N ) as the approximate solution ( β * , γ * , λ * ) of (6).

4. Convergence

In this section, we will use a variational inequality to prove the convergence of the algorithm. The Lagrange function of the model is given by:
L ( β , γ , λ ) = 1 2 Y X β B γ 2 + λ , R β d ,
where λ is the Lagrange multiplier.
The solution of Equation (11) is equivalent to finding ( γ * , β * , λ * ) S such that:
B T X β * + B T B γ * B T Y = 0 , ( X X T + ρ R R T ) 1 ( X T ( Y B γ * ) + R T ( ρ d λ * ) ) = 0 , R T β * d = 0 .
Let S * satisfy Equation (19); then, we define ω * = ( γ * , β * , λ * ) S * . Equation (19) is equivalent to a variational problem. We find ( γ * , β * , λ * ) S * such that the following variational inequality holds:
V I ( S , F ) : ( ω ω * ) F ( ω * ) 0 , ω S .
Here,
ω = γ β λ , F ( ω ) = B T X β + B T B γ B T Y ( X X T + ρ R R T ) 1 ( X T ( Y B γ ) + R T ( ρ d λ ) ) R T β d .
We need to use the positive definite matrix G:
G = μ I p X ^ T X ^ 0 0 0 ρ I p 0 0 0 1 ρ I p ,
where X ^ = ( X T , ρ R ) T . For the positive definite matrix G, the following conditions are satisfied: μ > τ ( X T X + ρ R T R ) and μ ( · ) is the spectral radius of the matrix.
In order to establish the convergence of the algorithm, the n + 1 th iteration value of the algorithm is taken as a variational inequality problem. The following lemma can be obtained.
Lemma 1. 
Let { ω n } denote the sequence generated by the algorithm; then, for any ω S ,
( ω ω n + 1 ) ( F ( ω n + 1 ) + M ( β n β n + 1 ) G ( ω n ω n + 1 ) ) 0 ,
where,
M = ρ R T ρ I p 0 p .
Lemma 2. 
Let { ω n } denote the sequence generated by the algorithm; then, for any ω * S * ,
( ω n ω * ) T G ( ω n ω n + 1 ) ( ω n ω n + 1 ) T G ( ω n ω n + 1 ) ( λ n λ n + 1 ) T ( β n β n + 1 ) .
Lemma 3. 
Let { ω n } denote the sequence generated by the algorithm; then, for any ω * S * ,
| | ω n + 1 ω * | | G 2 | | ω n ω * | | G 2 | | ω n ω n + 1 | | G 2
From Lemmas 1 and 2, it can be proved that the sequence { ω n } generated by this algorithm shrinks to the solution set S * . Lemma 3 shows that the sequence { ω n } generated by the algorithm shrinks to the solution set S, and the following corollary can be obtained from Lemma 3.
Corollary 1. 
Let the sequence be generated by the algorithm; then, we get:
  • lim k ω n ω n + 1 G = 0 ;
  • The sequence { ω n } is bounded;
  • For arbitrary ω * S * , the sequence ω n ω * G is non-increasing.
Theorem 1. 
Given any starting point ( γ 0 , β 0 , λ 0 ) S , for any ρ 0 , μ > τ ( X T X + ρ R T R ) , the sequence ω n = ( γ n , β n , λ n ) is generated by the algorithm and converges to ω = ( γ , β , λ ) , where ( γ , β , λ ) is the solution of the model.
Proof. 
From Property 1 of Corollary 3, we can get:
lim k | | γ n γ n + 1 | | = 0 , lim k | | β n β n + 1 | | = 0 , lim k | | λ n λ n + 1 | | = 0 .
By Property 2 of Corollary 3, let ω = ( γ , β , λ ) be one of the clusters, and let the sequence { ω n j } converge to the sequence { ω } , so we can obtain:
γ n j γ , β n j β , λ n j λ ,
and
lim k | | γ n j γ n j + 1 | | = 0 , lim k | | β n j β n j + 1 | | = 0 , lim k | | λ n j λ n j + 1 | | = 0 .
It is proved below that the cluster ω satisfies the optimality condition (19). From Equations (18) and (29), for any ω S , we can obtain:
lim j ( ω ω n j ) F ( ω n j ) 0 .
Then, from Equation (28), for any ω S , the above inequality is transformed into:
( ω ω ) F ( ω ) 0 .
Therefore, the cluster ω satisfies the optimality condition (19), i.e., ω S * . For any n 0 , by Property 3 of Corollary 3, we can obtain:
| | ω n + 1 ω | | G 2 | | ω n ω | | G 2 .
Through the above proof, we can find that the sequence { ω n } has a unique clustering point ω . That is, the sequence { ω n } converges to ω and has ( γ , β , λ ) as the solution of the model. The proof is complete.

5. Simulation and Application

5.1. Parameter Settings

The estimation of a high-dimensional partially linear model is performed through a numerical simulation based on a dataset with a sample size of n generated by the model. The random error terms are ε N ( 0 , σ 2 ) , and X obeys the p-dimensional multivariate normal distribution, i.e., X N ( 0 , Σ ) , where Σ = 0 . 5 j k ; j and k are the jth and kth components of the covariance, respectively. The variable Z obeys the uniform distribution of the interval 0 , 1 , i.e., Z U ( 0 , 1 ) and g ( z ) = 3 cos ( 2 π z ) . The parameter β = ( 1 , 2 , 0.5 , 1 , 0 , , 0 ) T satisfies the constraint β 5 = = β p . The estimation of the smooth function is performed by using cubic spline interpolation and three B-spline basis functions for the numerical simulation. The results are good.

5.2. Simulation Results

According to the above parameter settings, the algorithm proposed in this paper is used, and the specific results are shown in Table 1.
The simulation’s effect is expressed by the mean square error (MSE) of the parameter estimation, MSE = β ^ β 2 , where the sample sizes are n = 100 , 200 and the dimensions are p = 9 , 29 , 49 , 69 , 89 , 109 , 129 , 169 , 209 , 509 , 1009 . The dimensions are taken from small to large by determining the value of the sample size. This was done, on the one hand, to compare with the results of Wang’s [17] study and, on the other hand, to study the simulation’s effect in the case of high dimensions ( p n ). The effect of Wang’s study is expressed by the MSEW, and the effect of this paper is expressed by the MSEC.
According to the results in Table 1, the mean square error for this paper is slightly better than that of Wang’s study in the low-dimensional case, and the mean square error for this paper is slightly lower than that of Wang’s. The results are better in the high-dimensional case because the fitting effect of the augmented Lagrange multiplier method is better than that of the Lagrange multiplier method for high-dimensional data. For fixed values of p, the mean square error becomes larger with increasing σ , and the stability of the parameter estimation also becomes worse with the increase in σ . For fixed values of σ , the mean square error decreases with the increase in the dimensions p of the parameter. The method studied in this paper works better for parameter estimation in high-dimensional cases, and the higher the dimensionality, the better the stability of the parameter estimation.
A line plot of the mean square error for different variances was constructed with a sample size of 100 in order to specifically express the effects of this study (Figure 1, Figure 2 and Figure 3) and to provide a comparison with Wang’s results (Figure 4, Figure 5 and Figure 6).
Line plots of the mean square errors for different variances were constructed with a sample size of 200 to specifically express the effects of this study (Figure 7, Figure 8 and Figure 9) and for a comparison with Wang’s results (Figure 10, Figure 11 and Figure 12).
In these figures, it can be seen that the mean square error (MSEC) of the method studied in this paper is very small and is very close to zero in high dimensions. Moreover, the mean square error of this method is smaller than that of Wang’s method. Therefore, the method studied in this paper is more applicable in the case of high dimensions.

5.3. Application: Workers’ Wage Data Analysis

In order to test the algorithm proposed in this paper, we applied the algorithm to the practical problem of the analysis of workers’ wage data. The workers’ wage data were given by the 1985 Current Population Survey (CPS85) [29]. These data came from reality and are real. Moreover, the indicators of these data contained both quantitative data and classified data, so they were representative. In addition, the data were studied by other authors in the literature to facilitate comparison [30]. The data consisted of 534 samples of CPS85 personnel with 11 variables, which included wages and other characteristics of workers, such as gender, years of education, race, sex, marital status, years of work experience, occupational status, area of residence, and union membership. The wage level did not necessarily have a linear relationship with the years of work experience, so the importance of other variables for wages was mainly considered. The model was built as follows:
Y i = j = 1 10 X i j β j + m ( U i ) + ε i ,
where Y i is the wage of the ith worker, U i is the number of years of experience of the ith worker, X i j is the jth variable of the ith worker, and ε i N ( 0 , σ 2 ) .
We describe the use of the method proposed in this paper to study the important factors that affect wages in this section. In order to reduce the absolute differences between wages, avoid the influence of individual extreme values, and satisfy the assumptions of the linear model as much as possible, a logarithmic transformation was required for the variable of wages.
During the experiment, it was necessary to select training samples and test samples. If the proportion of training samples was large, the model may have been closer to a model trained with all samples. However, if the proportion of test samples was small, the evaluation results would not be accurate enough. If the proportion of test samples was large, that could lead to a large difference between the evaluation model and the previous one, thus reducing the authenticity of the evaluation. In all samples, the division ratio for the training samples and test samples was typically 7:3 to 8:2. For large amounts of data, ratios of 9:1 or even 99:1 can be used. Based on the sample size of CPS85, 75 percent of the samples were selected as training samples, and the method in this paper was used for parameter estimation training. The remaining 25 percent of the samples were used as test samples to predict the wages of workers, and the predicted values were expressed in y ^ i . The test samples were used to evaluate the prediction ability of the model, and the prediction effect was evaluated with the median absolute error (MAE) and the standard error (SE).
The MAE attenuates the effects of outliers. The loss is calculated by taking the median of all absolute differences between the real and the predicted value.
MAE = m e d a i n { | y 1 y ^ 1 | , | y 2 y ^ 2 | , , | y n y ^ n | } .
The SE is a measure of the precision of data and reflects the degree of dispersion of a whole sample from the sample’s mean.
SE = ( y i y ^ i ) 2 n .
The smaller the SE is, the greater the reliability of the prediction will be; otherwise, the reliability of the prediction is small.
The results are shown in Table 2.
As can be seen from the results in Table 2, the value of the MAE was small, indicating that the loss between the predicted and actual values of workers’ wages was lower. The values of the SE were all below 0.09, indicating that the prediction was reliable. In short, the MAE was low and the SE was small, so the parameter estimation method in this paper is relatively efficient.

6. Conclusions

The research in this paper considered a partially linear model with a restricted profile and used the least squares method to estimate the parameters with the purpose of minimizing the error. By constructing an augmented Lagrangian function under linear constraint conditions, the constrained optimization problem was transformed into an unconstrained optimization problem. The model was solved with the ADMM. The ADMM algorithm has the advantage that some large global problems can be solved by decomposing them into several smaller, more easily solvable local subproblems and then coordinating the solutions of the resulting subproblems to obtain the solution of the large global problem. The convergence of the algorithm was obtained by using the method of variational inequality. Through numerical simulations, the results showed that the method of this paper is suitable for parameter estimation in high-dimensional cases. Finally, this paper applied the algorithm to workers’ wage data from CPS85 and analyzed the important factors that affected wages.
In this paper, we used the ADMM algorithm to solve a high-dimensional partially linear model, and the effect was very good. The model in this paper is mainly for convex optimization problems. It can be used to solve other optimization problems, such as non-concave penalty optimization SCAD or MCP. This is a subject that will be studied further.

Author Contributions

Conceptualization, A.F. and X.C.; Methodology, A.F. and X.C.; Software, X.C.; Validation, Y.S. and J.F.; Formal analysis, X.C.; Resources, A.F.; Data curation, J.F.; Writing—original draft, X.C.; Writing—review & editing, A.F. and Y.S.; Visualization, X.C.; Supervision, Y.S.; Project administration, A.F. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the National Natural Science Foundation of China (Nos. 12071112, 11971149, 12101195).

Data Availability Statement

Not applicable.

Acknowledgments

We sincerely thank the three anonymous reviewers for their insightful comments, which greatly improved the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Engle, R.F.; Granger, C.W.J.; Rice, J.; Weiss, A. Semiparametric estimates of the relation between weather and electricity sales. J. Am. Stat. Assoc. 1986, 81, 310–320. [Google Scholar] [CrossRef]
  2. Heckman, N.E. Spline smoothing in a partly linear model. J. R. Stat. Soc. Ser. (Methodol.) 1986, 48, 244–248. [Google Scholar] [CrossRef]
  3. Xu, H.X.; Chen, Z.L.; Wang, J.F.; Fan, G.L. Quantile regression and variable selection for partially linear model with randomly truncated data. Stat. Pap. 2019, 60, 1137–1160. [Google Scholar] [CrossRef]
  4. Chen, W. Polynomial-based smoothing estimation for a semiparametric accelerated failure time partial linear model. Open Access Libr. J. 2020, 7, 1. [Google Scholar] [CrossRef]
  5. Auerbach, E. Identification and estimation of a partially linear regression model using network data. Econometrica 2022, 90, 347–365. [Google Scholar] [CrossRef]
  6. Härdle, W.; Liang, H.; Gao, J. Partially Linear Models; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  7. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
  8. Ranstam, J.; Cook, J.A. LASSO regression. J. Br. Surg. 2018, 105, 1348. [Google Scholar] [CrossRef]
  9. Chetverikov, D.; Liao, Z.; Chernozhukov, V. On cross-validated lasso in high dimensions. Ann. Stat. 2021, 49, 1300–1317. [Google Scholar] [CrossRef]
  10. Xie, H.; Huang, J. SCAD-penalized regression in high-dimensional partially linear models. Ann. Stat. 2009, 37, 673–696. [Google Scholar] [CrossRef]
  11. Zeng, L.; Xie, J. Group variable selection via SCAD-L 2. Statistics 2014, 48, 49–66. [Google Scholar] [CrossRef]
  12. Gao, L.; Li, X.; Bi, D.; Xie, Y. Robust Compressed Sensing based on Correntropy and Smoothly Clipped Absolute Deviation Penalty. In Proceedings of the 2020 IEEE 3rd International Conference on Information Communication and Signal Processing (ICICSP), Shanghai, China, 12–15 September 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 269–273. [Google Scholar]
  13. Zhang, C.H. Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 2010, 38, 894–942. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Breheny, P.; Huang, J. Penalized methods for bi-level variable selection. Stat. Its Interface 2009, 2, 369. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Ni, X.; Zhang, H.H.; Zhang, D. Automatic model selection for partially linear models. J. Multivar. Anal. 2009, 100, 2100–2111. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Chen, B.; Yu, Y.; Zou, H.; Liang, H. Profiled adaptive Elastic-Net procedure for partially linear models with high-dimensional covariates. J. Stat. Plan. Inference 2012, 142, 1733–1745. [Google Scholar] [CrossRef]
  17. Wang, X.; Zhao, S.; Wang, M. Restricted profile estimation for partially linear models with large-dimensional covariates. Stat. Probab. Lett. 2017, 128, 71–76. [Google Scholar] [CrossRef]
  18. Wei, C.H.; Wu, X.Z. Profile Lagrange multiplier test for partially linear varying-coefficient regression models. J. Syst. Sci. Math. Sci. 2008, 28, 416. [Google Scholar]
  19. Hestenes, M.R. Multiplier and gradient methods. J. Optim. Theory Appl. 1969, 4, 303–320. [Google Scholar] [CrossRef]
  20. Powell, M.J.D. A method for nonlinear constraints in minimization problems. Optimization 1969, 1, 283–298. [Google Scholar]
  21. Glowinski, R.; Marroco, A. Sur l’approximation, par éléments finis d’ordre un, et la résolution, par pénalisation-dualité d’une classe de problèmes de Dirichlet non linéaires. Revue française d’automatique, informatique, recherche opérationnelle. Anal. Numer. 1975, 9, 41–76. [Google Scholar]
  22. Boyd, S.; Parikh, N.; Chu, E.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar]
  23. Wahlberg, B.; Boyd, S.; Annergren, M.; Wang, Y. An ADMM algorithm for a class of total variation regularized estimation problems. IFAC Proc. Vol. 2012, 45, 83–88. [Google Scholar] [CrossRef] [Green Version]
  24. Yang, Y.; Sun, J.; Li, H.; Xu, Z. ADMM-CSNet: A deep learning approach for image compressive sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 521–538. [Google Scholar] [CrossRef] [PubMed]
  25. Forero, P.A.; Cano, A.; Giannakis, G.B. Consensus-Based Distributed Support Vector Machines. J. Mach. Learn. Res. 2010, 11, 1663–1707. [Google Scholar]
  26. Glowinski, R.; Song, Y.; Yuan, X.; Yue, H. Application of the Alternating Direction Method of Multipliers to Control Constrained Parabolic Optimal Control Problems and Beyond. Ann. Appl. Math. 2022, 38, 115–158. [Google Scholar] [CrossRef]
  27. Joshi, S.; Codreanu, M.; Latva-aho, M. Distributed SINR balancing for MISO downlink systems via the alternating direction method of multipliers. In Proceedings of the 2013 11th International Symposium and Workshops on Modeling and Optimization in Mobile, Ad Hoc and Wireless Networks (WiOpt), Tsukuba, Japan, 13–17 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 318–325. [Google Scholar]
  28. Gabay, D.; Mercier, B. A dual algorithm for the solution of nonlinear variational problems via finite element approximation. Comput. Math. Appl. 1976, 2, 17–40. [Google Scholar] [CrossRef]
  29. Berndt, E.R. The Practice of Econometrics: Classic and Contemporary; Addison-Wesley Publishing Company: Reading, MA, USA, 1991. [Google Scholar]
  30. Wang, X.; Wang, M. Adaptive group bridge estimation for high-dimensional partially linear models. J. Inequalities Appl. 2017, 2017, 1–18. [Google Scholar]
Figure 1. Folding line plot of the mean square error for a variance of 0.5.
Figure 1. Folding line plot of the mean square error for a variance of 0.5.
Mathematics 10 04767 g001
Figure 2. Folding line plot of the mean square error for a variance of 1.
Figure 2. Folding line plot of the mean square error for a variance of 1.
Mathematics 10 04767 g002
Figure 3. Folding line plot of the mean square error for a variance of 2.
Figure 3. Folding line plot of the mean square error for a variance of 2.
Mathematics 10 04767 g003
Figure 4. Folding line plot of the compared mean square error for a variance of 0.5.
Figure 4. Folding line plot of the compared mean square error for a variance of 0.5.
Mathematics 10 04767 g004
Figure 5. Folding line plot of the compared mean square error for a variance of 1.
Figure 5. Folding line plot of the compared mean square error for a variance of 1.
Mathematics 10 04767 g005
Figure 6. Folding line plot of the compared mean square error for a variance of 2.
Figure 6. Folding line plot of the compared mean square error for a variance of 2.
Mathematics 10 04767 g006
Figure 7. Folding line plot of the mean square error for a variance of 0.5.
Figure 7. Folding line plot of the mean square error for a variance of 0.5.
Mathematics 10 04767 g007
Figure 8. Folding line plot of the mean square error for a variance of 1.
Figure 8. Folding line plot of the mean square error for a variance of 1.
Mathematics 10 04767 g008
Figure 9. Folding line plot of the mean square error for a variance of 2.
Figure 9. Folding line plot of the mean square error for a variance of 2.
Mathematics 10 04767 g009
Figure 10. Folding line plot of the compared mean square error for a variance of 0.5.
Figure 10. Folding line plot of the compared mean square error for a variance of 0.5.
Mathematics 10 04767 g010
Figure 11. Folding line plot of the compared mean square error for a variance of 1.
Figure 11. Folding line plot of the compared mean square error for a variance of 1.
Mathematics 10 04767 g011
Figure 12. Folding line plot of the compared mean square error for a variance of 2.
Figure 12. Folding line plot of the compared mean square error for a variance of 2.
Mathematics 10 04767 g012
Table 1. Comparison of the mean square errors under different conditions.
Table 1. Comparison of the mean square errors under different conditions.
np σ MSECMSEW σ MSECMSEW σ MSECMSEW
10090.50.00260.285310.00260.305720.00260.3975
290.00250.76910.00261.02200.00311.5839
490.0021.12810.00261.51820.00332.6788
690.00352.41410.00423.23380.00575.2201
890.005734.94060.006856.53060.011199.8811
1092.5194 × 10−13160.76253.9866 × 10−13126.06994.7814 × 10−1357.2586
2095.2491 × 10−15215.58295.6079 × 10−15247.23016.2244 × 10−15313.6789
5091.7694 × 10−15867.25141.7710 × 10−15789.72761.9649 × 10−15636.5492
10091.2117 × 10−15446.74701.3816 × 10−15434.20371.4950 × 10−15429.8142
20090.00250.13560.00250.17730.00250.2733
490.00250.00250.00260.81080.00301.5062
890.00250.98570.00281.47250.00322.6198
1290.00291.64490.00322.14370.00423.4911
1690.00393.12270.00524.65910.00848.3544
2091.0400 × 10−126.82657.7886 × 10−1316.38848.6313 × 10−1340.1867
4094.2441 × 10−15221.48694.3281 × 10−15228.53445.2164 × 10−15243.1755
5092.2511 × 10−15266.57852.3479 × 10−15284.43852.6637 × 10−15324.3384
10091.8520 × 10−15369.43911.7060 × 10−15526.35151.7332 × 10−15843.3223
Table 2. The MAE and SE of predictions of workers’ wages.
Table 2. The MAE and SE of predictions of workers’ wages.
VariablesVariable DescriptionMAESE
eduNumber of years of education1.97770.0429
raceNW = 1, W = 00.74910.0861
sexF = 1, M = 00.75260.0860
hispanicHisp = 1, NH = 00.74910.0861
southS = 1, NS = 00.74910.0861
marriedMarried = 1, Single = 00.75210.0858
unionUnion = 1, Not = 00.74910.0860
ageAge1.98850.0517
sectorClerical = 1, Const = 1, Manag = 1, Manuf = 1
Prof = 1, Sales = 1, Service = 1, Other = 0
0.75690.0857
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Feng, A.; Chang, X.; Shang, Y.; Fan, J. Application of the ADMM Algorithm for a High-Dimensional Partially Linear Model. Mathematics 2022, 10, 4767. https://doi.org/10.3390/math10244767

AMA Style

Feng A, Chang X, Shang Y, Fan J. Application of the ADMM Algorithm for a High-Dimensional Partially Linear Model. Mathematics. 2022; 10(24):4767. https://doi.org/10.3390/math10244767

Chicago/Turabian Style

Feng, Aifen, Xiaogai Chang, Youlin Shang, and Jingya Fan. 2022. "Application of the ADMM Algorithm for a High-Dimensional Partially Linear Model" Mathematics 10, no. 24: 4767. https://doi.org/10.3390/math10244767

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop