Next Article in Journal
Information Transfer between Stock Market Sectors: A Comparison between the USA and China
Next Article in Special Issue
Machine Learning Photovoltaic String Analyzer
Previous Article in Journal
Decision Tree-Based Sensitive Information Identification and Encrypted Transmission System
Previous Article in Special Issue
Dynamic Analysis and Intelligent Control Strategy for the Internal Thermal Control Fluid Loop of Scientific Experimental Racks in Space Stations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Residual Learning for Nonlinear Regression

1
School of Mathematical and Statistical Sciences, Clemson University, Clemson, SC 29641, USA
2
State Key Laboratory of Atmospheric Boundary Layer Physics and Atmospheric Chemistry, Institute of Atmospheric Physics, Chinese Academy of Sciences, Beijing 100029, China
3
College of Earth Science, University of Chinese Academy of Sciences, Beijing 100049, China
4
State Key Laboratory of Numerical Modeling for Atmospheric Sciences and Geophysical Fluid Dynamics, Institute of Atmospheric Physics, Chinese Academy of Sciences, Beijing 100029, China
5
Forecast Weather (Suzhou) Technology Co., Ltd., Suzhou 215000, China
*
Authors to whom correspondence should be addressed.
Entropy 2020, 22(2), 193; https://doi.org/10.3390/e22020193
Submission received: 5 January 2020 / Revised: 31 January 2020 / Accepted: 4 February 2020 / Published: 7 February 2020
(This article belongs to the Special Issue Intelligent Tools and Applications in Engineering and Mathematics)

Abstract

:
Deep learning plays a key role in the recent developments of machine learning. This paper develops a deep residual neural network (ResNet) for the regression of nonlinear functions. Convolutional layers and pooling layers are replaced by fully connected layers in the residual block. To evaluate the new regression model, we train and test neural networks with different depths and widths on simulated data, and we find the optimal parameters. We perform multiple numerical tests of the optimal regression model on multiple simulated data, and the results show that the new regression model behaves well on simulated data. Comparisons are also made between the optimal residual regression and other linear as well as nonlinear approximation techniques, such as lasso regression, decision tree, and support vector machine. The optimal residual regression model has better approximation capacity compared to the other models. Finally, the residual regression is applied into the prediction of a relative humidity series in the real world. Our study indicates that the residual regression model is stable and applicable in practice.

1. Introduction

Functions are common conceptions in sciences and engineering that quantify the dependence or interaction between one variable and the others. Functions are classified as linear functions and nonlinear functions from the superposition principle prospective. Linear functions are analytical, easy to analyze mathematically, and satisfy the superposition principle, while nonlinear functions are complicated and even nonanalytical. People often use linear models to approximate linear functions such as multiple linear regression [1], stepwise linear regression [2], ridge regression [3], lasso regression [4], and elastic net regression [5], which do not work for nonlinear functions. However, nonlinear functions are more common in the real world. Therefore, the approximation or regression of nonlinear functions has gained a lot of attention and is more practical and meaningful in practice [6,7,8,9,10,11,12,13,14].
One of the classical approximations is the Weierstrass approximation theorem, which shows that every continuous function defined on a close interval can be uniformly approximated as closely as desired by a polynomial function [15]. Support vector regression machine [16,17] is also a well-known nonlinear approximation technique. However, it often takes a long time if support vector regression machines are trained on large datasets, and it is difficult to choose proper kernel functions. In addition, decision tree regression [18] is widely used in practice. However, decision tree can be very non-robust, and it is NP-complete to learn an optimal decision tree where NP is the abbreviation for nondeterministic polynomial time.
Since the 1980s and 1990s, artificial neural networks have been used to approach deterministic nonlinear continuous functions based on the universal approximation theorem, which states that a neural network with a single hidden layer can approximate any continuous function with compact support to arbitrary accuracy, when the width goes to infinity [19,20,21,22]. To improve the efficiency of tuning parameters of single hidden-layer feedforward networks, [23] proposes an incremental constructive method. One may randomly choose hidden nodes and only need to adjust the output weights so that the neural network is fully automatic and works as a universal approximator. To represent multidimensional nonlinear functions by feedforward neural network, [24] investigates the training process and the network approximation properties via linear algebra, which has faster execution speeds and better generalization properties. Neural networks are also used to approximate stochastic processes. [25] reports that some classes of artificial neural networks are capable of approximating prescribed stochastic processes with arbitrary accuracy in the mean square sense, which has improved approximating capabilities of artificial neural networks. [26] considers stochastic neural networks as a generalization of usually defined neural networks and shows that stochastic neural networks are more adherent to the real stochastic world. [27] also applies stochastic neural networks to approximate nonlinear input–output random transformations.
The classic universal approximation theorem relies on “fat” neural networks with infinite widths, which means there are infinitely many neurons in the hidden layer. Nevertheless, “tall” or deep neural networks are not covered in the theorem. Recently, a popular trend in machine learning is to build deeper neural networks (the so-called deep learning), ranging from LeNet [28], AlexNet [29], VGG-Net with tens of layers [30], to GoogLeNet [31] or residual neural network (ResNet) with hundreds of layers [32]. They play a central role in the progresses of machine learning and its applications. It’s observed that deeper networks perform better and approximate nonlinear functions better. [33,34] carry out research on the approximation property of a deep neural network where rectified linear units (ReLU) are used as activation functions. Their studies show that in fully connected networks, if each hidden layer has at least + 1 neurons where is the input dimension, the universal approximation theorem holds as the depth goes to infinity. [35] also shows that a deep ResNet containing only one neuron in every hidden layer can uniformly approximate any Lebesgue integrable function in dimensions.
The traditional ResNet behaves well in images process because of local convolution kernels and deep neural networks. However, the convolution kernels have no effects on the whole sequence. Thus, it is not suitable for the regression of nonlinear functions. Based on the original structure of ResNet, we build a new neural network for nonlinear regressions. Convolutional layers and pooling layers are replaced by fully connected layers in the residual block, so that the new residual neural network is more suitable for nonlinear regressions.
This paper is organized as follows. In the methodology part, the structure of regression model and datasets are introduced. In the results part, we evaluate the approximation capacity of residual regression models by performing numerical experiments. The effects of widths and depths are taken into account. When the optimal parameters are obtained, we compare the optimal residual regression with other linear and nonlinear regression techniques. We also employ the optimal residual regression into the prediction of relative humidity series. Finally, we draw conclusions.

2. Methodology

2.1. Architecture of Deep Regression Models

Residual neural network is one of the most successfully applied deep networks. [32] introduces residual shortcut connections and argues that they are indispensable for training very deep convolutional models, since the shortcuts introduce neither extra parameters nor computation complexity and increase the depth of neural network. Considering the input and output dimensions, residual shortcuts are classified as identity shortcuts and convolution shortcuts as shown in Figure 1. The identity blocks in panel (a) only require addition for tensors and can be directly used when the input and output have the same dimensions. Linear projection shortcuts are used when the dimensions are different. One linear projection shortcut consists of a 1 × 1 convolution unit and a Batch Normalization unit. Therefore, it is also called a convolution block as shown in panel (b). The 1 × 1 convolution unit can change the input dimension so that the input dimension can match with the output dimension, and then addition operations can be performed. The inputs of ResNet are often images with many channels, and it behaves well in vision imagery. However, the inputs of nonlinear functions are often one-dimensional vectors, and convolution is a not a global but rather a local operator. This means that if vectors are reshaped into matrices, the convolution kernels cannot affect the whole sequence but rather only parts of the sequence. This contradicts with the aim of nonlinear regression. Therefore, the original architecture of ResNet is not suitable for the nonlinear regression issues.
Based on the structure of ResNet, we build a new neural network for nonlinear regression. Its architecture is presented in Figure 2. Convolutional layers and pooling layers are replaced by fully connected layers (or dense layers) in the residual block. Batch Normalization layers from the primary model are kept in our new model, which act as a regularizer in some cases, eliminating the need for Dropout, and allow people to use much higher learning rates and care less about initialization [36]. Panel (a) shows identity blocks that are used when the input and output are in the same dimensions. The dense blocks in panel (b) correspond to convolution blocks in Figure 1 and are used when the dimensions are different.
Usually, deep learning uses a multilayer network and employs the gradient algorithm to train models, so executing deep learning requires heavy computation, and the learning is often trapped into a saddle point or local minima [37]. To tackle this issue, people propose the rectified linear unit (ReLU) as the activation function, whose gradient can be easily computed. ReLU is shown as follows
R e L U ( x ) = max ( 0 , x ) .
In this paper, the new residual regression model employs tens or hundreds of layers. Hence, to speed up the learning convergence, ReLU is applied as activation function in hidden layers. For the output (or top) layer, linear activation function is used to meet the nonlinear regression requirement.
The fundamental blocks of our model are shown in Table 1. The output dimension of the input block is equal to the input dimension of the first hidden layer. Usually, the input and output of an identity block have the same dimension ( N 1 ), but it is not true for the input dimension ( N 2 ) and output dimension ( N 3 ) of a dense block ( N 2 N 3 ). A dense block is usually inserted between two identity blocks when the output shape of one identity block is different from the input shape of the other. The structure of regression model is presented in Figure 3. The model begins with an input block and is followed by dense blocks and identity blocks. The output block is in the end. In this paper, every dense block is followed by two identity blocks, and then it is followed by one dense block, and so forth. The last two identity blocks are followed by the output layer.

2.2. Datasets

In this part, we introduce multiple nonlinear functions to evaluate the residual regression model. The order of nonlinearity corresponds to the number of minimum functions. Functions with the order of nonlinearity from 1 to 4 are shown as Equations (2)–(5) where x i is uniformly distributed in the interval [ 0 ,   4 ] .
y = min ( i = 0 6 ( x i ) i , 400 ) ,   x i ~ U ( 0 , 4 ) ,
y   y = min ( m i n ( i = 0 6 ( x i ) i , 800 ) , 400 ) ,   x i ~ U ( 0 , 4 ) ,
y = min ( m i n ( min ( i = 0 6 ( x i ) i , 1200 ) , 800 ) , 400 ) , x i ~ U ( 0 , 4 ) ,
y = min ( m i n ( min ( min ( i = 0 6 ( x i ) i , 1600 ) , 1200 ) , 800 ) , 400 ) , x i ~ U ( 0 , 4 ) .
In this paper, 10,000,000 samples are generated for each function. The datasets are shown in Figure 4. Panels (a), (b), (c), and (d) are samples corresponding to Equations (2)–(5) respectively. The residual regression models are trained on 6,750,000 samples and validated on 750,000 samples. The remaining 2,500,000 samples serve as testing data.

3. Results

3.1. Regression Models with Different Depths and Widths

As mentioned, the depth and width of ResNet can affect the approximation capacity. In order to evaluate the effects, we fix one factor and consider the influence of the other. Exactly speaking, when we assess the effect of depth, the width of every hidden layer is fixed, and the depth is changed. Nevertheless, when we consider the effect of width, the depth of ResNet is fixed, and the width is changed. Training data in this part are generated by Equation (4). Before training, the original data is standardized by the Min-Max scaler.
  w k ^ = max i w i w k max i w i min i w i ,  
where   w k ^ stands for standardized data of w k . The residual regression models are built by Keras using TensorFlow as the backend and are trained on computer clusters with 64 CPUs and 126 GB Random Access Memory (RAM). The type of CPUs used is an Intel(R) Xeon(R) CPU E5-2683 V4 working at 2.10 GH. Every core has two threads. Training processing is also accelerated by two graphics processing units (GPUs). The type of GPUs is a GeForce GTX 1080 Ti produced by NVIDIA, and every GPU has a 10421 MB memory. We use mean squared error (MSE) as the loss function.
l o s s = 1 n k = 1 n ( y k y k ^ ) 2 .
The Adam method is applied to minimize the loss function. It computes individual adaptive learning rates for different parameters [38] and combines the advantages of AdaGrad, which works well with sparse gradients [39], and RMSProp which works well in on-line and non-stationary settings [40].
An early stopping strategy is also used to avoid overfitting. When training large models, people often observe that training loss decreases over time, while validation loss begins to rise again. This means that a model with better validation loss can be obtained by returning to the time with the lowest validation loss, which is known as an early stopping strategy and is probably the most commonly used form of regularization in deep learning due to its effectiveness and simplicity [41]. The algorithm stops when no progress has been made over the best recorded validation loss for some pre-specified number (or patience) of epochs.
In this paper, the batch size for gradient descent is 5000, and the epoch number of training is 50. The patience of early stopping is 10 epochs. The train loss, validation loss, and testing loss are computed by standardized data and have the magnitude of 10−4. Table 2 shows the training process of regression models with different depths. Before considering the effects of depth and width, we have trained residual regression models with different widths for many times and found that testing loss is small when the width is fixed at 20. Hence by experience, the width of every hidden layer is initially fixed at 20.
From Table 2, we could see that when the depth is less than 100, testing losses range from 3.4855 × 10 4 to 5.3401 × 10 4 . The testing losses are in the same magnitude and have small changes. However, when the depth is equal to or beyond 100, the testing and validation losses increase greatly. The testing losses vary from 1.8493 × 10 3 to 2.4995 × 10 1 . Thus, the optimal depth of residual regression model is approximately 28, since the model has a minimum testing loss, i.e., 3.4855 × 10 4 . Besides, it is not true that if the neural network is deeper, it behaves better on approximation or regression. This is because neural networks use the back propagation (BP) algorithm to minimize their loss functions, but it is difficult to optimize very deep neural networks.
Based on the optimal depth, we consider the effect of width. Table 3 shows the training information of residual regression model with different widths. The depth is fixed to the optimal depth, i.e., 28. From Table 3, we could know that the optimal width is approximately 16, since the model has a minimum testing loss. When the width changes from 8 to 700, testing losses range from 2.1360 × 10 4 to 8.1343 × 10 4 , which have the same magnitude 10 4 and have small variations. Nevertheless, when the width is small (less than 4), testing and validation losses are greater than 1 × 10 3 . Hence, the approximation capacities of wide neural networks are stronger than very narrow neural networks. This is because residual regression models with small widths are too simple to approximate complex nonlinear mappings. Therefore, it is not recommended to set the residual regression model with a great depth (more than 100) or a small width (less than 4).

3.2. Trainng the Optimal Regression Model on Nonlinear Datasets

Table 4 shows the training information of an optimal regression model on simulated nonlinear data. The depth is fixed at 28 and the width 16. Preprocessing of data is the same as before. From Table 4, one could see that the maxima and minima of testing losses of the optimal regression model are 4.2117 × 10 4 and 2.550 × 10 5 , respectively. This indicates that the optimal model has small testing losses on these nonlinear datasets. Figure 5 shows the visualized comparisons between regression results of the optimal model and the real value of simulated datasets. For each dataset, 200 samples and the corresponding predictions are compared and plotted in Figure 5. From Table 4 and Figure 5, we could see that the optimal model behaves well in simulated nonlinear datasets.

3.3. Comparisons with Other Approximation Techniques

In this section, we compare the optimal residual model with other linear and nonlinear approximation techniques. Linear techniques include the linear regression, ridge regression, lasso regression, and elastic net regression, which combines ridge regression and lasso regression. Nonlinear techniques include the usual artificial neural network (ANN), decision tree, and support vector regression (SVR) machine. ANN has the same architecture (the same depth and width) as the optimal residual regression model except for the shortcut connections mentioned in Figure 2. ANN is modeled by Keras using tensorflow as a backend. Other regression models are built by the machine learning package scikit-learn in python [42]. A total of 10,000,000 samples generated by Equation (4) are used here. ANN is trained on 6,750,000 samples and validated on 750,000 samples. The remaining 2,500,000 samples are used as testing data. The epoch and patience of early warning for ANN are 50 and 10, respectively. Other models are trained on 7,500,000 samples and tested on the same 2,500,000 samples. Table 5 shows the comparison results including training time, validation loss, and testing loss. There are hyperparameters for every approximation technique excluding linear regression, which cannot be learned through training and must be set in advance. Hence, to get a good approximation, we use the grid search method to find the optimal values. Table 6 shows the information about hyperparameters, including the name, the range, and the corresponding optimal value of every hyperparameter. NA in Table 5 and Table 6 stands for Not Applicable.
From Table 5, one could observe that the testing losses of the linear models listed in the top are approximately 3.72 × 10 2 , which is much greater than the testing losses of the other nonlinear approximation techniques. This indicates that these linear models are not appropriate for nonlinear approximation. It is notable that support vector regression (SVR) machines with radial basis function (RBF) kernel have the maximum training time, which is close to 44 h, and the second greatest testing loss ( 1.2676 × 10 2 ). The testing loss of SVR is also greater than that of residual regression, ANN without shortcuts, and decision tree. These drawbacks of SVR limit its further applications in practice. It is also worthy to mention that the testing loss of decision tree regression is 4.58 times as great as that of the residual regression model. This implies that the residual regression model has better approximation capacities than decision tree. In addition, the influence of shortcut connections is investigated. ANN without shortcuts has the same depth and width as the optimal residual model, and it has the second smallest testing loss. The model is trained for nearly 27 min and early stops at 34 epochs. However, the optimal residual model is trained approximately 23 min through 50 epochs. This means that the residual regression model is more efficient and approximates nonlinear functions better.

3.4. Application of Residual Regression Model on Climate Data

In this paper, we employ the residual regression to approximate relative humidity. Relative humidity is defined as the ratio of the water vapor pressure to the saturated water vapor pressure at a given temperature. It is a key factor affecting cloud microphysics and dynamics, and it plays an important role in climate [43]. The formation of cloud condensation nuclei needs water vapor to be supersaturated in the air. However, currently, there is no widely accepted and reliable method to measure the supersaturated vapor pressure accurately [44], which means that the relative humidity is not accurate under supersaturation circumstance. Therefore, finding the nonlinear relationship between relative humidity and other factors is meaningful.
Our training data is from EAR5 hourly reanalysis datasets on the 1000 hPa pressure level [45]. Pressure, temperature, and specific humidity are used as input features. EAR5 is the fifth generation ECMEF (European Centre for Medium-Range Weather Forecasts) atmospheric reanalysis of the global climate, and it provides hourly outputs at a spatial resolution of 0.25°. The timestamp of the training data is from 00:00:00 to 23:00:00 on September 1, 2007. There are 24,917,709 samples in total. The optimal residual regression model is trained on 20,183,344 samples and validated on 2,242,594 samples. The remaining 2,491,771 samples are testing data. The other parameters are the same. A total of 200 observations and the corresponding predicted data are plotted in Figure 6. The relative error of testing data for relative humidity is 9%. We replace the testing dataset randomly for 10 times, and the averaged relative error is still 9%. This verifies that the residual regression model is stable and applicable in practice.

4. Conclusions

In this paper, we develop deep residual regression models for nonlinear regression. The traditional deep residual learning behaves well in the images process due to local convolution kernels and deep neural networks. However, the convolution kernels have no effects on the whole input sequence. Therefore, it is not suitable for the regression of nonlinear functions. We replace convolutional layers and pooling layers by fully connected layers to ensure that deep residual learning can be applied in nonlinear regression. The residual regression model is carefully and numerically evaluated on simulated nonlinear data, and the results show that the improved regression model works well. It is recommended to avoid setting the residual regression model into a great depth or a small width, since it has a great testing loss under these circumstances. In addition, we compare the residual regression model with other linear and nonlinear approximation techniques. It turns out that the optimal residual regression model has a better approximation capacity compared to others. Finally, the residual regression model is applied into the prediction of relative humidity, and we get a low relative error, which indicates that the residual regression model is stable and applicable in practice. In the future, we intend to apply the residual regression model on large eddy simulation (LES) datasets of turbulence to improve the subgrid-scale parameterizaitons of LES.

5. Patents

There aren’t any patents resulting from the work reported in this manuscript.

Author Contributions

Conceptualization, D.C.; methodology, D.C.; software, D.C., G.N.; validation, T.Y.; formal analysis, T.Y.; investigation, D.C.; resources, G.N., F.H.; data curation, D.C.; writing—original draft preparation, D.C.; writing—review and editing, D.C. and T.Y.; visualization, D.C.; supervision, F.H. and G.N.; project administration, D.C.; funding acquisition, F.H. and G.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by National Natural Science Foundation of China, grant number 11472272.

Acknowledgments

Dongwei gives thanks for the effective discussions with his colleagues in the meteorology group of Unis Co., Ltd. The authors thank all the anonymous reviewers for their constructive comments. The authors also thank all the editors for their careful proofreading. The code of optimal residual regression model could be found at https://github.com/DowellChan/ResNetRegression.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pandey, M.; Zakwan, M.; Sharma, P.K.; Ahmad, Z. Multiple linear regression and genetic algorithm approaches to predict temporal scour depth near circular pier in non-cohesive sediment. Ish. J. Hydraul. Eng. 2020, 26, 96–103. [Google Scholar] [CrossRef]
  2. Zhou, N.; Pierre, J.W.; Trudnowski, D. A stepwise regression method for estimating dominant electromechanical modes. IEEE Trans. Power Syst. 2011, 27, 1051–1059. [Google Scholar] [CrossRef]
  3. Ahn, J.J.; Byun, H.W.; Oh, K.J.; Kim, T.Y. Using ridge regression with genetic algorithm to enhance real estate appraisal forecasting. Expert Syst. Appl. 2012, 39, 8369–8379. [Google Scholar] [CrossRef]
  4. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  5. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  6. Bates, D.M.; Watts, D.G. Nonlinear Regression Analysis and Its Applications, 1st ed.; Wiley: Hoboken, NJ, USA, 1988. [Google Scholar] [CrossRef] [Green Version]
  7. Rhinehart, R.R. Nonlinear Regression Modeling for Engineering Applications: Modeling, Model Validation, and Enabling Design of Experiments, 1st ed.; Wiley: Hoboken, NJ, USA, 2016. [Google Scholar] [CrossRef]
  8. Ritz, C.; Streibig, J.C. Nonlinear Regression with R, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar] [CrossRef]
  9. Motulsky, H.; Christopoulos, A. Fitting Models to Biological Data Using Linear and Nonlinear Regression: A Practical Guide to Curve Fitting, 1st ed.; Oxford University Press: Oxônia, UK, 2004. [Google Scholar]
  10. Majda, A.J.; Harlim, J. Physics constrained nonlinear regression models for time series. Nonlinearity 2012, 26, 201. [Google Scholar] [CrossRef]
  11. Yagiz, S.; Gokceoglu, C. Application of fuzzy inference system and nonlinear regression models for predicting rock brittleness. Expert Syst. Appl. 2010, 37, 2265–2272. [Google Scholar] [CrossRef]
  12. Cobourn, W.G. An enhanced PM2. 5 air quality forecast model based on nonlinear regression and back-trajectory concentrations. Atmos. Environ. 2010, 44, 3015–3023. [Google Scholar] [CrossRef]
  13. Gunay, A. Application of nonlinear regression analysis for ammonium exchange by natural (Bigadic) clinoptilolite. J. Hazard. Mater. 2007, 148.3, 708–713. [Google Scholar] [CrossRef]
  14. Drignei, D.; Forest, C.E.; Nychka, D. Parameter estimation for computationally intensive nonlinear regression with an application to climate modeling. Ann. Appl. Stat. 2008, 2, 1217–1230. [Google Scholar] [CrossRef]
  15. Hipp, L.B. The Weierstrass Approximation Theorem. Master’s Thesis, University of South Carolina, Columbia, SC, USA, 2013. [Google Scholar]
  16. Drucker, H.; Burges, C.C.; Kaufman, L.; Smola, A.J.; Vapnik, V.N. Advances in Neural Information Processing Systems 9, 1st ed.; MIT Press: Cambridge, UK, 1996; pp. 155–161. [Google Scholar]
  17. Smola, A.J.; Schölkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef] [Green Version]
  18. Loh, W.Y. Classification and regression trees. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2011, 1, 14–23. [Google Scholar] [CrossRef]
  19. Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control. Signals. Syst. 1989, 2, 303–314. [Google Scholar] [CrossRef]
  20. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  21. Funahashi, K. On the approximate realization of continuous mappings by neural networks. Neural Netw. 1989, 2, 183–192. [Google Scholar] [CrossRef]
  22. Kurková, V. Kolmogorov’s theorem and multilayer neural networks. Neural Netw. 1992, 5, 501–506. [Google Scholar] [CrossRef]
  23. Huang, G.B.; Lei, C.; Chee, K.S. Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Netw. 2006, 17, 879–892. [Google Scholar] [CrossRef] [Green Version]
  24. Ferrari, S.; Stengel, R.F. Smooth function approximation using neural networks. IEEE Trans. Neural Netw. 2005, 16, 24–38. [Google Scholar] [CrossRef]
  25. Turchetti, C.; Massimo, C.; Paolo, C.; Simone, O. On the approximation of stochastic processes by approximate identity neural networks. IEEE Trans. Neural Netw. 1998, 9, 1069–1085. [Google Scholar] [CrossRef]
  26. Belli, M.R.; Conti, M.; Crippa, P.; Turchetti, C. Artificial neural networks as approximators of stochastic processes. Neural Netw. 1999, 12, 647–658. [Google Scholar] [CrossRef]
  27. Turchetti, C.; Crippa, P.; Pirani, M.; Biagetti, G. Representation of nonlinear random transformations by non-Gaussian stochastic neural networks. IEEE Trans. Neural Netw. 2008, 19, 1033–1060. [Google Scholar] [CrossRef] [PubMed]
  28. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  29. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012. [Google Scholar]
  30. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  31. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 29th IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar] [CrossRef] [Green Version]
  33. Lu, Z.; Pu, H.; Wang, F.; Hu, Z.; Wang, L. The expressive power of neural networks: A view from the width. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  34. Hanin, B.; Sellke, M. Approximating continuous functions by ReLU nets of minimal width. Arxiv Prepr. 2017, arXiv:1710.11278. [Google Scholar]
  35. Lin, H.; Jegelka, S. ResNet with one-neuron hidden layers is a Universal Approximator. In Proceedings of the 32nd Annual Conference on Neural Information Processing Systems, Montréal, QC, Canada, 2–8 December 2018. [Google Scholar]
  36. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 7–9 July 2015. [Google Scholar]
  37. Hara, K.; Saito, D.; Shouno, H. Analysis of function of rectified linear unit used in deep learning. In Proceedings of the International Joint Conference on Neural Networks, Killarney, Ireland, 12–16 July 2015. [Google Scholar] [CrossRef]
  38. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference for Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  39. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  40. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA Neural Netw. Mach. Learn. 2012, 4, 26–31. [Google Scholar]
  41. Ian, G.; Bengio, Y.; Courville, A. Deep Learning, 1st ed.; MIT Press: Cambridge, UK, 2016; pp. 241–249. [Google Scholar]
  42. Homepage of Scikit-Learn. Available online: https://scikit-learn.org/stable/ (accessed on 27 January 2020).
  43. Fan, J.; Zhang, R.; Li, G.; Tao, W. Effects of aerosols and relative humidity on cumulus clouds. J. Geophys. Res. Atmos. 2007, 112, D14204. [Google Scholar] [CrossRef]
  44. Shen, C.; Zhao, C.; Ma, N.; Tao, J.; Zhao, G.; Yu, Y.; Kuang, Y. Method to Estimate Water Vapor Supersaturation in the Ambient Activation Process Using Aerosol and Droplet Measurement Data. J. Geophys. Res. Atmos. 2018, 123, 10606–10619. [Google Scholar] [CrossRef]
  45. Download Page of ERA5 Hourly Data on Pressure Levels from 1979 to Present. Available online: https://cds.climate.copernicus.eu/cdsapp#!/dataset/10.24381/cds.bd0915c6?tab=form (accessed on 27 January 2020).
Figure 1. The shortcut connections of a deep residual neural network (ResNet) for the image process. (a) An identity block, which is employed when the input and output have the same dimensions. (b) A convolution block, which is used when the dimensions are different.
Figure 1. The shortcut connections of a deep residual neural network (ResNet) for the image process. (a) An identity block, which is employed when the input and output have the same dimensions. (b) A convolution block, which is used when the dimensions are different.
Entropy 22 00193 g001
Figure 2. The shortcut connections of ResNet for nonlinear regression. Convolution layers are replaced by dense layers. There are three hidden dense layers in each dense block and identity block. (a) An identity block, which is employed when the input and output have the same dimensions. (b) A dense block, which is used when the dimensions are different.
Figure 2. The shortcut connections of ResNet for nonlinear regression. Convolution layers are replaced by dense layers. There are three hidden dense layers in each dense block and identity block. (a) An identity block, which is employed when the input and output have the same dimensions. (b) A dense block, which is used when the dimensions are different.
Entropy 22 00193 g002
Figure 3. Structure of ResNet regression model. #1 means 1 dense block and #M means M identity blocks. In this paper 1 dense block and 2 identity blocks are stacked repeatedly between dashed lines.
Figure 3. Structure of ResNet regression model. #1 means 1 dense block and #M means M identity blocks. In this paper 1 dense block and 2 identity blocks are stacked repeatedly between dashed lines.
Entropy 22 00193 g003
Figure 4. Datasets for residual regression model. (a), (b), (c) and (d) are simulated data with a nonlinearity order from 1 to 4, respectively.
Figure 4. Datasets for residual regression model. (a), (b), (c) and (d) are simulated data with a nonlinearity order from 1 to 4, respectively.
Entropy 22 00193 g004
Figure 5. The results of optimal regression model on simulated nonlinear data. A red cross symbol (×) means predicted data and a green circle symbol (O) stands for simulated data generated by Equations (2)–(5).
Figure 5. The results of optimal regression model on simulated nonlinear data. A red cross symbol (×) means predicted data and a green circle symbol (O) stands for simulated data generated by Equations (2)–(5).
Entropy 22 00193 g005
Figure 6. The prediction of regression models for relative humidity. The red cross symbol (×) means predicted data, and the green circle symbol (O) stands for observations of EAR5.
Figure 6. The prediction of regression models for relative humidity. The red cross symbol (×) means predicted data, and the green circle symbol (O) stands for observations of EAR5.
Entropy 22 00193 g006
Table 1. Fundamental blocks of ResNet regression.
Table 1. Fundamental blocks of ResNet regression.
Type of BlockInput ShapeOutput ShapeActivation FunctionInput and Output Are of Same Size
Input(1, 6)(1, None *)ReLUFalse
Dense(1, N2)(1, N3)ReLUFalse
Identity(1, N1)(1, N1)ReLUTrue
Output(1, None)(1,1)LinearFalse
* None means the size of input in this dimension is uncertain.
Table 2. Residual regression model with different depths.
Table 2. Residual regression model with different depths.
Depth of ResNetNumber of ParametersStopping Epoch Training TimeTraining Loss (10−4)Validation Loss (10−4)Testing Loss (10−4)
1045815000:30:046.66373.65663.6687
1995813700:23:426.71964.40224.3824
28 *14,5814400:55:345.72213.52053.4855
3719,5815000:59:485.08775.11255.2183
4624,5813401:16:576.32285.37905.3401
5529,5814102:14:365.46764.30134.1437
8244,5814402:27:266.69243.84263.8526
10054,5815002:25:188.266719.00018.943
14579,5813903:46:5815.0002497.02499.5
244134,5815009:09:051273.01277.01279.0
* The bold implies that the depth of ResNet is optimal and has a minimum testing loss.
Table 3. Residual regression model with different widths.
Table 3. Residual regression model with different widths.
Width of ResNetNumber of ParametersStopping Epoch Training TimeTraining Loss (10−4)Validation Loss (10−4)Testing Loss (10−4)
11985000:41:3848.000547.00548.44
411255000:44:3615.00013.00013.255
831455000:44:135.66513.00783.0666
1260615000:20:545.28216.74556.6151
16 *98735000:22:224.75342.12222.1360
2014,5814400:55:345.72213.52053.4855
3030,2715000:54:054.86703.58463.5973
5078,4514200:18:285.52103.90343.9000
70149,0314400:21:205.17934.28864.2837
90242,0115000:22:114.51572.82372.8627
150655,3515001:02:575.43133.45653.4189
3002,570,7014700:45:005.67568.22588.1343
5007,084,5015001:16:356.26784.00933.9767
70013,838,3015001:56:216.86823.35803.4051
* The bold implies that the width of ResNet is optimal and has a minimum testing loss.
Table 4. Optimal regression model on nonlinear datasets.
Table 4. Optimal regression model on nonlinear datasets.
Nonlinearity of DatasetNumber of ParametersStopping EpochTraining Time Training Loss (10−4)Validation Loss (10−4)Testing Loss (10−4)
198734500:46:151.06190.25520.2550
298733500:36:098.25664.20144.2117
398735000:22:224.75342.12222.1360
498735000:57:263.65422.04392.0481
Table 5. Comparisons of regression techniques.
Table 5. Comparisons of regression techniques.
Regression Techniques UsedStopping EpochTraining TimeTraining Loss (10−4)Validation Loss (10−4)Testing Loss (10−4)
Linear regressionNA00:00:20371.61NA371.68
Ridge regressionNA00:00:33371.64NA371.59
Lasso regressionNA00:00:39371.51NA371.97
Elastic regressionNA00:34:48371.51NA372.00
Residual regression *5000:22:224.75342.12222.1360
ANN without shortcuts3400:27:077.2640 6.63796.6537
Decision tree regressionNA00:17:069.4083NA9.7777
Support vector regressionNA43:51:36126.75NA126.76
* The bold implies that the residual regression model has a minimum testing loss.
Table 6. Optimal hyperparameters of regression techniques.
Table 6. Optimal hyperparameters of regression techniques.
Regression Techniques UsedName of HyperparametersRange of HyperparametersOptimal Hyperparameters
Linear regressionNANANA
Ridge regressionPenalty parameter of L 2 norm α ;10−10, 109, 108
…, 109, 1010;
102
Lasso regressionPenalty parameter of L 1 norm α ;1010, 109, 108
…, 109, 1010;
105
Elastic regressionPenalty parameter α ; Ratio of L 1 norm ρ ;1010, 109,…, 1010; 0.0, 0.1,…,0.9, 1.0;104; 1.0;
Residual regressionWidth; Depth;NA16; 28;
ANN without shortcutsWidth; Depth;NA16; 28;
Decision tree regressionMaximum depth;1, 2,…,9, 10;10
Support vector regressionPenalty parameter α ;
RBF kernel parameter γ ;
Epsilon-tube parameter ε ;
Maximum iteration N ;
102, 103,…, 107;
10−4, 10−3,…, 101;
0.1;
5000;
102;
10;
0.1;
5000;

Share and Cite

MDPI and ACS Style

Chen, D.; Hu, F.; Nian, G.; Yang, T. Deep Residual Learning for Nonlinear Regression. Entropy 2020, 22, 193. https://doi.org/10.3390/e22020193

AMA Style

Chen D, Hu F, Nian G, Yang T. Deep Residual Learning for Nonlinear Regression. Entropy. 2020; 22(2):193. https://doi.org/10.3390/e22020193

Chicago/Turabian Style

Chen, Dongwei, Fei Hu, Guokui Nian, and Tiantian Yang. 2020. "Deep Residual Learning for Nonlinear Regression" Entropy 22, no. 2: 193. https://doi.org/10.3390/e22020193

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop