Abstract
9–12% Cr martensitic steels are widely used for critical components of new, high-efficiency, ultra-supercritical power plants because of their high creep and oxidation resistances. Due to the time consuming effort of obtaining creep properties for new alloys under high temperature creep conditions, in both short-term and long-term testing, it is often dealt with simplified models to assess and predict the future behavior of some materials. In this work, the total time to produce the material fracture is predicted according to models obtained using several linear techniques, since this property is really relevant in power plants elements. These models are obtained based on 344 creep tests performed on modified P92 steels. A multivariate analysis and a feature selection were applied to analyze the influence of each feature in the problem, to reduce the number of features simplifying the model and to improve the accuracy of the model. Later, a training-testing validation methodology was performed to obtain more useful results based on a better generalization to cover every scenario of the problem. Following this method, linear regression algorithms, simple and generalized, with and without enhanced by gradient boosting techniques, were applied to build several linear models, achieving low errors of approximately 6.75%. And finally, among them the most accurate model was selected, in this case the one based on the generalized linear regression technique.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Steel mills are highly involved in improving mechanical and creep resistance under conditions of high temperatures and long service times in the materials used to produce steels, among them martensitic steels of high percentage in chromium [1, 2]. Since long service times under creep conditions will produce creep damage in steel used to steam turbine in power plants, this improvement can enhance the efficiency of these plants. Optimizing of creep damage evolution can improve mechanical properties until fracture, prolonging the service life of equipment operating [3,4,5].
The purpose of this work is to generate knowledge that allows a better understanding of the different elements influence in the alloy and in the metallurgical mechanisms, allowing to improve martensitic steels materials creep properties. This knowledge will allow easier development of new advanced steels.
Specifically, steels with a content of 9–12% Cr are studied, focusing this study to improve creep resistance [6]. In this approach, several models will be developed to predict the short-long-term creep behavior of new steels based on previous creep behaviors of similar materials.
These models use a previous knowledge of the influence of each composition element of the alloy and the thermal treatment on the fracture time, applied to predict this life expectancy. These models have a direct application in the high-chromium steels studied in this work, but it is also useful for a wide range of steels such as the rest of stainless steels, microalloys, high strength and new generation steels.
Several techniques of linear regression were applied to predict material fracture life, specifically 9–12% Cr martensitic steels based on modified P92 steels. Also, to obtain a clear knowledge of the influence of each feature that define the composition of the material, the thermal pretreatment applied to the material, and the creep test conditions on the final material fracture life. In order to get this purpose, a study of the literature [7,8,9,10,11,12,13,14,15,16,17,18,19] focus on creep test of Cr martensitic steels were performed, obtaining a dataset with 344 instances that was used to train and test the models. Then, several techniques like data pre-processing, outlier detection, analysis of variance, analysis of covariance, analysis of correlation, multivariate data visualization, and principal components analysis were applied. This step studies the influence of each attribute with the output, the material fracture life, to finally select the most significant variables to perform the most accurate models to solve the proposed problem. And then several linear regression techniques were used and validated to build and select an accurate linear model that improves the amount of knowledge of the creep behavior of 9–12% Cr martensitic steels.
Linear regression algorithms, simple (simple linear regression) and generalized (generalized linear regression), with and without using enhanced by gradient boosting techniques (enhanced linear regression and enhanced generalized linear regression) were applied in this work.
These kind of methodologies that applies data analysis techniques and machine learning algorithms is gaining interest in industrial engineering fields [20,21,22] where a mathematical model that represents the problem can be helpful in the decision making. And within of the possible engineering problems, this methodology can be used for the development of new materials (advanced modified P92 steels) improving final properties and reducing time and money during the process.
2 Materials
In this work, short-term and long-term testing were studied. From this study, 344 instances based on these testing were contemplated based on significant information of the composition of the material, the previous thermal treatment performed on the material, and the parameters that control the creep test. More specifically the studied features were:
-
Composition of the material: weight percent (wt%) of the most applied elements of the chemical composition of modified P92 type steel according to ASTM A335 (C, Mn, Si, Cr, Ni, Mo, P, S, W, Nb, V, N, Al, B, Co, Cu).
-
Heat treatment of this steel before the creep test: temperature (NormTemp) and time (NormTime) for normalizing, and temperature (TempTemp) and time (TempTime) for tempering.
-
Control parameters of creep test: Temperature (Temp) and strain (Strain) applied on creep test performance.
-
Time to fracture (TimeFractPoint), that is the predicted value based on the rest of features.
Figure 1 shows the relationships between the variables under study.
3 Methodology
Based on the high number of variables and the complexity of the studied problem (prediction of fracture time of P92 steels or similar materials), the proposed methodology is applied in two steps, first a multivariate analysis and later a linear regression. With the final goal of predicting the material fracture time, but with an easy to understand model that, at the same time, could show information of the influence of the independent variables on the fracture time.
3.1 Multivariate Analysis
A multivariate analysis was performed to simplify the final models, building simpler models and as result, making a better generalization of the problem to improve the results in the validation stage. This analysis was conducted in five steps in order to detect wrong instances and redundant features:
-
Data pre-processing.
-
Outlier detection [23].
-
Analysis of correlation, variance and covariance [24].
-
Multivariate data visualization.
3.2 Simple Linear Regression
A probabilistic model of the expected value of an output variable is built based on several independent variables by means of a simple multivariate linear regression (LR) technique [27, 28] based on Eq. 1.
Where \( \upeta \) is the dependent variable; \( {\text{x}}_{\text{i}} \) are the independent variables; \( \upbeta_{\text{i}} \) are the weight of each independent variable and indicates the influence of these variables on the dependent variable; and finally \( \upvarepsilon \) is the bias based on a Gaussian distribution.
In this case, fracture time of the material is predicted based on 16 features that define the composition of the material, 4 features that define the steel heat treatment, and 2 features that define the creep test control.
3.3 Enhanced Linear Regression
Models based on simple linear regression algorithms can be improved applying gradient boosting [29, 30], combining the results of several simple linear regression models according to a cost function. This technique is call enhanced linear regression (LRB) and calculates and uses the obtained residuals from one model evaluation to train another model, that together make decrease the error according to a squared error loss function. This process is repeated until the training convergence is achieved [31]. The training stage tunes two significant parameters of the algorithm: the number of repetitions (mstop) and the shrinkage factor (shrink), to reduce a possible overtraining.
3.4 Generalized Linear Regression
Another applied technique is the generalized linear regression (GLM), which is based in a flexible generalization of the simple linear regression. In this technique, a linking function allows the optimization of the regression reducing the residual error but assuming different distributions [31,32,33,34,35,36]. Beforehand, the assumption of different distributions can built more accurate models in a not linear problem, like the case of short and long-term creep tests.
The application of the generalized regression algorithm is based on three components: a prediction based on a simple linear regression (Eq. 1); a linking function that is reversible and transforms the variables according to Eq. 2; and a random component based on the chosen distribution (Gaussian, binomial, gamma, etc.), that based on the independent features, specifies the possible distribution of the dependent feature, η.
To apply this technique, it was assumed that the distribution of the original features of the dataset studied in this work were Gaussian, and that subsequently the three link functions shown in Table 1 were applied.
3.5 Enhanced Generalized Linear Regression
As in the enhanced simple linear regression, models based on generalized linear regression technique can be improved using boosting techniques. In this case, L2 boosting technique [37, 38] was applied to maximize the accuracy minimizing the descending gradient function error. The boosted generalized linear regression algorithm (GLRB) [39,40,41,42] obtains the residuals in an iterative way according to a squared error loss function until the training convergence is achieved. Also, the training stage tunes two significant parameters of the algorithm: the number of repetitions (mstop) and the number of variables (based on the Akaike criterion), to reduce a possible overtraining and to simplify the model.
3.6 Validation Method
A validation method to obtain comparable results from the built models was performed. This method started with a normalization between 0 and 1 of the instances that formed the dataset. The benefits of this normalization were two: first that the range of the weight that each feature had in the output variable were equal in such a way that the impact of the features in the prediction could be observed in an easy way; and second that the final accuracy could be improved. Later, a training-testing simple validation method was conducted, where an 80% of the original dataset was randomly selected to build and train the models and the remaining 20% of the original dataset were used to test and validate the models. In the training stage, trying to avoid overtraining, 10-fold cross validation was performed. This cross validation allows having a more trustful error to make a preliminary selection of the obtained models from the parameters tuning. In this way, the most accurate models during training stage were selected to be tested and validated with the testing dataset. And finally based on these last results, the model with lower error in the testing stage was selected.
The work conducted in this study was performed using the statistical software tool R x64 v3.4.1 [43].
3.7 Accuracy Criteria
Models accuracy must be measured in order to evaluate the performance of the model prediction. There are several criteria to measure this accuracy when models predict a numeric feature. Some of the most used criteria are computational validation errors, that calculate the relation between the values predicted by the model and the real values measured during the performance of creep tests. In this case, the following criteria were applied:
-
Mean absolute error (MAE) (Eq. 3)
$$ {\text{MAE}} = \frac{1}{n}\mathop \sum \limits_{k = 1}^{d} \left| {m_{k} - p_{k} } \right| $$(3) -
Root mean squared error (RMSE) (Eq. 4)
$$ {\text{RMSE}} = \sqrt {\frac{1}{\text{n}}\mathop \sum \limits_{{{\text{k}} = 1}}^{\text{d}} \left( {{\text{m}}_{\text{k}} - {\text{p}}_{\text{k}} } \right)^{2} } $$(4) -
Correlation coefficient (CORR) (Eq. 5)
$$ {\text{CORR}} = \frac{{\mathop \sum \nolimits_{{{\text{k}} = 1}}^{\text{d}} \frac{{\left( {{\text{p}}_{\text{k}} - {\overline{\text{p}}}} \right) \cdot \left( {{\text{m}}_{\text{k}} - {\overline{\text{m}}}} \right)}}{{{\text{n}} - 1}}}}{{\sqrt {\mathop \sum \nolimits_{{{\text{k}} = 1}}^{\text{d}} \frac{{\left( {{\text{p}}_{\text{k}} - {\overline{\text{p}}}} \right)^{2} }}{{{\text{n}} - 1}}} \cdot \mathop \sum \nolimits_{{{\text{k}} = 1}}^{\text{d}} \frac{{\left( {{\text{m}}_{\text{k}} - {\overline{\text{m}}}} \right)^{2} }}{{{\text{n}} - 1}}}} $$(5)
Where \( {\text{d}} \) is the number of instances used from the database to validate the model, \( {\text{m}} \) are the real values, \( {\text{p}} \) are the predicted values, \( \overline{\text{m}} = \frac{1}{\text{n}}\mathop \sum \nolimits_{{{\text{k}} = 1}}^{\text{d}} {{\rm m}}_{\text{k}} \) and \( \overline{\text{p}} = \frac{1}{\text{d}}\mathop \sum \nolimits_{{{\text{k}} = 1}}^{\text{d}} {{\rm p}}_{\text{k}} \).
4 Results and Discussion
The dataset of 9–12% Cr martensitic steels based on modified P92 standard specification was formed from a set of creep experiments obtaining from the related literature. The studied features and ranges were selected according to the proposed in the standard specification of the rule. Previously to apply the validation method, a multivariate analysis was performed on the dataset. Then, based on the results of the multivariate analysis the final dataset was set, and like it was commented previously the dataset was split in two new datasets, the training dataset and the testing dataset. Using the training dataset, the models were built applying 10-fold cross validation process combined with a tuning of the most significant parameters of each algorithm. This process allows to optimize the final accuracy that subsequently will be tested and validate using the testing dataset. One example of the results obtained during the training process, using cross validation and parameter tuning is shown in Fig. 2.
After the analysis of this parameter tuning, the final results obtained during the training stage are listed in Table 2.
After the selection of the parameter values to be used in each algorithm during the training stage to improve the accuracy criteria, the models were built. And then were tested with the 20% of the instances that were split previously from the initial dataset and were never used on the training stage. The validation with this testing dataset gives a better idea of the generalization of the problem since these instances were not used to built the models. The results obtained during the testing and validation stages are shown in Table 2.
In this case, the model that gets the most accurate results during the training stage gets the most accurate results during the testing stage. That means that a possible overfitting during the training stage is avoided using 10-fold cross validation, and also that the results obtaining during the validation indicate that the model based on GLM algorithm has an accurate generalization of the problem. That shows that the prediction of the fracture time on modified P92 steels is reliable according to the obtained error.
Then, the model based on the GLM algorithm gets the best performance during all the proposed stages. For this reason, the linear regression model that is defined in Eq. 6 can be considered like an accurate predictor of the material fracture life expectancy, with a RMSE of 6.75% and a correlation of 59.91% based on the testing stage results.
5 Conclusions
The results show that linear modeling techniques can predict some material properties, such as fracture time in creep conditions, with high accuracy. And not only a good prediction can be obtained, also and based on the join of multivariate data analysis techniques and linear model prediction a hidden knowledge of the process can be obtained. In this case, obtained models predict fracture time of 9–12% Cr martensitic steels with low errors of approximately a RMSE of 6%. Also the most influential variables in the process have been detected, and their weight in the process was determined. Therefore, it can be concluded that these techniques are useful in industrial problems, such as the one presented, and of significant help for developers of new materials in order to improve the final properties of the product.
References
Sachadel, U.A., Morris, P.F., Clarke, P.D.: Design of 10% Cr martensitic steels for improved creep resistance in power plant applications. J. Mater. Sci. Technol. 29(7), 767–774 (2013)
Morris, P.F., Sachadel, U.A., Clarke, P.D.: Design of heat treatments for 9–12% Cr steels to optimise creep resistance for power plant applications. In: Proceedings of 9th Liège Conference on Materials for Advanced Power Engineering, Liège, Belgium, pp 554–564 (2010)
Mayer, K.H., Bendick, W., Husemann, R.V., Kern, T., Scarlin, R.B.: International Joint Power Generation Conference, PWR, vol. 33, pp. 831–841. ASME, New York (1998)
Gold, M., Jaffee, R.I.: Materials for advanced steam cycles. J. Mater. Energy Syst. 6(2), 130–145 (1984)
Viswanathan, R., Bakker, W.: Materials for ultrasupercritical coal power plants - boiler materials: Part 1. J. Mater. Eng. Perform. 10(1), 81–95 (2001)
Lanin, A.A., Grin, E.A.: An approach to assessment of the lifetime characteristics of steels under creep conditions using fracture mechanics criteria. Therm. Eng. 65(4), 239–245 (2018)
Hald, J.: Prospects for martensitic 12% Cr steels for advanced steam power plants. Trans. Indian Inst. Met. 69(2), 183–188 (2016)
Kimura, K., Kushima, H., Sawada, K.: Long-term creep deformation property of modified 9Cr–1Mo steel. Mater. Sci. Eng., A 510, 58–63 (2009)
Sklenička, V., Kuchařová, K., Svoboda, M., Kloc, L., Buršık, J., Kroupa, A.: Long-term creep behavior of 9–12%Cr power plant steels. Mater. Charact. 51(1), 35–48 (2003)
Fujita, T., Asakura, K., Sawada, T., Takamatsu, T., Otoguro, Y.: Creep rupture strength and microstructure of Low C-10Cr-2Mo heat-resisting steels with V and Nb. Metall. Trans. A 12(6), 1071–1079 (1981)
Liu, Y., Tsukamoto, S., Sawada, K., Abe, F.: Role of boundary strengthening on prevention of type IV failure in high cr ferritic heat-resistant steels. Metall. Mater. Trans. A 45(3), 1306–1314 (2014)
Mishnev, R., Dudova, N., Kaibyshev, R.: On the origin of the superior long-term creep resistance of a 10% Cr steel. Mater. Sci. Eng., A 713, 161–173 (2018)
Abe, F.: Creep behavior, deformation mechanisms, and creep life of Mod.9Cr-1Mo steel. Metall. Mater. Trans. Phys. Metall. Mater. Sci. 46(12), 5610–5625 (2015)
Aghajani, A., Somsen, Ch., Eggeler, G.: On the effect of long-term creep on the microstructure of a 12% chromium tempered martensite ferritic steel. Acta Mater. 57(17), 5093–5106 (2009)
Tamura, M., Kumagai, T., Miura, N., Kondo, Y., Shinozuka, K., Esaka, H.: Effect of martensitizing temperature on creep strength of modified 9Cr steel. Mater. Trans. 52(4), 691–698 (2011)
Sawada, K.: Effect of W on recovery of lath structure during creep of high chromium martensitic steels. Mater. Sci. Eng., A 267(1), 19–25 (1999)
Sklenicka, V., Kucharova, K., Svobodova, M., Kral, P., Kvapilova, M., Dvorak, J.: The effect of a prior short-term ageing on mechanical and creep properties of P92 steel. Mater. Charact. 136, 388–397 (2018)
Haney, E.M., et al.: Macroscopic results of long-term creep on a modified 9Cr–1Mo steel (T91). Mater. Sci. Eng., A 510–511, 99–103 (2009)
Fedoseeva, A., Dudova, N., Kaibyshev, R.: Creep strength breakdown and microstructure evolution in a 3% Co modified P92 steel. Mater. Sci. Eng., A 654, 1–12 (2016)
Fernandez Martinez, R., Iturrondobeitia, M., Ibarretxe, J., Guraya, T.: Methodology to classify the shape of reinforcement fillers: optimization, evaluation, comparison, and selection of models. J. Mater. Sci. 52(1), 569–580 (2017)
Fernandez Martinez, R., Lostado Lorza, R., Santos Delgado, A.A., Piedra Pullaguari, N.O.: Optimizing presetting attributes by softcomputing techniques to improve tapered roller bearings working conditions. Adv. Eng. Softw. 123, 13–24 (2018). https://doi.org/10.1016/j.advengsoft.2018.05.005
Fernandez Martinez, R., Jimbert, P., Ibarretxe, J., Iturrondobeitia, M.: Use of support vector machines, neural networks and genetic algorithms to characterize rubber blends by means of the classification of the carbon black particles used as reinforcing agent. Soft. Comput. (2018). https://doi.org/10.1007/s00500-018-3262-2
Hair, J.F., Black, W.C., Babin, B.J., Anderson, R.E.: Multivariate Data Analysis, 7th edn. Pearson, Upper Saddle River (2010)
Fernandez Martinez, R., Martinez-de-Pison Ascacibar, F.J., Pernia Espinoza, A.V., Lostado Lorza, R.: Predictive modeling in grape berry weight during maturation process: comparison of data mining, statistical and artificial intelligence techniques. Span. J. Agric. Res. 9(4), 1156–1167 (2011)
Fernandez Martinez, R., Okariz, A., Ibarretxe, J., Iturrondobeitia, M., Guraya, T.: Use of decision tree models based on evolutionary algorithms for the morphological classification of reinforcing nano-particle aggregates. Comput. Mater. Sci. 92, 102–113 (2014)
Jolliffe, I.T.: Principal Component Analysis, 2nd edn. Springer, New York (2002). https://doi.org/10.1007/b98835
Wilkinson, G.N., Rogers, C.E.: Symbolic descriptions of factorial models for analysis of variance. Appl. Stat. 22, 392–399 (1973)
Chambers, J.M.: In: Chambers, J.M., Hastie, T.J. (eds.) Statistical Models in S. Wadsworth & Brooks/Cole (1992)
Friedman, J.H.: Greedy function approximation: a gradient boosting machine. Technical report, Department of Statistics, Sequoia Hall, Stanford University, Stanford California 94305 (1999)
Friedman, J.H.: Stochastic gradient boosting. Technical report, Department of Statistics, Sequoia Hall, Stanford University, Stanford California 94305 (1999)
Wang, Z.: HingeBoost: ROC-based boost for classification and variable selection. Int. J. Biostat. 7(1), 13 (2011)
McCullagh, P., Nelder, J.A.: Generalized Linear Models. Chapman and Hall, London (1989)
Dobson, A.J.: An Introduction to Generalized Linear Models. Chapman and Hall, London (1990)
Hasti, T. J., Pregibon, D.: In: Chambers, J.M., Hastie, T.J (eds.) Statistical Models in S. Wadsworth & Brooks/Cole (1992)
Venables, W.N., Ripley, B.D.: Modern Applied Statistics with S. Springer, New York (2002). https://doi.org/10.1007/978-0-387-21706-2
Fox, J.: Applied Regression Analysis and Generalized Linear Models, 3rd edn. McMaster University, SAGE Publications, Inc, Los Angeles (2015)
Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Proceedings of 13th International Conference on Machine Learning, San Francisco, CA, pp. 148–156 (1996)
Buehlmann, P.: Boosting for high-dimensional linear models. Ann. Stat. 34, 559–583 (2006)
Buehlmann, P., Yu, B.: Boosting with the L2 loss: regression and classification. J. Am. Stat. Assoc. 98, 324–339 (2003)
Buehlmann, P., Hothorn, T.: Boosting algorithms: regularization, prediction and model fitting. Stat. Sci. 22(4), 477–505 (2007)
Hothorn, T., Buehlmann, P., Kneib, T., Schmid, M., Hofner, B.: Model-based boosting 2.0. J. Mach. Learn. Res. 11, 2109–2113 (2010)
Hofner, B., Mayr, A., Robinzonov, N., Schmid, M.: Model-based boosting in R: a hands-on tutorial using the R package mboost. Comput. Stat. 29(1–2), 3–35 (2014)
R Development Core Team: R: a language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria (2017). http://www.R-project.org/
Acknowledgment
The authors wish to thanks to the Basque Government through the KK-2018/00074 METALCRO.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Fernandez Martinez, R., Jimbert, P., Barbero, J.I., Callejo, L.M., Somocueto, I. (2019). Material Fracture Life Prediction Using Linear Regression Techniques Under High Temperature Creep Conditions. In: Nyström, I., Hernández Heredia, Y., Milián Núñez, V. (eds) Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications. CIARP 2019. Lecture Notes in Computer Science(), vol 11896. Springer, Cham. https://doi.org/10.1007/978-3-030-33904-3_50
Download citation
DOI: https://doi.org/10.1007/978-3-030-33904-3_50
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-33903-6
Online ISBN: 978-3-030-33904-3
eBook Packages: Computer ScienceComputer Science (R0)