Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

LSTM input timestep optimization using simulated annealing for wind power predictions

  • Muhammad Muneeb

    Roles Conceptualization, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    muneebsiddique007@gmail.com

    Affiliation Department of Electrical Engineering and Computer Science, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates

Abstract

Wind energy is one of the renewable energy sources like solar energy, and accurate wind power prediction can help countries deploy wind farms at particular locations yielding more electricity. For any prediction problem, determining the optimal time step (lookback) information is of primary importance, and using information from previous timesteps can improve the prediction scores. This article uses simulated annealing to find an optimal time step for wind power prediction. Finding an optimal timestep is computationally expensive and may require brute-forcing to evaluate the deep learning model at each time. This article uses simulated annealing to find an optimal time step for wind power prediction. The computation time was reduced from 166 hours to 3 hours to find an optimal time step for wind power prediction with a simulated annealing-based approach. We tested the proposed approach on three different wind farms with a training set of 50%, a validation set of 25%, and a test set of 25%, yielding MSE of 0.0059, 0.0074, and 0.010 for each wind farm. The article presents the results in detail, not just the mean square root error.

Introduction

Countries are moving towards renewable energy sources [13] due to the recent increase in global warming, and sources like solar energy and wind energy can play a crucial role in reducing carbon dioxide emissions in the environment. Wind power is one energy source that can help generate free electricity without pollutants. Wind power depends on the location where windmills are installed and on wind speed and direction, and in these articles [47], researchers illustrated the features which can be used for wind power prediction.

This article focuses on improving the performance of a deep learning algorithm for wind power prediction and reducing computational time. Previous timesteps are used to predict the current timestep for any prediction problem [8, 9]. For instance, information from timestep 1,2 and 3 can be combined with timestep 4 to predict the wind power for timestep 4, but finding the optimal lookback is difficult because we do not know how much past information should be included to make the current prediction. There are many ways to find that information. For instance, we can use the brute-force technique from lookback = 1 to 500 and run LSTM for each lookback. There are several issues associated with this technique. Firstly, as we increase the lookback variable, the data size grows exponentially, resulting in an exponential increase in training time. Secondly, we do not know how long we should train the LSTM model to obtain optimal results [1012]. Thirdly, the optimal time step for the training data may not be optimal for the test set. These issues raise concerns for a better approach in which we can find an optimal lookback or how many previous timestep information should be used to predict the current timestep, and for how long we should train the machine learning model to yield an optimal performance (formally known as a number of epochs).

In this paper, we used a simulated annealing-based [13] + LSTM approach for wind power prediction, allowing us to find an optimal lookback in a limited number of epochs resulting in reduced training time.

The paper focuses on wind power prediction, but it is important to notice that it can also be used for other predictions. The previous timestep inclusion to predict the current timestep can significantly improve the performance. Optimization algorithms like particle swarm optimization [14], genetic algorithm, and hill-climbing can be used to find an optimal lookback. We used simulated annealing, which improves the current results generated using a genetic algorithm.

The important point to notice here is that the contribution of this article is not just using the simulated annealing with LSTM for wind power prediction but also how we used it to find an optimal look back, and it involves running some other computational steps which are documented in the methodology.

In these papers [1517], researchers compared wind power prediction based on physical, statistical, and hybrid methods over different time scales.

There are many algorithms for wind power prediction like auto-regression moving average model [18], LSTM with particle swarm optimization [19], LSTM with extrapolation capability [20], double-stage hierarchical approach for energy management [21], and Cluster-Based Ensemble Regression models [22].

Following are the scientific contribution of this research work.

  • We integrated simulated annealing (optimization algorithm) and LSTM (time series forecasting algorithm) to find an optimal lookback to reduce the computation time, find an optimal lookback, and improve the forecasting performance.
  • The proposed integration is also valid for other deep learning algorithms like BILSTM, GRU, and machine learning-based regression algorithms with minor modifications. It is also one of the future directions that we considered for optimization.

Materials and methods

This section explains the preprocessing performed on the dataset, the difference between three LSTM models, and the integration of simulated annealing with LSTM for wind power prediction.

Dataset division

The dataset we used for the analysis is available on this link. There are about 16 features and one power variable to predict.

The training set is 50%, the test set is 25%, and the validation is 25%. The next step is to transform the dataset based on the lookback parameter, which shows how many previous time step information is to be included to make the current prediction. The transformation plays a key role in the whole process, and it can drastically increase the size of the dataset and computation time. To understand the dataset transformation step refer to Figs 13. In general, we do not know how many previous time step information should be included for the current time prediction, so an efficient approach like simulated annealing is required to find that optimal lookback. Consider Fig 1 each row represents the feature values at each time step. In Fig 2 the data is transformed to 4 time step. In Fig 3 data is transfored to 3 time step.

thumbnail
Fig 1. This figure illustrates the form of the dataset for lookback = 1.

https://doi.org/10.1371/journal.pone.0275649.g001

thumbnail
Fig 2. This figure illustrates the form of the dataset for lookback = 2.

https://doi.org/10.1371/journal.pone.0275649.g002

thumbnail
Fig 3. This figure illustrates the form of the dataset for lookback = 3.

https://doi.org/10.1371/journal.pone.0275649.g003

The transformation step for any number of lookbacks is explained in the code. What follows describes the simulated annealing and LSTM working.

Simulated annealing

Simulated Annealing (SA) is an optimization technique for locating global optima. Simulated annealing employs the objective function of an optimization problem, which in our case is the MSE. The root mean square or explained variance can also be used as an optimization function, but we used the mean square error.

The method works similarly to a hill-climbing algorithm; instead of only choosing the optimal step, it chooses a random move. It constantly changes the current solution if the chosen move improves the solution. Otherwise, the procedure will proceed with a probability of fewer than one. With the “badness” of the maneuver, the chance of changing the current solution falls exponentially.

This likelihood of changing the current solution is also determined by the parameter T (Temperature). Uphill movements are more common with higher T values. Fig 4 shows the simulated annealing algorithm.

Table 1 shows the simulated annealing parameters.

thumbnail
Table 1. This table shows the hyper-parameters for simulated annealing.

https://doi.org/10.1371/journal.pone.0275649.t001

This work extends our previously proposed algorithm, which used a genetic algorithm with LSTM for time step optimization. In that algorithm, we optimized the lookback and the number of neurons in each layer, which took a lot of time. The focus of that work was on improving the performance but not reducing the computation time, so we decided to use simulated annealing to improve the performance (lookback) and reduce the computation time. The success of a genetic algorithm depends on the number of generations and the number of instances in each generation. LSTM has to be executed 100 times for ten generations and ten instances, whereas the simulated annealing requires only a small number of iterations to find an optimal lookback.

LSTM

LSTM is a powerful timeseries prediction algorithm used for genetics [23], windpower prediction [24, 25], text processing [26], and human action prediction.

The LSTM comprises three sections, as illustrated in the diagram Fig 5, each of which serves a different function. The first component determines whether the last timestamp’s information should be remembered or is irrelevant and can be ignored. The cell attempts to learn new information from the input in the second section. Finally, the cell sends updated information from the current timestamp to the next timestamp in the third component. The gates are the three components of an LSTM cell. The Forget gate is the first component, the Input gate is the second, and the Output gate is the third.

Fig 5 shows the architecture of an LSTM. Eqs 13 represents the functions in LSTM cell. F, C, I, and O are the forget, Candidate, Input, and Output gates. (1) (2) (3)

Table 2 shows the LSTM architecture for the deep learning model.

thumbnail
Table 2. This table shows the hyper-parameters for machine learning.

https://doi.org/10.1371/journal.pone.0275649.t002

Simulated annealing objective function

This section discusses and explains the objective function of simulated annealing.

The objective of simulated annealing is to reduce the validation mean square error (See Eq 4), which is calculated using y_validation_data and predicted_values = (LSTM.predict(X_validation_data)). (4)

In Eq 4, n represents the number of samples in the validation set, x represents the predicted samples, y represents the actual samples, and i represents the ith instance.

The following text unpacked the objective function, which minimizes the mean square error MSE.

  1. minimize(MSE)
  2. minimize(objectivefunction(lookback))—The objective function takes lookback as an input.
  3. minimize(objectivefunction(LSTMtrainingfunction(lookback)))—The objective function takes lookback as an input, trains the LSTM model, and returns the MSE.

The computation cost of the proposed algorithm

The following calculation shows the time for simple LSTM when lookback is increased from 1 to N.

Above mentioned equations show the training time for LSTM for lookback 1 to N, and the total time is shown in Eq 5. (5)

The following calculation shows the time when simulated annealing is used to find an optimal lookback. Consider the worst-case scenario in which N is selected for 20 iterations. (6)

Compare Eq 5 with Eq 6. The computation time of Eq 6 (when simulated annealing is used with LSTM for prediction) is far less when lookback is increased from 1 to N. If the number of iterations = N, the cost for both approaches becomes the same.

Stress analysis on lookback

When simulated annealing is used to find the optimal lookback, we specify a particular range in which simulated annealing should look for the next lookback. Increasing lookback directly affects the dataset’s size because this increase results in data replication. Consider the following calculation to understand the size of the dataset in memory as the lookback is increased linearly.

Eq 7 shows the relationship between the size of the dataset and lookback. (7)

  • For lookback = 0; total size = (10 − 0)(1 + 0) = 10
  • For lookback = 1; total size = (10 − 1)(1 + 1) = 18
  • For lookback = 2; total size = (10 − 2)(1 + 2) = 24
  • For lookback = 3; total size = (10 − 3)(1 + 3) = 28
  • For lookback = 4; total size = (10 − 4)(1 + 4) = 30
  • For lookback = 5; total size = (10 − 5)(1 + 5) = 30
  • For lookback = 6; total size = (10 − 6)(1 + 6) = 28
  • For lookback = 7; total size = (10 − 7)(1 + 7) = 24
  • For lookback = 8; total size = (10 − 8)(1 + 8) = 18
  • For lookback = 9; total size = (10 − 9)(1 + 9) = 10

If we increase the lookback, the size of the dataset increase, so we cannot consider all the lookbacks when using simulated annealing. There must be a specific bound on lookback; otherwise, the system memory would not be able to handle it.

Results

This section compares each method’s computation time and performance: Naïve LSTM (A simple LSTM model with lookback = 1 and epochs = 200), Simple LSTM (A simple LSTM model with lookback = 1-500 and epochs = 30), and LSTM with Simulated Annealing (LSTM model with 20 iterations of simulated annealing, 30 epochs, and lookback = 1-500). We observed that LSTM’s performance for windpower prediction could be increased by mutating two parameters: the first one is the number of epochs (the number of times that the learning algorithm will work through the entire training), and the second is the lookback (Number of previous time steps to predict the current time step or the window size in terms of LSTM). For Naïve LSTM we considered 500 epochs and loopback = 1 it took 134 seconds yield MSE of about 0.25 for windfarm 1. It was executed quickly, but the difference between the predicted values and actual values was very high, as shown in Table 3.

thumbnail
Table 3. This table shows the result for LSTM with epochs = 500 and lookback = 1.

https://doi.org/10.1371/journal.pone.0275649.t003

For Simple LSTM we considered epochs = 30 and lookback = [1-500] to find an optimal solution. We considered lookback from 1 to 500 because we do not know where the optimal lookback exists, and to find an optimal lookback iteration, an overall lookback is required. This brute force technique works, but the computation time increases exponentially. The time for epochs = 200 and lookback = 1-500 is 166 hours as shown in Fig 6, and it is extremely computationally expensive. Moreover, the size of the dataset is increased as we increase the lookback, as shown in Fig 7. The difference (MSE, RMSE, r2_score, explained variance, and MAE) between the actual values and predicted values are shown in Fig 8.

thumbnail
Fig 6. This figure shows the computation time in seconds for lookback 1 to 500 and 30 epochs.

https://doi.org/10.1371/journal.pone.0275649.g006

thumbnail
Fig 7. This diagram shows the relationship between data size and lookback.

As we increase the lookback, the size of the data increases exponentially.

https://doi.org/10.1371/journal.pone.0275649.g007

thumbnail
Fig 8. This figure shows the results for 200 epochs at each lookback starting from 1 to 500.

MSE decreases as we increase the lookback.

https://doi.org/10.1371/journal.pone.0275649.g008

The last method is LSTM with Simulated Annealing in which we tried to reduce the computation time to find an optimal lookback without affecting the performance. We used simulated annealing in two steps: the first is to find the optimal lookback, and the second is to use the optimal lookback and increase the epochs to 200. For simulated annealing-based LSTM (20 iterations of simulated annealing to find an optimal lookback and 30 epochs), the computation time was 2.67 hours.

In Tables 46 columns 2, 3, 4, 5, 6, 7, 8, 9, 10 and 11 show the validation (mse, mae, rmse, explained variance, r2_score) and test (mse, mae, rmse, explained variance, r2_score) respectively for each iteration of simulated annealing. Table 7 shows the final results for 3 wind farms. The MSE for wind farms 1,2 and 3 is reduced to 0.0059,0.0074 and 0.010, respectively.

thumbnail
Table 4. Iteration 8 for lookback = 236 yields the minimum validation evaluation metrics.

https://doi.org/10.1371/journal.pone.0275649.t004

thumbnail
Table 5. Iteration 0 for lookback = 221 yields the minimum validation evaluation metrics.

https://doi.org/10.1371/journal.pone.0275649.t005

thumbnail
Table 6. Iteration 1 for lookback = 231 yields the minimum validation evaluation metrics.

https://doi.org/10.1371/journal.pone.0275649.t006

thumbnail
Table 7. This table shows the final result for 200 epochs and the optimal time step found using simulated annealing.

https://doi.org/10.1371/journal.pone.0275649.t007

The two-tailed p-values between each metric when simulated annealing is not used are shown in Table 8.

It is essential to understand overfitting and underfitting when simulated annealing is used for optimization.

The overfitting and underfitting of simulated annealing can be inferred from the results. In the case of simulated annealing, the optimal lookback can be a sub-optimal solution, but whether the optimal/sub-optimal lookback overfits/underfits on the training or not can be inferred from the test MSE.

The test data is not used to find an optimal look back, so the LSTM performance without simulated annealing and with simulated annealing can be used to see whether the model overfits or not on the training data. So, we see that MSE on test data is reduced with simulated annealing, which means the model does not overfit the training data when simulated annealing is used to find an optimal lookback. As far underfitting is concerned, the model does not underfit because when optimal lookback is used to train the model, the training MSE is reduced compared to when simulated annealing is not used.

Discussion

This section discusses the results generated by all two algorithms: Simple LSTM (A simple LSTM model with lookback = 1-500 and epochs = 30), and LSTM with Simulated Annealing (LSTM model with 20 iterations of simulated annealing, 30 epochs, and lookback = 1-500).

The first interesting observation is reduced computation time. Consider the results for windfarm 1 Table 4. The simulated annealing algorithm starts with lookback = 220, increasing the iterations. It mutates the lookback, rearranges the training data accordingly, and reports the evaluation metrics. The results for lookback = 236, shown in row 8, are the best, so we used that lookback to train the model for 500 epochs. Compare the same process with the simple LSTM, which iterates from lookback 1 to 500 and reports the epochs as shown in the Fig 8. The computation is the cost high in the simple LSTM.

Second, lookback for wind farms 1, 2, and 3 are different because the lookback is optimized for a particular wind farm. Third, we considered minimizing mean absolute error as an objective function. However, minimization of other parameters like mean explained variance or mean square error can also be considered an objective function.

The computation cost of the whole process depends on the number of iterations of the simulated annealing. Consider the number of epochs = 30, lookback in the range [1-500], and 20 iterations of simulated annealing. For each iteration, the computational time will be different because it depends on the lookback, as shown in the diagram Fig 6. The max time for each iteration is within the range [Time algorithm takes for lookback 1 = T1—Time algorithm takes for lookback 500 = T500]. The maximum time for 20 iterations can be 20*T500 if simulated annealing chooses lookback = 500 for all 20 iterations.

The time algorithm takes when lookback is linearly increased from 1 to 500 is the sum of all look back, which is greater than 20*T500. After finding the best lookback, the final step is to use LSTM with the best lookback for a specific number of epochs. The number of epochs in the final iteration is greater than the number of epochs used for the simulated annealing. So that is how simulated annealing reduces the execution time without reducing the performance.

Conclusion

This section includes the concluding remarks, future directions, and the limitation of the proposed algorithm. In this article, we proposed a simulated and healing-based LSTM for wind power prediction, which reduces the time to find an optimal lookback for LSTM prediction reducing the loss function. Rather than using simulated annealing, a particle swarm optimization algorithm or a combination of genetic algorithms and simulated annealing can be used for more robust predictions.

Due to limited space, we considered only 500 lookbacks (2009-07-02-13 to 2009-07-23-07 = 21 previous days), but in reality, data for more than one month should be considered to make the current prediction, which we believe can significantly increase the prediction. It also depends on the season; for example, the wind velocity is not constant throughout a year for a particular area, so if we consider large lookback information, the model’s performance may degrade.

We considered wind power prediction, and in the future, we plan to do it for other forecasting problems related to the energy sector like solar power prediction. Second, we considered only one optimization algorithm, simulated annealing, but optimization algorithms like particle swarm optimization and genetic algorithm can also be used to benchmark the performance. Benchmarking is necessary because different optimization algorithms yield different optimal lookbacks, and simulated annealing tries only a limited number of lookbacks (same as the number of iterations) to find an optimal one. Third, we considered only the lookback parameter for optimization, but simulated annealing can also be used to optimize multiple parameters like the number of epochs, which will undoubtedly increase the overall computation cost. The LSTM can be replaced with BILSTM, GRU, and machine learning-based regression algorithms with minor modifications.

Following are the computer and library specs used for implementing models and generating results. The system specifications are: Intel(R) Core(TM) 7-9750H CPU @ 2.60Hz, 16 GB RAM as well as a NVIDIA GeForce RTX 2060 GPU, running Microsoft Windows 10. Moreover, the development specifications are: Cuda compilation tools release 10.0, V10.0.130, Deep Learning framework Keras 2.4.3, Python 3.6.8, and Tensorflow 2.3.1.

All the code and dataset used to demonstrate the metholodgy of this paper is available at this link. NaiveLSTM directory contains a simple LSTM code (window size 1 and 500 epochs) for wind power prediction for 3 datasets. SimpleLSTM directory contains a simple LSTM code (lookback [1-500] and 200 epochs) for wind power prediction for 3 datasets. SimulatedAnnealing/code—find optimal lookback.py file contains a simulated annealing LSTM code (to find optimal lookback and 30 epochs) for wind power prediction for 3 datasets. SimulatedAnnealing/code—use optimal lookback.py file contains a simulated annealing LSTM code (to find optimal lookback and 30 epochs) for wind power prediction for 3 datasets.

References

  1. 1. Liu B, Chen J, Wang H, Wang Q. Renewable Energy and Material Supply Risks: a Predictive Analysis Based on An LSTM Model [Internet]. Vol. 8, Frontiers in Energy Research. Frontiers Media SA; 2020. Available from: http://dx.doi.org/10.3389/fenrg.2020.00163.
  2. 2. Gernaat DEHJ, de Boer HS, Daioglou V, Yalew SG, Müller C, van Vuuren DP. Climate change impacts on renewable energy supply [Internet]. Vol. 11, Nature Climate Change. Springer Science and Business Media LLC; 2021. p. 119–25. Available from: http://dx.doi.org/10.1038/s41558-020-00949-9
  3. 3. Hernandez RR, Jordaan SM, Kaldunski B, Kumar N. Aligning Climate Change and Sustainable Development Goals With an Innovation Systems Roadmap for Renewable Power [Internet]. Vol. 1, Frontiers in Sustainability. Frontiers Media SA; 2020. Available from: http://dx.doi.org/10.3389/frsus.2020.583090.
  4. 4. Saroha S, Kumar Aggarwal S, Rana P. Wind Power Forecasting [Internet]. Forecasting in Mathematics—Recent Advances, New Perspectives and Applications. IntechOpen; 2021. Available from: http://dx.doi.org/10.5772/intechopen.94550.
  5. 5. Pasari S, Shah A, Sirpurkar U. Wind Energy Prediction Using Artificial Neural Networks [Internet]. Sustainable Production, Life Cycle Engineering and Management. Springer International Publishing; 2020. p. 101–7. Available from: http://dx.doi.org/10.1007/978-3-030-44248-4_10
  6. 6. Peiris AT, Jayasinghe J, Rathnayake U. Forecasting Wind Power Generation Using Artificial Neural Network: “Pawan Danawi”—A Case Study from Sri Lanka [Internet]. Vallée F, editor. Vol. 2021, Journal of Electrical and Computer Engineering. Hindawi Limited; 2021. p. 1–10. Available from: http://dx.doi.org/10.1155/2021/5577547
  7. 7. Santhosh M, Venkaiah C, Vinod Kumar DM. Current advances and approaches in wind speed and wind power forecasting for improved renewable energy integration: A review [Internet]. Vol. 2, Engineering Reports. Wiley; 2020. Available from: http://dx.doi.org/10.1002/eng2.12178.
  8. 8. Kaushik S, Choudhury A, Sheron PK, Dasgupta N, Natarajan S, Pickett LA, et al. AI in Healthcare: Time-Series Forecasting Using Statistical, Neural, and Ensemble Architectures [Internet]. Vol. 3, Frontiers in Big Data. Frontiers Media SA; 2020. Available from: http://dx.doi.org/10.3389/fdata.2020.00004 pmid:33693379
  9. 9. Liu X, Gherbi A, Li W, Cheriet M. Multi Features and Multi-time steps LSTM Based Methodology for Bike Sharing Availability Prediction [Internet]. Vol. 155, Procedia Computer Science. Elsevier BV; 2019. p. 394–401. Available from: http://dx.doi.org/10.1016/j.procs.2019.08.055
  10. 10. Naidu K, Ali MS, Abu Bakar AH, Tan CK, Arof H, Mokhlis H. Optimized artificial neural network to improve the accuracy of estimated fault impedances and distances for underground distribution system [Internet]. Pławiak P, editor. Vol. 15, PLOS ONE. Public Library of Science (PLoS); 2020. p. e0227494. Available from: http://dx.doi.org/10.1371/journal.pone.0227494 pmid:31999711
  11. 11. Motta D, Santos AÁB, Machado BAS, Ribeiro-Filho OGV, Camargo LOA, Valdenegro-Toro MA, et al. Optimization of convolutional neural network hyperparameters for automatic classification of adult mosquitoes [Internet]. Zhang J, editor. Vol. 15, PLOS ONE. Public Library of Science (PLoS); 2020. p. e0234959. Available from: http://dx.doi.org/10.1371/journal.pone.0234959 pmid:32663230
  12. 12. Bell D, Gaillard F. Epoch (machine learning) [Internet]. Radiopaedia.org. Radiopaedia.org; 2017. Available from: http://dx.doi.org/10.53347/rid-56141.
  13. 13. Aarts EHL, Korst JHM, van Laarhoven PJM. A quantitative analysis of the simulated annealing algorithm: A case study for the traveling salesman problem [Internet]. Vol. 50, Journal of Statistical Physics. Springer Science and Business Media LLC; 1988. p. 187–206. Available from: http://dx.doi.org/10.1007/bf01022991.
  14. 14. Zameer A, Muneeb M, Mirza SM, Raja MAZ. Fractional-order particle swarm based multi-objective PWR core loading pattern optimization [Internet]. Vol. 135, Annals of Nuclear Energy. Elsevier BV; 2020. p. 106982. Available from: http://dx.doi.org/10.1016/j.anucene.2019.106982
  15. 15. Wang X, Guo P, Huang X. A Review of Wind Power Forecasting Models [Internet]. Vol. 12, Energy Procedia. Elsevier BV; 2011. p. 770–8. Available from: http://dx.doi.org/10.1016/j.egypro.2011.10.103
  16. 16. Chen Q, Folly KA. Short-Term Wind Power Forecasting Using Mixed Input Feature-Based Cascade-connected Artificial Neural Networks [Internet]. Vol. 9, Frontiers in Energy Research. Frontiers Media SA; 2021. Available from: http://dx.doi.org/10.3389/fenrg.2021.634639.
  17. 17. Deng YC, Tang XH, Zhou ZY, Yang Y, Niu F. Application of machine learning algorithms in wind power: a review [Internet]. Energy Sources, Part A: Recovery, Utilization, and Environmental Effects. Informa UK Limited; 2021. p. 1–22. Available from: http://dx.doi.org/10.1080/15567036.2020.1869867
  18. 18. Ren Z, Huang C, Li M. Research on Wind Power Prediction [Internet]. 2019 IEEE 3rd Conference on Energy Internet and Energy System Integration (EI2). IEEE; 2019. Available from: http://dx.doi.org/10.1109/EI247390.2019.9061851.
  19. 19. Zhang Y, Li R, Zhang J. Optimization scheme of wind energy prediction based on artificial intelligence [Internet]. Vol. 28, Environmental Science and Pollution Research. Springer Science and Business Media LLC; 2021. p. 39966–81. Available from: http://dx.doi.org/10.1007/s11356-021-13516-2. pmid:33763837
  20. 20. Hossain M, Mekhilef S, Afifi F, Halabi LM, Olatomiwa L, Seyedmahmoudian M, et al. Application of the hybrid ANFIS models for long term wind power density prediction with extrapolation capability [Internet]. Hu X, editor. Vol. 13, PLOS ONE. Public Library of Science (PLoS); 2018. p. e0193772. Available from: http://dx.doi.org/10.1371/journal.pone.0193772 pmid:29702645
  21. 21. Zheng D, Eseye AT, Zhang J, Li H. Short-term wind power forecasting using a double-stage hierarchical ANFIS approach for energy management in microgrids [Internet]. Vol. 2, Protection and Control of Modern Power Systems. Springer Science and Business Media LLC; 2017. Available from: http://dx.doi.org/10.1186/s41601-017-0041-5.
  22. 22. Tasnim S, Rahman A, Oo AMT, Haque ME. Wind Power Prediction Using Cluster Based Ensemble Regression [Internet]. Vol. 16, International Journal of Computational Intelligence and Applications. World Scientific Pub Co Pte Lt; 2017. p. 1750026. Available from: http://dx.doi.org/10.1142/s1469026817500262
  23. 23. Muneeb M, Henschel A. Eye-color and Type-2 diabetes phenotype prediction from genotype data using deep learning methods [Internet]. Vol. 22, BMC Bioinformatics. Springer Science and Business Media LLC; 2021. Available from: http://dx.doi.org/10.1186/s12859-021-04077-9 pmid:33874881
  24. 24. Shao B, Song D, Bian G, Zhao Y. Wind Speed Forecast Based on the LSTM Neural Network Optimized by the Firework Algorithm [Internet]. Liu W, editor. Vol. 2021, Advances in Materials Science and Engineering. Hindawi Limited; 2021. p. 1–13. Available from: http://dx.doi.org/10.1155/2021/4874757
  25. 25. Zhang Y. Research on Text Classification Method Based on LSTM Neural Network Model [Internet]. 2021 IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC). IEEE; 2021. Available from: http://dx.doi.org/10.1109/IPEC51340.2021.9421225.
  26. 26. Serpush F, Rezaei M. Complex Human Action Recognition Using a Hierarchical Feature Reduction and Deep Learning-Based Method [Internet]. Vol. 2, SN Computer Science. Springer Science and Business Media LLC; 2021. Available from: http://dx.doi.org/10.1007/s42979-021-00484-0 pmid:33615240