Next Article in Journal
Investigation on the Mineral Catalytic Graphitization of Anthracite during Series High Temperature Treatment
Next Article in Special Issue
Multitarget Intelligent Recognition of Petrographic Thin Section Images Based on Faster RCNN
Previous Article in Journal
Gold–Sulfide Mineralization in the Manitanyrd Region, Polar Urals, Russia
Previous Article in Special Issue
Coal Structure Prediction Based on Type-2 Fuzzy Inference System for Multi-Attribute Fusion: A Case Study in South Hengling Block, Qinshui Basin, China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recursive Feature Elimination and Neural Networks Applied to the Forecast of Mass and Metallurgical Recoveries in A Brazilian Phosphate Mine

by
Fernanda Gontijo Fernandes Niquini
1,*,
André Miranda Brito Branches
1,
João Felipe Coimbra Leite Costa
1,
Gabriel de Castro Moreira
1,
Claudio Luiz Schneider
2,
Florence Cristiane de Araújo
3 and
Luciano Nunes Capponi
3
1
PPGE3M (Programa de Pós-Graduação em Engenharia de Minas, Metalúrgica e de Materiais), Universidade Federal do Rio Grande do Sul, Av. Bento Gonçalves, 9500-Setor 4-Prédio 74-Sala 211, Campus do Vale, Porto Alegre 91509-900, Brazil
2
Centro de Tecnologia Mineral (CETEM), 900-Ilha da Cidade Universitária, Av. Pedro Calmon, Rio de Janeiro 21941-596, Brazil
3
Mosaic Fertilizantes P&k Ltd., Rodovia MGC, Km 196.5, Tapira 38185-000, Brazil
*
Author to whom correspondence should be addressed.
Minerals 2023, 13(6), 748; https://doi.org/10.3390/min13060748
Submission received: 15 March 2023 / Revised: 19 May 2023 / Accepted: 20 May 2023 / Published: 31 May 2023

Abstract

:
Geometallurgical models are commonly built by combining explanatory variables to obtain the response that requires prediction. This study presented a phosphate plant with three concentration steps: magnetic separation, desliming and flotation, where the yields and recoveries corresponding to each process unit were predicted. These output variables depended on the ore composition and the collector concentration utilized. This paper proposed a solution based on feature engineering to select the best set of explanatory variables and a subset of them able to keep the model as simple as possible but with enough precision and accuracy. After choosing the input variables, two neural network models were developed to simultaneously forecast the seven geometallurgical variables under study: the first, using the best set of variables; and the second, using the reduced set of inputs. The forecasts obtained in both scenarios were compared, and the results showed that the mean squared error and the root mean squared error increase in all output variables evaluated in the test set was smaller than 2.6% when the reduced set was used. The trade-off between simplicity and the quality of the model needs to be addressed when choosing the final neural network to be used in a 3D-block model.

1. Introduction

Geometallurgical models are essential for the mining industry. These models combine geology, geochemistry, mineralogy, mine planning, rock mechanics, and metallurgy [1] to make predictions about mineral processing performance. This performance can be measured in several ways, depending on the goals to be achieved. Comminution energy requirements; ore resistance to impact breakage (A × b); Ball Mill Work Index [2,3,4,5,6] throughput [3,7,8]; particle size in the product—defined as the size at which 80% of the particles are smaller (commonly known as P80) [9]; yields and recoveries [6,10,11,12,13,14,15]; the contaminant grades in tailings [16,17]; and acid consumption [18,19] are variables frequently modeled in geometallurgical studies. The most common input variables used to build geometallurgical models include the mineralogical features; the geological descriptions (containing lithological, textural, alteration, and weathering information); the chemical composition; and the specific gravity of the ore. All of these generate many variables that need to be analyzed and are related to plant performance.
The data collected from the ore vary in importance in terms of their effects on the metallurgical performance of the process. Proper selection of the features is an important step in geometallurgical modeling. The selection of variables, also called feature selection, obtains the subset of input data that contributes the most to determining the dependent variables (target/output) to be modelled. The advantages of the feature selection procedure are:
  • Removing redundant and irrelevant variables;
  • Decreasing noise in the forecast, improving the efficiency of the model;
  • Reducing the risk of overfitting;
  • Reducing the computational cost of processing.
The process starts by deciding which variables should be used as inputs in the geometallurgical model. It is prevalent to start with a correlation matrix analysis aimed at identifying the input variables with the strongest relationships with the outputs to be modelled. If the selected variables present a strong correlation coefficient among themselves, this can lead to a collinearity problem [20]. Furthermore, when the number of variables increases, analyzing the correlation matrix can become a complex problem due to the number of cells to interpret. Other methods that are frequently used to reduce the number of variables are listed in Table 1, along with the core idea of the method, disadvantages, and the appropriate references.
Recursive feature elimination (RFE) [28] is another method used to select a subset of features to reduce dimensionality. The idea is to start with a model containing all of the available variables and iteratively remove one or more features in each algorithm step, until only one variable remains. The improvement in the recursive feature elimination, in relation to the backward elimination, is that the RFE allows the user to choose which technique will be used to elaborate the forecast model. It allows the use of support vector regression, random forest, linear regression, and other techniques which contain a measure of the importance of each variable during the modeling. In contrast, the backward elimination was conceived to be used only with linear regression models. Furthermore, the RFE allows the user to remove more than one variable in each iteration, while the backward elimination was initially developed to remove one variable at a time.
The RFE technique was chosen to be applied in this study for three main reasons: first, the relation between the number of samples and the number of variables was not a problem for the database under study. There were 3056 samples of ten input variables (nine chemical components and the collector concentration used in the flotation test), making techniques such as PCA, MAF, and super secondary not advantageous due to the difficulty in interpreting the resulting variables. The second reason was that the application of clustering to choose variables demands more time for interpretation. This study’s core objective was to develop a semi-automatic workflow to be applied and quickly updated whenever new samples were introduced into the database. The third reason is that RFE is an improvement when compared to the stepwise methods, more specifically the backward elimination, allowing the use of other algorithms in its core and expanding the modeling options under evaluation. Another advantage of the RFE algorithm is its use combined with k-fold cross-validation [29], allowing the user to evaluate the variability of the results by varying the samples used.
Thus, the workflow proposed in this paper was to carry out a feature selection study using RFE to reduce model complexity and, after the most important features were selected, to elaborate a neural network model to predict the processing plant response for any ore sample. Some examples of neural networks applied to forecast geometallurgical variables, as proposed in this paper, are listed below:
  • Niquini and Costa used neural networks to forecast yield and metallurgical recovery of all plant outputs in a gold [15] and a zinc [14] processing plants. The results showed that neural networks are able to generate accurate and precise predictions while satisfying the mass and metallurgical balance constraints;
  • Both and Dimitrakopoulos [30] applied neural networks to forecast throughput and compared the results against the throughput predicted by linear regression. The NN showed better results and the predictions obtained with its use can be integrated into production scheduling;
  • Gholami et al. [31], used neural networks to forecast four geometallurgical variables in a copper mine, using as inputs particle size distribution, collector and frother concentrations, solids content, pH and mineralogical variables;
  • Jorjani et al. [32] created a neural network model to forecast La, Ce, Y and Nd recoveries from an apatite concentrate. The input variables were leaching time, pulp density, agitation rate and acid concentration;
  • Nakhaei et al. [33] developed a neural network model to forecast Cu and Mo grades and recoveries using as inputs Cu and Mo grades measured in the flotation circuit as well as some flotation circuit operational variables;
  • Srivastava et al. [34] elaborated a neural network model to predict the power in a SAG mill using, as inputs, the mill power observed in the previous time increments and the feed rate in the evaluated time;
  • Panda and Tripathy [35] created a neural network to forecast the grade and metallurgical recovery of chromite using operational variables as inputs.
After developing the neural network model, the process response can be included in a 3D block model used for mine planning. The reduced input scenario was compared to the results obtained when the best scenario, using all variables available, was used to build the neural network forecasting model. By doing this, it was possible to determine the impact of removing features from the neural network inputs.

Process Modelling and Geometalurgical Modelling

A general solution for any flowsheet was established as early as 1979 [36], allowing for accurate solutions for very complex and extensive mineral-processing plant flowsheets. The performance of a mineral-processing plant can be predicted using a plant-wide simulator [37]. However, the simulation prediction accuracy depends on several factors related to the unit operation models and the ore model (also known as a geometallurgical model). These two models cannot be completely separated and, in many cases, the unit operation models contain ore model parameters such as comminution parameters for breakage and selection functions. Nevertheless, it is advantageous to understand the problem as if the unit operation models and the ore models are two separate, but not independent, entities. The reason for this is that, in the case of mine planning, the plant may, and should, be considered as a fixed entity, a group of processes that will perform only as a function of the ore that is fed to it. System data may change as a function of the ore that is fed to the plant, and this includes solids flowrates, water flowrates, and particle-size distributions. If we look at this problem from this perspective, it is clear that the ore model is the main factor determining plant performance and the accuracy of the predictions made.
At the heart of the mineral-processing plant that was considered in this study, is the flotation circuit. Although flotation is widely used to concentrate large tonnages of a variety of ore minerals, it is a challenging process to model because of the large number of micro-processes involved. King [37] gives an excellent review of the state of the art of flotation modeling.
In King’s approach, the flotation process requires four models, for the pulp phase, the bubble phase, the froth phase, and the entrained phase. Each of these models requires several parameters, some of which are difficult to measure in practice. Many of the micro-processes involved in flotation can be bundled in flotation rates. The onset of the development of kinetic models is due to Sutherland [38]. Since then, several authors have contributed to understanding flotation kinetics through developing particle-bubble collision efficiency models [39,40,41]. Clift, Grace and Weber [42] and Schulze [43] contributed with hydrodynamic considerations of bubble motion. Bubble-rise velocity models were developed by Karamanev and Nikolov [44]. The micro-process of bubble loading and its influence in flotation rates were investigated by Tomlinson and Fleming [45], King, Hatton and Hulbert [46] and Bradshaw and O’Connor [47]. Recovery as a function of particle size was modelled by Trahar and Warren [48].
Another important process in flotation is particle detachment, as described by Drzymala [49]. A simple model for the froth phase was developed by Murphy, Zimmerman and Woodburn [50]. Several other authors have made excellent contributions to understanding and modelling each of the processes involved in flotation. The point here is to show that it is not feasible to include every aspect of flotation simulation into a geometallurgical model. For example, Neethling and Cilliers [51] provided a significantly more accurate and realistic model for the froth phase that includes coalescence. However, the computation is complex, and it is not likely that one simulation could be accomplished in a fraction of a second.
While plant-wide simulation is suited for predicting yields and recoveries in the context of mine planning, it would require a detailed characterization of the ore such as, but not limited to, complete mineralogical assembly and liberation by size information for every drill-core sample that is produced in the mine. At this stage in the development of characterization techniques, this would consume too much time and would also be financially unattractive. Process modelling is best employed in the context of flowsheet optimization and for scaling up unit operations. In these applications, it is feasible to obtain fundamental laboratory-scale parameters for a given ore sample. In the context of geometallurgical modelling for mine planning, the approach that is described in this paper is better suited because the number of samples that need to be characterized and tested is large, the flowsheet does not change, or, more accurately, the geometallurgical model is developed for a given flowsheet and it is valid while the flowsheet is not changed, and there is no need for scale-up parameters as all unit operations have already been scaled at the time the plant was designed.

2. Data Processing

2.1. Geological Context

The geological features of the region were studied extensively [52,53,54,55] due to the anomalous occurrences of phosphate, titanium, niobium, and rare earth elements. The anomalies are related to the weathering of Cretaceous alkaline intrusive rocks. The weathering layer often reaches several tens of meters in depth and the unaltered rock is only usually observed in cores from deep diamond drilling.
In the study area, anomalous phosphate concentrations occur due to the presence of apatite. Naturally, the alteration is more intense closer to the surface and the apatite there is almost completely leached. The supergenic concentration of this mineral results from the solubilization and leaching of the more unstable components contained in the original rocks, such as mafic minerals and carbonates, mainly pyroxenites.
During the weathering process, at first, carbonates and mafic minerals are leached, while apatite crystals remain, as well as titanium, the latter in the form of perovskite. With the progress of such transformations, apatite is converted into secondary minerals, from the crandalite group, of reduced solubility and with lesser economic value. Meanwhile, the perovskite transforms into anatase, which gradually becomes residually concentrated. This process has led to the current weathering profile of the study area, where the intermediate-upper layer shows a high content of titanium in the form of anatase, and the intermediate-lower portion presents higher grades of phosphatic ore and lower concentrations of titanium.
The lower portion of the weathering profile comprises semi-altered rocks, which may marginally contain economic concentrations of apatite, but predominately comprise mafic minerals, carbonates and magnetite. Variations due to the heterogeneity of the original magma lead to higher or lower occurrences of minerals such as perovskite, phlogopite, and vermiculite.
The samples taken for this study corresponded to the intermediate-lower portion of the profile, which was classified as ore during the geological logging of the diamond-drilled cores.

2.2. Dependent and Independent Variables

The database under study contained 3056 individual geometallurgical samples, with Al2O3, BaO, CaO, Fe2O3, MgO, Nb2O5, P2O5, SiO2, and TiO2 grades analyzed from vertical diamond drilling. The feature selection study also considered the collector (fatty acid) concentration used in the flotation batch tests as an input variable. This concentration was determined by the laboratory technician, taking into account some sample aspects such as: apatite crystal cleanliness; P2O5 grade after desliming and demagnetization; and ore texture. The milling time adopted in the processing, the depressor (corn starch) concentration, the pH, the flotation time, the strength of the magnetic field and the solid concentration in the flotation pulp were kept under constant values to not affect the results; the influence of these variables on the test responses was nil, and they were not considered as input variables.
Mineralogy and rock texture, although very important aspects of a geometallurgical study, were not directly available for this particular study as specific features, but were certainly reflected in the wide assembly of grade variables considered. Table 2 summarizes the ROM information and the collector concentration, showing that the mean grade of P2O5 was 9.74%.
Understanding the relationship between independent and dependent variables requires knowledge of the mineral processing plant’s flowsheet (Figure 1). All geometallurgical samples were submitted to batch tests that mimicked the flowsheet: starting with a magnetic separation step, followed by desliming and flotation.
Magnetics, limes, tailings, and apatite concentrate are plant products. Therefore, each batch test generated the relative fraction of the initial mass destined for each product, making it possible to calculate the yield corresponding to each plant output. In this paper the term “yield” was used to express the ratio between weight or throughput of a stream and the throughput of the plant feed [56,57]. The term “recovery” was used to express the fraction of apatite recovered in a stream with respect to the feed stream.
Let Qstream represent the solids flowrate of a given stream. Let Ystream represent the yield of a given stream, then Equations (1) to (4) follow.
Y M a g n e t i c s = Q M a g n e t i c s Q R O M
Y S l i m e s = Q S l i m e s Q R O M
Y C o n c e n t r a t e = Q C o n c e n t r a t e Q R O M
Y T a i l i n g s = Q T a i l i n g s Q R O M
Thus, four yield values were available (totalling 100%): magnetics, slimes, concentrate, and tailings yield.
Chemical analysis is routinely carried out for the ROM, the flotation feed, and the flotation concentrate streams of the batch tests. The grades in the flotation tailings stream can be calculated from the measured grades in the flotation feed and concentrate streams. The grades in the combined magnetics and slime streams can be calculated using the grades and yields measured in the ROM and in the flotation feed. Let the metallurgical recovery of P2O5 in any stream be RStream and the grade of P2O5 in any stream be gStream. The metallurgical recoveries of P2O5 were calculated using the yields and grades as shown in Equations (5) to (7). Considering all output points with chemical information (calculated or analysed), it is possible to determine the P2O5 metallurgical recovery with respect to the ROM P2O5 content, as presented in Equations (6) to (8). It should be noted that when these three variables are added, the result equals 100%, as should be the case with the yield variables.
R S l i m e s + M a g n e t i c s = Y S l i m e s + Y M a g n e t i c s × g S l i m e s + M a g n e t i c s g R O M
R C o n c e n t r a t e = Y C o n c e n t r a t e × g C o n c e n t r a t e g R O M
R T a i l i n g s = Y T a i l i n g s × g T a i l i n g s g R O M
Therefore, the dependent variables to be modeled are the slimes, magnetics, concentrate, and tailings yields, as well as the metallurgical recoveries in the slime + magnetic, in the flotation concentrate, and the flotation tailings. Table 3 presents the descriptive statistics for all dependent variables under study. On average, 47.81% of all P2O5 and 13.74% of all mass fed in the plant are destined for the flotation concentrate.

2.3. Database Partition

Twenty percent of the samples were randomly selected and set apart to be used as a test set to evaluate the model’s performance. All models presented were elaborated using the remaining 80% of the samples in the database. The k-fold technique [29] was used to determine the optimum parameters for each proposed model. Then, the models obtained were applied to the test dataset.

3. Materials and Methods

The methodology proposed in this work was based on two main methods: artificial neural network (ANN) and recursive feature elimination, both belonging to the machine learning framework.
As explained before, RFE is a technique used in, so-called, ‘Feature Engineering’ aimed at identifying the most relevant input variables to predict either one or a set of output variables. The goal is to use the least possible number of variables, thus simplifying the model while maintaining good accuracy in the predictions.
ANNs are computing systems designed to most efficiently map the relationships between input and output variables in complex structures. These two subjects are more extensively described in the following sections, and the overall methodology can be outlined as follows:
(1)
Selection of the most relevant input variables using RFE;
(2)
The construction of ANN models to predict metallurgical attributes using the input variables selected in the previous step;
(3)
Checking the performance of the built model by forecasting the test dataset.

3.1. Recursive Feature Elimination

As already mentioned, one must decide how many variables will be used as inputs in the prediction model. It is common for the geometallurgical dataset to contain data that is either redundant or not relevant to the prediction of the process performance. Recursive feature elimination (RFE) [28] is an algorithm for automatic feature selection that uses a prediction algorithm in its core to rank the inputs by relevance using a predefined metric. The process starts by creating a predictive model with all the information, organizing the information, and measuring model performance. The less critical predictor is removed at each step. The model is rebuilt iteratively until the desired set of inputs is achieved.
Here, the most relevant subset was found using the RFE version with cross-validation (RFECV) [58], partitioning the dataset tenfold, three times (ten-fold validation with three repetitions). Three different machine learning multitarget predictive models were used in its core: linear regression [59], random forest [60], and CatBoost regressor [61], which will be briefly explained below.
Linear regression is one of the most commonly used techniques for building geometallurgical models [15]. Its goal is to find the equation which minimizes the prediction’s mean squared error using ordinary least squares. To rank the variables for importance, the recursive feature elimination uses the p-value of each variable, removing the one corresponding to the higher p-value in each algorithm iteration, until only the variable with the smallest p-value remains.
The random forest algorithm establishes cutoffs in the input variables, which help to estimate the output variables under study. The greater the number of times an input variable is selected to best describe the output, the greater its importance in the model’s creation. This importance is used inside the RFE to rank the input variables.
The CatBoost algorithm was developed based on the decision trees and gradient boosting theory. The main idea of boosting is to sequentially combine many weak models (small decision trees). Because gradient boosting fits the decision trees sequentially, the fitted trees will learn from former trees’ mistakes and reduce the errors. Adding a new function to the existing ones is continued until the selected loss function is no longer minimized [61]. Similar to other decision tree-based algorithms, CatBoost uses feature importance to rank the relevance of all inputs for the model being created.
It should be noted that only techniques which present a measure of variables’ importance can be used in RFE. Artificial neural networks, ACE [62], and others which do not provide this kind of information, cannot be used in combination with the recursive feature elimination.
To evaluate the performance of each model built, the following metrics were chosen: mean absolute error (MAE), square root of the mean squared error (RMSE), mean absolute percentage error (MAPE), median absolute percentage error (MdAPE), and the Pearson correlation coefficient. Once this study dealt with multitarget models, the mean value of each variable metric was taken to give a general idea of the model’s precision and accuracy. For example, the MAE value was calculated for each of the seven output variables under study. The mean of the seven MAE values was then calculated and this was named the global MAE. The same procedure was used for the other metrics.
The feature selection process was developed in three stages. First, the three predictive algorithm performances were evaluated with RFECV for each subset of relevant inputs in decreasing order until only the most important variable remained, obtaining the global MAE value in each algorithm step. The lowest global MAE value indicated the best predictive algorithm and optimum subset size.
In the second stage, the algorithm selected previously was used inside the RFECV, and the global MAPE for the optimum set was calculated, this being the reference value. All subsets of the variables were analysed (with a lower number of variables than the best case), which provided a global MAPE increase of a maximum of 5%, compared with the reference value. Thus, the smallest subset within the percentage cut-off was chosen as the optimum subset.
In the final third stage, to highlight the importance of the features, an adaptation of the SHAP-values algorithm [63] was used to show the importance of each variable as percentage values.
After the feature selection step, the development of two neural networks started: the first one using the optimum set of variables as inputs (in this case, all available variables) and the second using the variables contained in the optimum subset as inputs. Using this technique to achieve the forecast models was preferred to the CatBoost algorithm, once the neural network results were comparable to the CatBoost results while presenting the advantage of providing the yields and metallurgical balances between the forecasts. Details about the artificial neural network technique are given below.

3.2. Artificial Neural Networks

The artificial neural network technique (ANN) is versatile in geometallurgical model construction. This method can forecast variables with nonlinear relationships with the input parameters, such as the yields and metallurgical recoveries. Additionally, it can simultaneously predict two or more dependent variables using one single network structure. Due to these characteristics and a history of good results [12,14,15], ANN was chosen to predict the geometallurgical responses studied herein. A brief description of the theory related to ANN is given below. For more details about neural networks, their mathematics, and their applications in the mining industry, the reader should refer to Niquini and Costa [14].
Determining which, and how many, variables should be used as inputs is necessary, given that different structures will lead to different results. The independent variables in neural networks are named “input variables” (represented by X), while the dependent variables are named “output variables” (represented by Y). Therefore, the oxide grades in the ROM and the collector dosage inserted in the flotation tests are information that can be used as inputs. The yields and metallurgical recoveries are in the network outputs.
In addition to the input and output layers, there are three other fundamental structures in the ANN: the hidden layer(s), the network weights, and the activation functions (Figure 2).
The user determines the number of hidden nodes and the number of hidden layers. The hidden layer is the intermediary part of the network, where the intermediate calculation to generate the network’s final forecast is carried out. Each hidden node is denoted by Z.
The lines that connect the input layer to the hidden layer and the hidden layer to the output layer are named network weights. They determine the nodes with a greater or a lesser influence in each output node forecast: the higher the weight (in absolute terms), the more significant the influence of the related node, and vice-versa. The determination of α and β weight is usually made using the backpropagation algorithm (details in [64]). Equation (8), an adaptation of the notation given by Izenman [64], shows the position of the weights inside the network forecast function.
y ^ i , k = g k j = 0 t α k j f j m = 0 r β j m x i , m
where i refers to the sample number in the database, with i = 1, …, n (where n is the total number of samples); k is the index which represents the output variable under study, with k = 1, …, s (where s is the number of output nodes); j is the index which represents the hidden nodes, with j = 0, …, t (where t is the total number of hidden nodes); and m represents the input nodes, with m = 0, …, r (where r is the total number of input nodes in the network).
The functions g and f, in Equation (8), and shown in Figure 2, are the activation functions: a set of predetermined functions which are fed with the linear combination of the input nodes (x) and its β weights, in the case of the f function, or fed with the linear combination of the hidden nodes (z) and its α weights, in case of the g function. Of all the activation functions that exist in the literature, the following were used in this study:
Linear function:
φ v = v
Sigmoid function:
φ v = 1 1 + e v
ReLU function:
φ v = m a x 0 , v
Softplus function:
φ v = l n 1 + e v
Tanh function:
φ v = ( e v e v ) ( e v + e v )
Using Equation (8) and the activation functions mentioned above, it is possible to forecast the geometallurgical variables of interest. The models can predict the processing plant performance in each block of the 3D block model, aiding in routine mine scheduling procedures.

4. Results and Discussion

The first step in elaborating the forecast model for the seven output variables consisting of yields and recoveries, was to determine which input variables would be used in the neural network. Following the methodology proposed, the first stage included finding the best predictive algorithm and subset size. Figure 3 presents the performance of the three algorithms for a different number of features. As can be seen, the CatBoost Regressor algorithm is the best for any quantity of inputs, with an optimal performance being achieved by using 10 features in this scenario. Figure 3 also shows that the global MAE presents a lower rate of decrease from four or more inputs. Hence, it is necessary to analyse whether the size reduction of any chosen subset will significantly affect the model’s performance and which variables are selected in each scenario.
The CatBoost algorithm was applied for the optimum set of variables (10 features) and for the subsets containing four to nine features to reduce model complexity. The results for four, five, and six features are presented due to their importance. A ten-fold cross-validation technique was used, considering different randomly selected seeds. The global MAPE value was calculated for each set size, as presented in Table 4. The increase in the global MAPE value, compared to the best case, is presented in Table 5.
By analyzing the results, it is possible to conclude that, by choosing the six most relevant features, the global MAPE increment is lower than 5% when comparing this metric against the global MAPE obtained for the best case. Thus, the scenario using six variables was chosen to continue the study.
In the next step of the analysis, the CatBoost hyperparameters were calibrated for the two sets of variables: the best one, with all variables available, and the second, with six variables chosen in the reduced scenario. The SHAP-values algorithm was then used to estimate each input variable’s importance (at percentage levels) in both scenarios. Figure 4 illustrates the importance of the optimum scenario (10 features), while Figure 5 presents this information for the subset with six variables. The graphics generated by the algorithm can be interpreted as follows: the X axis represents the importance of each feature to the model (the higher the value, the more important the feature), while the Y axis shows the input variables included in the evaluation. The horizontal bar colors are related to the output variables that are to be predicted (the larger is the color bar, more important the feature is to forecast the respective output behavior). More details about the SHAP values technique can be read in [65].
Amongst the grades, the most important feature to describe the geometallurgical variables was the CaO concentration in the ROM, followed by the MgO concentration in the best scenario and Fe2O3 concentration in the sub-optimum scenario. The importance of collector dosage is notorious in both scenarios; being even more important to forecast the flotation concentrate and the tailings recoveries. It is known that the collector dosage cannot influence the desliming and demagnetization processes; thus, the influence of this variable on the yield of slimes, the yield of magnetics and the recovery of slimes + magnetics presented in Figure 4 and Figure 5, are related to a cross-relation with other variables, for example: if the sample has a lot of apatite grains with magnetite, part of the magnetic material will be concentrated in the magnetic separator which will present a high yield, but another part of the apatite grains with magnetite will follow through to flotation. When entering in the flotation step, the laboratory technician will see the apatite grains with magnetite in the sample and will add a higher collector dosage, in order to improve the flotation behavior. Thus, the collector dosage was correlated with the magnetic mass yield not because the first affected the latest, but because the first was calibrated to deal with the characteristic of the sample not concentrated in the magnetic separation step.
BaO, Nb2O5, SiO2 and TiO2 grades were the variables removed in the reduced scenario. It is known that barite and pyrochlore, the minerals that carry BaO and Nb2O5, are present in the deposit and they may affect flotation when their grades are significant; more specifically the barite which competes for the collector of apatite in flotation. Because these minerals are not usually present in high quantities, their influence in the global scenario is reduced, so the grades of BaO and Nb2O5 can be removed. The TiO2 grade was probably selected for removal by the algorithm because it is correlated with the other two variables that were kept in the model (correlation of 0.48 with Fe2O3 and 0.46 with MgO). The same occurred with SiO2, which is correlated with the grades of Fe2O3 (0.68), P2O5 (0.56) and MgO (0.53).
Next, the neural network’s hyperparameters were calibrated, starting with the complete model with the optimum set of variables, presented in Figure 4. Several tests were carried out to identify the best learning rate [66] to adjust this neural network, and the value found was 0.001. One thousand epochs were used. Two hidden layers were used: the first one with eight hidden nodes and a softplus activation function and the second with nine hidden nodes and a linear activation function. The linear activation function was used in the output layer. These sets of hyperparameters were selected because they provided the smallest global MAE for the validation set. The metrics generated by this network are summarized in Table 6. The first value in each cell represents the metric mean of the ten-fold validation, and the second represents the standard error. At the bottom of the table, the test set results obtained when the training and validation sets are combined to create the final model are presented. The MdAPE metric was selected to be analyzed because the outliers (illustrated in Figure 6) inflated the mean absolute percentage error (MAPE), which may lead to an erroneous model interpretation.
The correlation coefficient between the forecasts and the real values for the flotation concentrate yield was 0.866, and the MdAPE found was 14.897 in the test set. For the flotation concentrate P2O5 recovery, the correlation coefficient was 0.625, associated with a MdAPE value of 13.925. The slimes yield and the slimes + magnetics P2O5 recoveries also showed good MdAPE values (12.449 and 16.199, respectively), although the correlation coefficient between the real values and the forecasts is not very high. The magnetics yields and tailings P2O5 recovery were the variables with the highest MdAPE values (27.851 and 34.133, respectively).
After completing the analysis with the optimum scenario, the neural network’s calibration was conducted using the six variables presented in Figure 5. Two hidden layers were used: the first with 10 hidden nodes and a sigmoid activation function and the second with 10 hidden nodes and a linear activation function. The linear activation function was used in the output layer. The best learning rate found to adjust this network was 0.002 and one thousand epochs were used. The metrics generated by using this network are summarized in Table 7.
The results presented in Table 7 show that for the reduced set of input variables, the metrics observed were very similar to the ones seen in the optimum scenario, especially for the variables with the most significant economic impact, namely the flotation concentrate yield and P2O5 recovery.
Using six and 10 features, a comparison of the two scenarios can be summarized by calculating the relative differences, as shown in Equation (14). Table 8 summarizes the relative differences observed between the Pearson correlation coefficients, MAE, RMSE, and MdAPE mean values.
R e l a t i v e   D i f f e r e n c e = M e t r i c   V a l u e 6 f e a t u r e s M e t r i c   V a l u e 10 f e a t u r e s M e t r i c   V a l u e 10 f e a t u r e s 100 %
By analysing the relative difference in the test set, it is possible to conclude that when the correlation coefficient decreases, the fall is not marked (smaller than 0.7%), showing that the reduction in the number of input variables did not impact the model’s precision. In most cases, an improvement in the correlation coefficient value is noted when the reduced set of variables is used.
Removing four variables from the inputs did not significantly affect the metrics. The relative differences in the MAE and RMSE values showed an increase in errors lower than 2.6% when the reduced set of variables is used instead of the optimum set. The most significant impact observed was in the MdAPE values for the flotation concentrate mass and metallurgical recoveries, showing an increase of 8.5% and 8.8% in the test set, respectively, when the reduced scenario was adopted. The reduced model can be considered as a good option to forecast the metallurgical variables of interest because the MdAPE increase was lower than 4% for all other output variables.
Considering the above results, it can be concluded that it is possible to work with the reduced model without losing much, in terms of model precision and accuracy. There is a gain in simplicity once adding variables in the process can represent a considerable increase in the workload, not only in computer power but also in time spent by the analyst to elaborate, update, and evaluate the models created to estimate the input variables in the block model using kriging or simulation. Analysing through the geostatistical point of view, passing from 10 to six variables to be modelled implies a work-time reduction of 40%, a considerable amount.

5. Conclusions

Reducing the number of variables leads to a lower computational cost. The gain is also felt in the geostatistical workflow, once the number of input variables to be integrated to the resource block model is reduced. However, those benefits are counterbalanced by higher errors in the forecasting of some variables. The trade-off between a small error and a higher labor force is important when applying such models to a complete industrial operation. To help with this decision, it is important to quantify the errors related to the two approaches and compare them, as shown in this paper.
In this case study, the reduction from 10 to six input variables showed that dimension reduction can lead to models capable of returning correlation values close to the ones observed when the complete set of variables is used, with a small loss in the reduced scenario, when it occurs. The increase in MAE and RMSE was smaller than 2.6% for all output variables in the test set when the sub-optimum set of variables was used instead of the set with all 10 features. The four removed variables make sense considering the points discussed in chapter 4: the BaO and Nb2O5 are related to the presence of barite and pyrochlore, which only affects the process in high quantities (which is rare); and SiO2 and TiO2 grades are correlated with other variables that were maintained in the model, making their presence redundant. Thus, it was decided to select the model elaborated with the reduced number of input variables, due to its quality, and considering the practicality of the whole process, as this is a workflow being suggested for an active mining operation where time response represents a crucial aspect.
We would like to stress that each case study is unique. It is possible to achieve smaller errors in the test set with a reduced set of variables compared with the optimum scenario, especially if small datasets are used, due to overfitting (caused by a high number of variables and a small number of samples). It is also possible to find higher errors that do not justify the dimensionality reduction. Thus, it is important to follow the proposed workflow and analyze the errors in both scenarios to choose the best option for each mine under study.
It is important to mention that the geometallurgical ore model herein described is essentially statistical in nature. Nevertheless, the model is powerful enough to include some of the physical–chemistry aspects of the process. For example, the properly carried out investigation has led to the establishment of a quantitative relationship describing the deleterious effect of the presence of carbonates in the plant’s feed, understood as the ratio between CaO and P2O5, proving that important geochemical/mineralogical relationships were mathematically predicted, giving an interpretation of the results that makes geological sense.
Another important point that needs to be stressed is related to the estimation of the collector dosage in the block model. In this case study, this explanatory variable was attributed to the blocks by the nearest neighbor estimation method. Thus, mining regions which require a smaller collector dosage can be recognized, and areas which demand a higher collector dosage are also noted.

Final Thoughts about Mineral Processing

Usually, the processing plant’s human operator(s) has a well-established objective of a certain number of tonnes of the product (concentrate) with a required grade or quality. Such quality may include different grades (for example, SiO2 and Fe2O3, but most certainly P2O5). A conveyor belt equipped with a scale provides the tonnage of the filtered product, so this information is readily available through the control system. At short time intervals, samples are taken and sent for analysis so that, given a short delay, product quality and moisture content are also available to the operator. Given that the plant is operating at a steady pace, the operator may decide to change the processing parameters so that the production objectives are met during his shift. The changes that can be made are relatively small, such as in the reagent dosage, aeration rate, and perhaps also the froth levels, and froth-wash-water flowrate.
The point here is that most of the plant’s processing parameters do not change at all, and for the few parameters that do change, so that production requirements can be met, the impacts on the production rate and quality are well-known. Therefore, it is safe to assume that the plant and all its processing parameters can either be changed to meet the required quality or not. Simply put, if the quality can be met, the rock is ore, and if not, the rock is waste. The broad question that is left concerns the metallurgical recovery of P2O5 as a function of ore quality. In other words, for the development of the geometallurgical model, the main concern is with the ore model.

Author Contributions

Conceptualization, F.G.F.N. and J.F.C.L.C.; methodology, F.G.F.N. and A.M.B.B.; software, F.G.F.N. and A.M.B.B.; investigation, F.G.F.N.; data curation, F.C.d.A. and L.N.C.; writing—original draft preparation, F.G.F.N. and C.L.S.; writing—review and editing, J.F.C.L.C., C.L.S., G.d.C.M., A.M.B.B., L.N.C. and F.C.d.A.; project administration, J.F.C.L.C., L.N.C. and F.C.d.A.; funding acquisition, J.F.C.L.C. and C.L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by Fundação Luiz Engert.

Data Availability Statement

The data is unavailable due to confidentiality.

Acknowledgments

The authors would like to thank the mining company which kindly provided the data used in this study and Carla Grasso for all technical discussions, which made this work possible and the results applicable to the industry.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Williams, S.R.; Richardson, J.M. Geometallurgical Mapping: A New Approach that Reduces Technical Risk; SGS Minerals Services: Truro, UK, 2004; 13p. [Google Scholar]
  2. Lopera, P.A. Geometallurgical Mapping and Mine Modelling—Comminution Studies: La Colosa Case Study, AMIRA P843A. Master’s Thesis, University of Tasmania, Hobart, Australia, 2014. [Google Scholar]
  3. Newton, M.J.; Graham, J.M. Spatial modelling and optimisation of geometallurgical indices. In Proceedings of the First AUSIMM International Geometallurgy Conference, Brisbane, QLD, Australia, 5–7 September 2011; pp. 247–261. [Google Scholar]
  4. Keeney, L.; Walters, S.G. A Methodology for geometallurgical mapping and orebody modelling. In Proceedings of the First AUSIMM International Geometallurgy Conference, Brisbane, QLD, Australia, 5–7 September 2011; pp. 217–225. [Google Scholar]
  5. Montoya, P.A.; Keeney, L.; Jahoda, R.; Hunt, J.; Berry, R.; Drews, U.; Chamberlain, V.; Leichliter, S. Techniques applicable to prefeasibility projects—La Colosa case study. In Proceedings of the First AUSIMM International Geometallurgy Conference, Brisbane, QLD, Australia, 5–7 September 2011; pp. 103–111. [Google Scholar]
  6. Sepulveda, E.; Dowd, P.A.; Xu, C.; Addo, E. Multivariate Modelling of Geometallurgical Variables by Projection Pursuit. Math. Geosci. 2016, 49, 121–143. [Google Scholar] [CrossRef]
  7. Wirfiyata, F.; Mccaffery, K. Applied geo-metallurgical characterisation for life of mine throughput prediction at Batu Hijau. In Proceedings of the International Autogneous Grinding, Semiautogenous Grinding and High Pressure Grinding Roll Technology, Westmount, QC, USA, 25–28 September 2011; pp. 1–22. [Google Scholar]
  8. Johnson, C.L.; Browning, D.A.; Pendock, N.E. Hyperspectral Imaging Applications to Geometallurgy: Utilizing Blast Hole Mineralogy to Predict Au-Cu Recovery and Throughput at the Phoenix Mine, Nevada. Econ. Geol. 2019, 114, 1481–1494. [Google Scholar] [CrossRef]
  9. Lishchuk, V.; Lund, C.; Ghorbani, Y. Evaluation and comparison of different machine-learning methods to integrate sparse process data into a spatial model in geometallurgy. Miner. Eng. 2019, 134, 156–165. [Google Scholar] [CrossRef]
  10. Fernandes, F.G.; Cabral, I.E. Regression model utilization to estimate the mass recovery of a phosphate mine in Brazil. Rem Rev. Esc. Minas 2016, 69, 75–77. [Google Scholar] [CrossRef]
  11. Vieira, M.C.A. Metodologia Para Prever Recuperação de Zinco em Planta de Beneficiamento. Master’s Thesis, Programa de Pós-Graduação em Engenharia de Minas, Metalurgia e de Materiais, Universidade Federal do Rio Grande do Sul, Porto Alegre, Brazil, 2016. [Google Scholar]
  12. Niquini, F.G.F.; Costa, J.F.C.L. Comparison between regression models and neural networks Applied to forecast geometallurgical variables. In Mining Goes Digital, Proceedings of the 39th International Symposium ’Application of Computers and Operations Research in the Mineral Industry’ (APCOM 2019), Wroclaw, Poland, 4–6 June 2019; CRC Press: Boca Raton, FL, USA, 2019; pp. 463–472. [Google Scholar]
  13. Andrade, L.B.; Cabral, I.E.; Costa, J.F.C.L. Modelling and forecasting geometallurgical recovery at a phosphate mine. In Mining Goes Digital, Proceedings of the 39th International Symposium ’Application of Computers and Operations Research in the Mineral Industry’ (APCOM 2019), Wroclaw, Poland, 4–6 June 2019; CRC Press: Boca Raton, FL, USA, 2019; pp. 551–556. [Google Scholar]
  14. Niquini, F.G.F.; Costa, J.F.C.L. Mass and Metallurgical Balance Forecast for a Zinc Processing Plant Using Artificial Neural Networks. Nat. Resour. Res. 2020, 29, 3569–3580. [Google Scholar] [CrossRef]
  15. Niquini, F.G.F.; Costa, J.F.C.L. Forecasting mass and metallurgical balance at a gold processing plant using modern multivariate statistics. REM -Int. Eng. J. 2020, 73, 571–578. [Google Scholar] [CrossRef]
  16. Knight, R.; Olson Hoal, K.; Abraham, A.P.G. Three-dimensional geometallurgical data integration for predicting concentrate quality and tailings composition in a massive sulfide deposit. In Proceedings of the First AUSIMM International Geometallurgy Conference, Brisbane, QLD, Australia, 5–7 September 2011; pp. 227–232. [Google Scholar]
  17. Louwrens, E.; Napier-Munn, T.; Keeney, L. Geometallurgical characterisation of a tailings storage facility—A novel approach. In Proceedings of the Tailings and Mine Waste Management for the 21st Century, Sydney, Australia, 27–28 July 2015; pp. 125–131. [Google Scholar]
  18. Boisvert, J.B.; Rossi, M.E.; Ehrig, K.; Deutsch, C.V. Geometallurgical Modeling at Olympic Dam Mine, South Australia. Math. Geosci. 2013, 45, 901–925. [Google Scholar] [CrossRef]
  19. Macmillan, E.; Ehrig, K.; Liebezeit, V.P.; Kittler, P.; Lower, C. Use of geometallurgy to predict tailings leach acid consumption at Olympic Dam. In Proceedings of the First AUSIMM International Geometallurgy Conference, Brisbane, QLD, Australia, 5–7 September 2011; pp. 93–102. [Google Scholar]
  20. Hair, J.F.; Black, W.C.; Babin, B.J.; Anderson, R.E.; Tatham, R.L. Multivariate Data Analysis, 6th ed.; Pearson University Press: Upper Saddle River, NJ, USA, 2006. [Google Scholar]
  21. Pearson, K. On lines and planes of closest fit to systems of points in space. Lond. Edinb. Dublin Philos. Mag. J. Sci. 1901, 2, 559–572. [Google Scholar] [CrossRef]
  22. Switzer, P.; Green, A.A. Min/Max Autocorrelation Factors for Multivariate Spatial Imaging: Technical Report No. 6; Department of Statistics, Stanford University: Stanford, CA, USA, 1984; 14p. [Google Scholar]
  23. Long, C.A.; Sokal, R.R.; Sneath, P.H.A. Principles of Numerical Taxonomy. J. Mammal. 1965, 46, 111–112. [Google Scholar] [CrossRef]
  24. King, B. Step-Wise Clustering Procedures. J. Am. Stat. Assoc. 1967, 62, 86–101. [Google Scholar] [CrossRef]
  25. Chavent, M.; Genuer, R.; Saracco, J. Combining clustering of variables and feature selection using random forests. Commun. Stat. 2016, 50, 426–445. [Google Scholar] [CrossRef]
  26. Park, C.H. A Feature Selection Method Using Hierarchical Clustering. In Mining Intelligence and Knowledge Exploration; Lecture Notes in Computer, Science; Prasath, R., Kathirvalavakumar, T., Eds.; Springer: Cham, Switzerland, 2014; Volume 8284. [Google Scholar]
  27. Derksen, S.; Keselman, H.J. Backward, forward and stepwise automated subset selection algorithms: Frequency of obtaining authentic and noise variables. Br. J. Math. Stat. Psychol. 1992, 45, 265–282. [Google Scholar] [CrossRef]
  28. Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene Selection for Cancer Classification using Support Vector Machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
  29. Hastie, H.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning. Data Mining, Inference, and Prediction; Springer Series in Statistics Springer New York Inc.: New York, NY, USA, 2009. [Google Scholar]
  30. Both, C.; Dimitrakopoulos, R. Applied Machine Learning for Geometallurgical Throughput Prediction—A Case Study Using Production Data at the Tropicana Gold Mining Complex. Minerals 2021, 11, 1257. [Google Scholar] [CrossRef]
  31. Gholami, A.; Asgari, K.; Khoshdast, H.; Hassanzadeh, A. A hybrid geometallurgical study using coupled Historical Data (HD) and Deep Learning (DL) techniques on a copper ore mine. Physicochem. Probl. Miner. Process. 2022, 58, 147841. [Google Scholar] [CrossRef]
  32. Jorjani, E.; Bagherieh, A.H.; Mesroghli, S.; Chehreh Chelgani, S.; Jorjani, H. Prediction of yttrium, lanthanum, cerium and neodymium leaching recovery from apatite concentrate using artificial neural networks. In Proceedings of the MetPlant 2008, Perth, Australia, 18–19 August 2008; pp. 433–446. [Google Scholar]
  33. Nakhaei, F.; Mosavi, M.R.; Sam, A.; Vaghei, Y. Recovery and grade accurate prediction of pilot plant flotation column concentrate: Neural network and statistical techniques. Int. J. Miner. Process. 2012, 110–111, 140–154. [Google Scholar] [CrossRef]
  34. Srivastava, V.; Bandopadhyay, S.; Akdogan, G.; Ghosh, T. Dynamic neural network based model for SAG mill power prediction. In Proceedings of the 38th International Symposium on the Application of Computers and Operations Research in the Mineral Industry, Golden, CO, USA; 2017; Volume 38, pp. 227–233. [Google Scholar]
  35. Panda, L.; Tripathy, S.K. Performance prediction of gravity concentrator by using artificial neural network-a case study. Int. J. Min. Sci. Technol. 2014, 24, 461–465. [Google Scholar] [CrossRef]
  36. Ford, M.A. Simulation of Ore dressing Plants. Ph.D. Thesis, University of the Witwatersrand, Johannesburg, South Africa, 1979. [Google Scholar]
  37. King, R.P. Flotation. In Modeling and Simulation of Mineral Processing Systems, 2nd ed.; Schneider, C.L., King, E.A., Eds.; SME: Englewood, CO, USA, 2012; pp. 341–415. [Google Scholar]
  38. Sutherland, K.L. Physical chemistry of flotation XI. Kinetics of the flotation process. J. Phys. Chem. 1948, 52, 394–425. [Google Scholar] [CrossRef]
  39. Flint, L.R.; Howarth, W.J. The collision efficiency of small particles with spherical air bubbles. Chem. Eng. Sci. 1971, 26, 115–168. [Google Scholar] [CrossRef]
  40. Reay, D.; Ratcliffe, G.A. Removal of fine particles from water by dispersed air flotation: Effect of bubble size and particle size on collision efficiency. Can. J. Chem. Eng. 1973, 51, 178–185. [Google Scholar] [CrossRef]
  41. Yoon, R.H.; Luttrell, G.H. The effect of bubble size on fine particle flotation. Miner. Process. Extr. Metall. Rev. 1989, 5, 101–122. [Google Scholar] [CrossRef]
  42. Clift, R.; Grace, J.R.; Weber, M.E. Bubbles, Drops and Particles; Academic Press: Cambridge, MA, USA, 1978. [Google Scholar]
  43. Schulze, H.J. Physico-Chemical Elementary Processes in Flotation; Elsevier: Amsterdam, The Netherlands, 1993. [Google Scholar]
  44. Karamanev, D.G.; Nikolov, L.N. Free rising spheres do not obey Newton’s Law for free settling. AIChE J. 1992, 38, 1843–1846. [Google Scholar] [CrossRef]
  45. Tomlinson, H.S.; Fleming, M.G. Flotation Rate Studies. Proc. 6th International Mineral Processing Congress; Roberts, A., Ed.; Pergamon Press: Oxford, UK, 1965. [Google Scholar]
  46. King, R.P.; Hatton, T.A.; Hulbert, D.G. Bubble loading during flotation. Trans. Instn. Min. Metall. 1996, 83, C112–C115. [Google Scholar]
  47. Bradshaw, D.J.; O’Connor, C.T. Measurement of the sub-process of bubble loading in flotation. Miner. Eng. 1996, 9, 443–448. [Google Scholar] [CrossRef]
  48. Trahar, W.J.; Warren, L.J. The floatability of very fine particles—A review. Int. J. Miner. Process. 1976, 3, 103–131. [Google Scholar] [CrossRef]
  49. Drzymala, J. Characterization of materials by Hallimond tube flotation. Part 2: Maximum size of floating particles and contact angle. Int. J. Miner. Process. 1994, 42, 153–168. [Google Scholar] [CrossRef]
  50. Murphy, D.G.; Zimmerman, W.; Woodburn, E.T. Kinematic model of bubble motion in a flotation froth. Powder Technol. 1996, 87, 3–12. [Google Scholar] [CrossRef]
  51. Neethling, S.J.; Cilliers, J.J. A visual kinematic model of flowing foams incorporating coalescence. Powder Technol. 1999, 101, 249–256. [Google Scholar] [CrossRef]
  52. Woolley, A.R. Alkaline Rocks and Carbonatites of the World. Part 1: North and South America; British Museum: London, UK, 1987; 216p. [Google Scholar]
  53. Brod, J.A. Petrology and Geochemistry of the Tapira Alkaline Complex, Minas Gerais State, Brazil. Ph.D. Thesis, Durham University, Durham, UK, 1999. [Google Scholar]
  54. Ribeiro, C.C. Geologia, Geometalurgia, Controles e Gênese dos Depósitos de Fósforo, Terras Raras e Titânio do Complexo Carbonatítico Catalão I, GO. Ph.D. Thesis, Universidade de Brasília, Brasília, Brazil, 2008. [Google Scholar]
  55. Grasso, C.B. Geologia e Geometalurgia dos Minérios de Fosfato na Base do Manto de Intemperismo da Mina do Barreiro, Araxá. Ph.D. Thesis, Universidade de Brasília, Brasília, Brazil, 2005. [Google Scholar]
  56. Gaudin, A.M. Quantifying Concentrating Operations. In Principles of Mineral Dressing, 1st ed.; McGraw-Hill Book Company, Inc.: New York, NY, USA; London, UK, 1939. [Google Scholar]
  57. Tarján, G. Fundamentals and genarl principles of Separation. In Mineral Processing Volume 1, 1st ed.; Akadémiai Kiadó: Magyarország, Budapest, 1981. [Google Scholar]
  58. Misra, P.; Yadav, A.S. Improving the Classification Accuracy using Recursive Feature Elimination with Cross-Validation. Int. J. Emerg. Technol. 2020, 11, 659–665. [Google Scholar]
  59. Kavitha, S.; Varuna, S.; Ramya, R. A comparative analysis on linear regression and support vector regression. In Proceedings of the 2016 Online International Conference on Green Engineering and Technologies (IC-GET), Online, 19 November 2016. [Google Scholar]
  60. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  61. Dorogush, A.V.; Ershov, E.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar]
  62. Breiman, L.; Friedman, J. Estimating Optimal Transformations for Multiple Regression and Correlation. J. Am. Stat. Assoc. 1985, 80, 580–598. [Google Scholar] [CrossRef]
  63. Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  64. Izenman, A.J. Modern Multivariate Statistical Techniques; Springer: New York, NY, USA, 2008. [Google Scholar]
  65. Molnar, C. Interpretable Machine Learning, 2nd ed.; Lulu.com: Morrisville, NC, USA, 2020. [Google Scholar]
  66. Bishop, C.M. Neural Networks for Pattern Recognition; Clarendon Press: Oxford, UK, 1995. [Google Scholar]
Figure 1. Plant flowsheet.
Figure 1. Plant flowsheet.
Minerals 13 00748 g001
Figure 2. Example of a neural network with three input nodes and two output nodes.
Figure 2. Example of a neural network with three input nodes and two output nodes.
Minerals 13 00748 g002
Figure 3. Results from predictive performances for all RFECV core algorithms.
Figure 3. Results from predictive performances for all RFECV core algorithms.
Minerals 13 00748 g003
Figure 4. Features importance for the best model (10 inputs).
Figure 4. Features importance for the best model (10 inputs).
Minerals 13 00748 g004
Figure 5. Features importance for the reduced model (six inputs).
Figure 5. Features importance for the reduced model (six inputs).
Minerals 13 00748 g005
Figure 6. Absolute percentage error (APE) for the test set.
Figure 6. Absolute percentage error (APE) for the test set.
Minerals 13 00748 g006
Table 1. Methods used to reduce the number of variables.
Table 1. Methods used to reduce the number of variables.
MethodCore IdeaDisadvantagesReference
Super SecondaryLinearly combine several input variables into a single variable
-
Possible loss of interpretability of the models that are generated;
-
Requires previous knowledge to determine which variables are to be considered for amalgamation;
-
High number of samples needed to avoid the attribution of inflated weights caused by poorly conditioned matrices.
[18]
PCA and MAFCombination of variables
-
Difficulty in interpreting the resulting models that use the factors;
-
Total variability of data is not considered if not all the components are used.
[21,22]
Agglomerative hierarchical cluster analysisGroup variables instead of samples by similarity
-
Demands time for interpreting the results.
[23,24,25,26]
Stepwise Methods (forward, backward and stepwise)Find the subset of variables that minimizes the residual sum of squares
-
Do not ensure that collinearity problems will not occur in models and that irrelevant variables will not be part of the selected subset.
[27]
Table 2. ROM descriptive statistics.
Table 2. ROM descriptive statistics.
VariableNMeanStd. Dev.MinimumQ1MedianQ3Maximum
A l 2 O 3 ( % )30564.422.150.063.154.155.2534.57
BaO (%)30560.340.640.010.150.240.3827.21
CaO (%)305613.654.390.0210.9113.4616.1542.89
F e 2 O 3 ( % ) 305627.337.217.5922.6826.5031.1268.64
MgO (%)30564.092.500.012.173.865.6117.85
N b 2 O 5 ( % ) 30560.140.260.010.060.110.1712.00
P 2 O 5 ( % ) 30569.743.080.367.809.4611.3431.63
S i O 2 (%)305623.028.051.3217.8423.0827.9088.40
T i O 2 ( %)30568.454.110.135.747.6410.4139.31
Collector dosage (g/t)3056627.80751.3020.00200.00350.00700.005000.00
Table 3. Dependent variables descriptive statistics. Values in %.
Table 3. Dependent variables descriptive statistics. Values in %.
Variable (%)NMeanStd. Dev.MinimumQ1MedianQ3Maximum
YSlimes305642.8510.104.0936.0742.3749.5281.62
YMagnetics305610.798.090.044.749.2514.9568.35
YConcentrate305613.746.940.348.4713.1418.1052.55
YTailings305632.6111.861.9123.1932.1741.5268.84
RSlimes+Magnetics305634.8110.251.8628.0233.9341.1182.51
RConcentrate305647.8116.280.6836.9949.9860.2092.08
RTailings305617.3813.290.007.2013.3024.6975.10
Table 4. Global MAPE values observed for the differing number of input variables.
Table 4. Global MAPE values observed for the differing number of input variables.
MetricCatboost Algortithm with
10 Variables6 Variables5 Variables4 Variables
Global MAPE6.076.366.506.75
Table 5. Increase in global MAPE when subsets of variables were used.
Table 5. Increase in global MAPE when subsets of variables were used.
ScenarioCatboost Algortithm withIncrease In Global MAPE (%) Compared Agaist the Best Case
16 variables4.6
25 variables7.1
34 variables11.1
Table 6. ANN metrics using 10 input variables.
Table 6. ANN metrics using 10 input variables.
VariableSetMean ± Std. Error
CorrelationMAERMSEMdAPE
YConcentrateTraining0.854 ± 0.0012.664 ± 0.0133.599 ± 0.01314.938 ± 0.2
YSlimes0.608 ± 0.0026.241 ± 0.0218.042 ± 0.03312.297 ± 0.067
YMagnetics0.855 ± 0.0013.05 ± 0.0134.156 ± 0.01525.445 ± 0.16
YTailings0.837 ± 0.0015.158 ± 0.0366.534 ± 0.03513.711 ± 0.077
RConcentrate0.667 ± 0.0029.296 ± 0.03612.174 ± 0.04414.564 ± 0.17
RSlimes+Magnetics0.456 ± 0.0026.964 ± 0.0169.149 ± 0.02216.155 ± 0.048
RTailings0.688 ± 0.0016.941 ± 0.0319.596 ± 0.02236.768 ± 0.231
YConcentrateValidation0.849 ± 0.0082.708 ± 0.063.641 ± 0.08215.482 ± 0.473
YSlimes0.585 ± 0.0176.35 ± 0.0788.272 ± 0.13412.239 ± 0.233
YMagnetics0.847 ± 0.0093.115 ± 0.0674.28 ± 0.13525.999 ± 0.727
YTailings0.83 ± 0.0065.24 ± 0.0776.664 ± 0.0913.909 ± 0.26
RConcentrate0.655 ± 0.0149.432 ± 0.20812.291 ± 0.25314.774 ± 0.416
RSlimes+Magnetics0.435 ± 0.0177.059 ± 0.1369.236 ± 0.1816.348 ± 0.644
RTailings0.676 ± 0.0137.065 ± 0.1919.728 ± 0.19137.596 ± 1.316
YConcentrateTest0.8662.6313.60114.897
YSlimes0.5836.3898.41612.449
YMagnetics0.8373.2764.60227.851
YTailings0.8335.3536.74114.263
RConcentrate0.6259.47912.84313.925
RSlimes+Magnetics0.4247.0739.27616.199
RTailings0.6587.18110.42034.133
Table 7. ANN metrics using six input variables.
Table 7. ANN metrics using six input variables.
VariableSetMean ± Std. Error
CorrelationMAERMSEMdAPE
YConcentrateTraining0.855 ± 0.0012.664 ± 0.0253.598 ± 0.01815.122 ± 0.32
YSlimes0.599 ± 0.0016.334 ± 0.0228.093 ± 0.02112.523 ± 0.062
YMagnetics0.846 ± 0.0013.195 ± 0.0194.319 ± 0.01826.523 ± 0.149
YTailings0.835 ± 0.0015.165 ± 0.0216.528 ± 0.02513.692 ± 0.072
RConcentrate0.667 ± 0.0029.32 ± 0.06912.221 ± 0.0614.601 ± 0.302
RSlimes+Magnetics0.463 ± 0.0026.957 ± 0.029.121 ± 0.03116.231 ± 0.082
RTailings0.692 ± 0.0026.902 ± 0.0559.561 ± 0.02836.608 ± 0.457
YConcentrateValidation0.851 ± 0.0082.706 ± 0.0563.634 ± 0.07515.564 ± 0.346
YSlimes0.582 ± 0.0126.412 ± 0.0658.205 ± 0.09212.732 ± 0.23
YMagnetics0.84 ± 0.0073.225 ± 0.0734.372 ± 0.09826.718 ± 0.788
YTailings0.829 ± 0.0065.236 ± 0.0716.629 ± 0.09713.855 ± 0.281
RConcentrate0.657 ± 0.0129.385 ± 0.20712.32 ± 0.26114.674 ± 0.404
RSlimes+Magnetics0.447 ± 0.0177.002 ± 0.1359.168 ± 0.18116.463 ± 0.276
RTailings0.682 ± 0.0116.985 ± 0.1339.659 ± 0.17436.91 ± 0.964
YConcentrateTest0.8592.6713.65016.169
YSlimes0.5866.4708.43712.656
YMagnetics0.8383.3124.59928.397
YTailings0.8275.4906.85414.806
RConcentrate0.6349.39512.63715.157
RSlimes+Magnetics0.4437.0069.19116.189
RTailings0.6657.24410.23334.327
Table 8. Relative differences between the sub-optimum scenario and the optimum.
Table 8. Relative differences between the sub-optimum scenario and the optimum.
VariableSetRelative Difference (%)
CorrelationMAERMSEMdAPE
YConcentrateTraining0.1%0.0%0.0%1.3%
YSlimes−1.5%1.5%0.6%1.9%
YMagnetics−1.1%4.8%3.9%4.2%
YTailings−0.2%0.1%−0.1%−0.1%
RConcentrate0.0%0.3%0.4%0.3%
RSlimes+Magnetics1.5%−0.1%−0.3%0.5%
RTailings0.6%−0.6%−0.4%−0.4%
YConcentrateValidation0.2%−0.1%−0.2%0.5%
YSlimes−0.5%1.0%−0.8%4.1%
YMagnetics−0.8%3.5%2.1%2.8%
YTailings−0.1%−0.1%−0.5%−0.4%
RConcentrate0.3%−0.5%0.2%−0.7%
RSlimes+Magnetics2.8%−0.8%−0.7%0.7%
RTailings0.9%−1.1%−0.7%−1.8%
YConcentrateTest−0.7%1.5%1.4%8.5%
YSlimes0.3%1.3%0.3%1.7%
YMagnetics0.1%1.1%−0.1%1.9%
YTailings−0.7%2.6%1.7%3.8%
RConcentrate1.4%−0.9%−1.6%8.8%
RSlimes+Magnetics4.5%−1.0%−0.9%−0.1%
RTailings1.1%0.9%−1.8%0.6%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Niquini, F.G.F.; Branches, A.M.B.; Costa, J.F.C.L.; Moreira, G.d.C.; Schneider, C.L.; Araújo, F.C.d.; Capponi, L.N. Recursive Feature Elimination and Neural Networks Applied to the Forecast of Mass and Metallurgical Recoveries in A Brazilian Phosphate Mine. Minerals 2023, 13, 748. https://doi.org/10.3390/min13060748

AMA Style

Niquini FGF, Branches AMB, Costa JFCL, Moreira GdC, Schneider CL, Araújo FCd, Capponi LN. Recursive Feature Elimination and Neural Networks Applied to the Forecast of Mass and Metallurgical Recoveries in A Brazilian Phosphate Mine. Minerals. 2023; 13(6):748. https://doi.org/10.3390/min13060748

Chicago/Turabian Style

Niquini, Fernanda Gontijo Fernandes, André Miranda Brito Branches, João Felipe Coimbra Leite Costa, Gabriel de Castro Moreira, Claudio Luiz Schneider, Florence Cristiane de Araújo, and Luciano Nunes Capponi. 2023. "Recursive Feature Elimination and Neural Networks Applied to the Forecast of Mass and Metallurgical Recoveries in A Brazilian Phosphate Mine" Minerals 13, no. 6: 748. https://doi.org/10.3390/min13060748

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop