Next Article in Journal
Correction: Ghnatios et al. A Regularized Real-Time Integrator for Data-Driven Control of Heating Channels. Computation 2022, 10, 176
Previous Article in Journal
Creation of a Simulated Sequence of Dynamic Susceptibility Contrast—Magnetic Resonance Imaging Brain Scans as a Tool to Verify the Quality of Methods for Diagnosing Diseases Affecting Brain Tissue Perfusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Group Classification for the Search and Identification of Related Patterns Using a Variety of Multivariate Techniques

by
Nisa Boukichou-Abdelkader
1,2,
Miguel Ángel Montero-Alonso
3,* and
Alberto Muñoz-García
4
1
School of Doctorate in Sciences, Technologies and Engineering, University of Granada, 18012 Granada, Spain
2
Data Science Unit, Health Innovation of La Rioja, Rioja Health Foundation, CIBIR, 26006 Logroño, Spain
3
Department of Statistic and Operational Research, University of Granada, 18016 Granada, Spain
4
Department of Statistic, University Carlos III of Madrid, 28903 Madrid, Spain
*
Author to whom correspondence should be addressed.
Computation 2024, 12(3), 55; https://doi.org/10.3390/computation12030055
Submission received: 27 January 2024 / Revised: 16 February 2024 / Accepted: 2 March 2024 / Published: 9 March 2024

Abstract

:
Recently, many methods and algorithms have been developed that can be quickly adapted to different situations within a population of interest, especially in the health sector. Success has been achieved by generating better models and higher-quality results to facilitate decision making, as well as to propose new diagnostic procedures and treatments adapted to each patient. These models can also improve people’s quality of life, dissuade bad health habits, reinforce good habits, and modify the pre-existing ones. In this sense, the objective of this study was to apply supervised and unsupervised classification techniques, where the clustering algorithm was the key factor for grouping. This led to the development of three optimal groups of clinical pattern based on their characteristics. The supervised classification methods used in this study were Correspondence (CA) and Decision Trees (DT), which served as visual aids to identify the possible groups. At the same time, they were used as exploratory mechanisms to confirm the results for the existing information, which enhanced the value of the final results. In conclusion, this multi-technique approach was found to be a feasible method that can be used in different situations when there are sufficient data. It was thus necessary to reduce the dimensional space, provide missing values for high-quality information, and apply classification models to search for patterns in the clinical profiles, with a view to grouping the patients efficiently and accurately so that the clinical results can be applied in other research studies.
MSC Classification:
65C20; 65C60; 68W40

1. Introduction

Large amounts of information are currently generated every day in all areas of knowledge, without any evaluation of the quality of the information that is stored. This frequently occurs in the health sector, which requires the rapid structuring and organization of the data created and stored in large repositories.
The goal is to give them greater value through the transformation and generation of high-quality data for the development of reliable research that can facilitate decision making for the improvement of therapies, treatments, and diagnoses, and the prevention of chronic diseases, which facilitate a better quality of life for patients [1].
Recent advances in bioinformatics propose the use of Artificial Intelligence (AI) and Machine Learning (ML) algorithms as the basis for analyses. The construction and improvement of tested and reliable models that help to prevent and recognize these diagnoses sufficiently in advance are also relevant. These advances make it possible to identify, explain, and group the patterns that support various hypotheses in ongoing research or to solve a specific case.
There is a wide range of statistical methods [2] that have had a great impact on the field of Computational Statistics. These include cluster analysis and classification techniques [3,4] that can be used to separate groups with similar characteristics. Correspondence Analysis (CA) [5] separates groups without considering affinities but is an excellent support for global visualization in clinical cases. Moreover, Decision Trees (DT) [6,7] are a valuable tool for the detection of an optimal classification of different classes of variables. In a DT, the leaves of a tree are shown as outputs in various layers or levels that are possible nodes of the results. This is achieved by applying a small set of hierarchical decision rules that result in the final decision, which is displayed in the form of a tree.
Furthermore, this splitting rule, which measures the quality of a partition, can be performed by using two types of measure: (i) a default measure known as the “Gini Index” [8] that measures impurity; and (ii) “entropy” that measures information gain.
For this purpose, the combination of the previous values is minimized or maximized (entropy or Gini, respectively). This means that all the categories or instances of a class can join the positive part of the condition, while the rest remain in the opposite (negative) part. Therefore, this process only considers a node to be pure if 100% of the instances of the node correspond to a specific category of target variable (class).
The learning process of this classification mainly focuses on building a function capable of distinguishing between different output classes or groups represented by a specific use case. The separation function is known as the discriminant function and uses the values of the input variables defined by the use case itself, which is studied by creating a separation boundary. In this sense, there are different types of discriminant function that help to determine the different areas of the space in which each of the classes in the same case belongs. This is accomplished by using simple functions, such as linear ones, or other, more complex polynomial functions.
Data mining functionalities, such as cluster analysis, become more complex as the number of dimensions increases, even though their only objective is to extract profiles or patterns from large volumes of data using efficient algorithmic processes that help to analyze large data records.
In short, all these Machine Learning techniques, together with the support of data mining, are used to evaluate or measure the quality of the estimated models using different methods, such as the K-means algorithm; the K-Nearest Neighbor or KNN; and metrics (classification rate or success rate through the confusion matrix). Other metrics are also available, depending on the level of measurement and quality specified in the models that are built and adjusted [9,10,11,12,13].
The information provided by these measurements supports the results of the search and identification of the best grouping option that can classify the patients that make up the dataset. The objective is to find any behavioral pattern that separates them into different sections of a related clinical profile based on their characteristics. Therefore, additional support is required to solve the difficulties in this area. This can be achieved thanks to the analyses performed, which, in turn, improve and complete the decision-making process.
The research study presented in this paper used computational learning methods to facilitate the search for patterns in high-dimensional datasets, where it was necessary to apply classification procedures to provide added value to the clinical information collected. In addition, filtering mechanisms were used to select the most relevant data to validate decision making, since these algorithms would probably have limitations when reproduced in other models with the same clinical approach.

2. Materials and Methods

This research identified and searched for clinical profiles using different classification methods to group different types of patients based on their characteristics and affinities, with a view to finding the optimal pattern model for the clinical data.
The approach was carried out in various stages using the database cited in [14]. The sample obtained was of 5178 hospitalized patients, who were suffering from an exacerbation of chronic obstructive pulmonary disease (COPD). From this sample, 32 epidemiological, clinical, and outcome variables were selected, and optimal classification of the patients was achieved.
Firstly, the data were represented in 2D, since the first two axes are the ones that most contribute to the variability. The next step was to apply supervised and unsupervised analysis techniques for the classification of the variables in the dataset by applying different testing mechanisms (Cluster, Correspondence (CA) and Decision Trees (DT)).
The methods applied for the separation and fusion of the groups were the following: (1) Complete Linkage, minimizing the maximum distance; (2) simple linkage, minimizing the minimum distance; (3) average, calculating the average distance; and (4) Ward, obtaining the minimum variance, in order to select the best among them.
The best classification method was found to be cluster analysis, which led to the creation of three optimal final groups [14]. Analytical exploration ended with a description of the grouping or classification pattern for each of the group profiles. This was accomplished by highlighting their characteristics, with a view to specifying the result obtained through the optimal identification of separation, as represented by the final clinical profiles generated for this case.
However, even though other techniques for evaluating results were analyzed, in this study, our main focus was on cluster analysis since it is the best method to obtain our objectives. This is because the primary aim of cluster analysis is to classify individuals by forming groups or clusters in such a way that the patients within each cluster present more homogeneity in terms of the values adopted for each type of variable.
It goes without saying that in this type of technique, the concept of spatial distance is relevant since it defines the degree of similarity or dissimilarity (in terms of distance ratio) in each of the groupings. In other words, one observation is more similar to another when the distance separating them is as small as possible (i.e., it has a minimum distance), or when the similarity is at its maximum.
For this case, although there are different distance methods (such as Euclidean, Levenshtein’s, or Jaccard’s coefficients), this study used Euclidean distance to evaluate the grouping patterns, which are key to this approach. In parallel, it was also necessary to simplify the analysis of the characteristics of each valuable, and thus give added value to the analytical exploration of the groups formed. This provided a higher level of quality and validity to the final profiles generated.
Cluster analysis methods are based on the concept of hierarchy [15] and can be either hierarchical or non-hierarchical. Hierarchical methods are composed of two types: (i) associative or agglomerative, and (ii) dissociative or divisive. Non-hierarchical methods are based on partitioning [16].
The approaches to clustering methods include the following: (i) hierarchical algorithms that create a hierarchical decomposition of the dataset based on some kind of criteria, and whose main objective is to build a dendrogram in the form of a tree; (ii) partitioning algorithms, which build different partitions by evaluating them according to the selected criterion, such as the K-means procedure, where the initial clustering parameter must be defined; (iii) methods based on connectivity and density functions; (iv) methods based on grids using a multi-level granularity structure; and finally, (v) methods based on different cluster models, one of which must be chosen as the best adjustable model that achieves the optimum result.
For our study, there was thus a large amount of available information as well as a great variety of Machine Learning techniques that could be used in the clustering algorithm of clinical data stored for classification. To identify the models of related patterns within a dataset, the bases of the desired model must be established and correctly defined a priori. Only in this way can different optimal solutions be generated, and the best one found reflects all the individuals of the data explored.
Therefore, as previously mentioned, classification methods can be either non-hierarchical or hierarchical.
Non-hierarchical methods can be of different types [15,16,17] depending on the selected criterion, such as the K-means procedure, K-medoids, and Partitioning around Medoids (PAM). The optimal estimation measures include the Elbow method, Silhouette index, and Calinski–Harabasz index. The other function-based methods are based on density functions (DBSCAN algorithm), connectivity, or grids, not to mention those models based on the search for the optimum.
In contrast, hierarchical methods [18,19,20,21] can be (i) agglomerative or associative (bottom-up) or (ii) divisive or dissociative (top-down). Both use different techniques, such as Nearest Neighbor (Single Linkage); Farthest Neighbor (Complete Linkage); Average (Average Linkage); Ward Method (Minimum Variance); Centroid Method (Centroid Linkage); Median Method; and Associative analysis (pertaining to divisive methods). In addition, the divisive methods are Unidimensional (Monothetics) or Multidimensional (Polythetics).
To perform analysis, the statistical software R [22] was used to apply various classification analysis techniques, namely Cluster, CA [23], and DTs in order to find and optimize the search for patterns that best group these data.
Using the database cited in [14], our study analyzed a sample of 5178 hospital patients suffering from an exacerbation of chronic obstructive pulmonary disease (COPD). In this context, we selected 32 epidemiological, clinical, and outcome variables to optimize dimensional reduction and obtain a coherent classification of patients.
This selection of variables was carried out in the original database because it was the most clinically relevant to classify these individuals based on their characteristics and added risk levels due to advanced disease. This study also addressed the problem of dimensionality using different techniques to reduce the dimensional space of the set of data variables. Of the methods tested (i.e., Principal Components Analysis [PCA], Random Forest by the Gini index and Information Value by weight of evidence [RF&IV], and Parallel Analysis with simulated data and data resampling [PA-RES]), PCA was found to yield the best results [14].
At the same time, we searched for clinical profiles by applying different learning techniques [17,24] through supervised and unsupervised analysis to identify the best possible classification grouping fit for the data, without any loss of relevant information. Based on their internal and visual characteristics, various group profiles were generated. Since these were very similar within the group and different from each other with respect to the other groups, our analysis resulted in a set of patterns based on their own clinical properties.

3. Results

The three analytical procedures implemented and the results obtained using each of them are described as follows. We searched for and identified suitable clinical profiles, producing a classification grouping adjusted to these patient data, while simultaneously providing differential and additional values of segmentation analysis for decision making.

3.1. Unsupervised Cluster Analysis

In recent years, unsupervised clustering analysis has been a research focus in healthcare-related fields, such as Clinical Research, Engineering, particularly bioinformatics and Applied Science in general. It is a good alternative application that is used to find clustering solutions for different observations or individuals.
The application of cluster analysis or clustering of cluster analysis as unsupervised classification [18,25,26] was performed with the R cluster package [27]. This software is used to identify groups of similar individuals, discover the distribution of patterns and correlations in large datasets, and assess the quality of the resulting groups.
More specifically, cluster analysis was selected for our study because it is an unsupervised group classification technique, where the classes are not predefined. It segments a heterogeneous population into a number of homogeneous subgroups or partitions called clusters.
This study showed that cluster analysis was able to generate a set of clusters from the same dataset, which were combined into a single final cluster in order to improve the quality of the individual patient data clusters. Spatial clustering was determined by identifying subspaces for cluster formation and classifying the patient data into different segments or groups of clusters.
Nevertheless, conventional spatial clustering algorithms tend to explore dense groups in all possible subspaces. This can lead to what is known as the curse of dimensionality, which materializes with the increase in the number of dimensions and subspaces to explore. This, in turn, produces an exponential increase in the number of subspace clusters.
However, in our study, this problem was solved by previously using different approaches to reduce dimensionality and select the most important features. This avoided redundant information in the clusters created in the different subspaces [12,13,14].
Furthermore, the quality of clustering depends both on the similarity measure, which is usually a distance function (e.g., Euclidean distance for numerical attributes), and on its implementation. Since distance functions are very sensitive to the type of variables used (continuous, categorical nominal, or ordinal) and the range or scale with which they are measured, it was necessary to apply a normalization process to standardize all the variables to the same scale. This ensured that within the clustering formation process, all the variables have the same importance or weight, and also that there is no overfitting by any of them in the generation of the final groups.
As previously mentioned, there are different cluster analysis methods based on hierarchies [24,25]. For this reason, after the implementation of each hierarchical technique, using normalized data and the Euclidean distance measure, the cophenetic correlation coefficients between the distance matrix and the cophenetic matrix were calculated for each of them. The results obtained with each method were 0.53 for the complete one; 0.76 for the simple one; 0.74 for the average; and 0.29 for the Ward. The best method was cluster 2, with a cophenetic coefficient of 0.76, which corresponded to simple linkage. Nevertheless, there were a few differences between cluster 2 and the cluster 3 (average method), with a cophenetic coefficient of 0.74.
In addition, the previous visualization of the results obtained with PCA [14] (using the first two components) found four groups of profiles organized by affinity and characteristics for the information in this dataset. These profiles were classified into different groups of five, four, and three. In this way, we obtained different dendrograms or hierarchical trees of the cluster, as well as their factorial maps, which improved the visual representation of the groups formed by the individuals.
In this case, the classification of the four groups showed that these individuals could be grouped somewhat more closely because they were nearer to each other and probably quite similar internally. This was especially true for clusters 1 and 3, where a new mixed cluster group was generated in order to reduce to the maximum (optimum) number of final clusters. Figure 1 shows the hierarchical tree formed by the three new clusters (K = 3 is optimal), which are more homogeneous within the same profile group and more heterogeneous among themselves, which makes it easier to interpret the final profiles generated.
In contrast, as there is no additional information available for the optimal number of groups K (except that which was extracted using PCA, where there could be four groups, and the one reflected in the visual support of AC analysis, with three possible groups), the K-means algorithm was applied for a range of K values. The K value was thus changed to 20, 15, 10, and even 5 for the different methods. However, this was performed only for the two best cluster methods, depending on the cophenetic coefficient, Simple and Average. The goal was to identify that value from which the reduction in the total sum of intra-cluster variance was no longer substantial. For this purpose, we applied the Elbow strategy or method, with the function “fviz_nbclust” of the R package factoextra [28], which automates this process and generates a representation of the results that help to detect the inflection point of change. This facilitated the selection of the most effective number of clusters.
Therefore, the visual results of the K-means and the correlation coefficients showed that the best method was simple linkage (defined by clustering 2) of minimum distance, with 76.5%, which was composed of four groups. This indicated the possibility of generating a classification with an optimal number (K) of fewer than four groups. Similarly, final analysis revealed that the optimal number ranged from five to three clusters of groups. This signified that the number of clusters could be optimized to three cluster profiles.
Accordingly, the number of groups (clusters) was obtained by determining the optimal value (profiles) on the basis of criteria such as the Silhouette, Elbow, and Calinski–Harabasz methods or the K-means algorithm, which confirmed this final division value for different risk levels. To evaluate the classification performance, we used quality and accuracy metrics (with cut-offs greater than 0.80) to assess the clinical usefulness of the model.
In view of these data (K-means algorithm and Elbow method), a second method called Silhouette was applied to verify the results. Although Silhouette is very similar to Elbow, the difference lies in the fact that the average of the coefficients or Silhouette indices of all the observations of individuals is maximized (Figure 2). Therefore, the results obtained from the Silhouette curve show a definitive number of three optimal clusters representing the best separation of the data, which reaffirms what was already visualized in the previous outputs.
Finally, after also exploring CA, which is described as follows, we visualized the three output clusters of groups based on our results, as well as their high and low correlation levels in the optimal clusters generated of the variables representing the characteristics of individuals (Table 1 and Glossary). In this case, the sample comprised 5178 hospitalized patients suffering from an exacerbation of Chronic Obstructive Pulmonary Disease (eCOPD). The analysis of this sample was based on the differences in the stages and progression of disease, clinical pathologies, and the characteristics of the patients, which were extremely diverse.
The classification of the three groups of optimal profiles is described as follows.
Cluster 1 was formed by patients with high values for the following variables (in descending order from the highest): Spirometry FVC in % of theoretical (FVCP), Spirometry performed prior to admission or at discharge (SPIROMETRY_PA), Tobacco Habit (SMOKING_HABIT), Weight (WKG), and Duration of Hospital Admission (DUR_ADM). These patients also had low values for the following variables (in ascending order from the lowest): Systolic Blood Pressure (SBP), Diastolic Blood Pressure (DBP), Relational Ratio FEV1/FVC previous or discharge spirometry (FEVCVF), Heart Rate (HR), and FEV1 spirometry in % of theoretical (FEV1P).
Cluster 2 was formed by patients with high values for the following variables: Systolic Blood Pressure (SBP), Diastolic Blood Pressure (DBP), Heart Rate (HR), Spirometry performed prior to admission or at discharge (SPIROMETRY_PA), Ventilatory Support (VS), and Respiratory Rate (RR). These patients also had low values for the following variables: Weight (WKG), Body Mass Index (BMI), Duration of Hospital Admission (DUR_ADM), Deaths at 90 days (DEATH_90DAYS), Cerebrovascular Disease (CVD), and Cardiovascular Comorbidity (CCVSDM).
Cluster 3 was formed by patients with high values for the following variables: FEV1/FVC ratio spirometry prior to admission or discharge (FEVCVF), FEV1 spirometry in % of theoretical (FEV1P), AGE, Cerebrovascular Disease (CVD), Congestive Heart Failure (CHF), Vascular Disease (VD) and Cardiovascular Comorbidity (CCVSDM). They also had low values for the following variables: Spirometric FVC in % of theoretical (FVCP), Spirometry performed prior to admission or at discharge (SPIROMETRIA_PA_), Tobacco Habit (SMOKING_HABIT), Diastolic Blood Pressure (DBP), Heart Rate (HR), and Ventilatory Support (VS).

3.2. Correspondence Analysis

The Correspondence Analysis package “ca” [29] was used as a part of the main analysis (Cluster) to complete the result of the classification in groups. This visually descriptive multivariate technique represents the relationship between variables using an association measure in a low-dimensional space with the least possible loss of information [30,31]. In other words, it is a dimension reduction method, equivalent to PCA, Factor (FA) and Discriminant (DA) analysis for categorical, nominal or ordinal variables. It thus helps to visualize the multidimensional point cloud in two dimensions.
CA shows that the inertia of the first dimensions indicate whether there were strong relationships between the variables and suggest the number of dimensions that need to be studied, which provided additional support for our exploration.
Based on the results in Table 2, the first two dimensions expressed 38.19% of the total inertia of the dataset. This means that the total variability in the row (or column) cloud is explained by the 1:2 plane. Obviously, this is an intermediate, though substantial, percentage for these data. The first plane represents a part of the variability in the data, which is much larger than the reference value (7.45% equal to the 95th percentile of the distribution of inertia percentages obtained by simulating 101 tables of data of equivalent size on the basis of a uniform distribution). Based on these figures, the variability explained by this plane is very important.
In consonance with PCA and in the event that better results are needed, the dimensions that also express a high percentage of the total inertia could be considered. This would mean interpreting those greater than or equal to the third to complete the information.
The CA results indicated that an estimate of the correct number of axes to interpret should restrict the analysis to the description of the first eight axes, especially since they carry the vast majority of real information. They also showed an amount of inertia 76.65% higher than that obtained by the 95th percentile of random distributions (28.61%). The most relevant description was found to be located on these axes, the first two of which contribute more to the total inertia. More specifically, the contribution of the first axis was 24.13%, whereas that of the second one was 14.05%.
As confirmation, these axes can be visualized on a sedimentation graph. This is another way of viewing the correct final axes, along with which the optimal number of components is shown. Since the values above point 1.0 were the most acceptable, this verified that the eight previous components (eigenvalues greater than one) were still the ones with the majority of the valid information. The red dashed line of the screen plot also indicates what the contribution would be (in terms of the percentage of variability in each dimension) if they were homogeneous.
In addition, this describes dimensional contributions in the 1: 2 plane for the different variables (columns) in order of importance on the axis. As reflected, there are three variables with a high percentage of contribution to the axis. There are also two more variables (just the cut of the dashed line) with a lower percentage of contribution to the plane. However, they must also be taken into account in the analysis of this axis (1–2), as well as the rest of variables with less weight, but with the same relevance.
In parallel, a matrix of correlations between variables (strong–weak) was also visualized in the analysis outputs. This ensured that these were high as one of the fundamental requirements. This was achieved by dimensionally representing the contribution of each attribute variable on each axis, especially the 1:2 plane, and by highlighting dimensions 3 and 5, given their importance and despite their complex description.
Based on the results obtained with the different methods, the optimum number of groups was found in three profiles, which verified the outputs of cluster analysis.
As this procedure was an additional support for the main analysis (Cluster), we visualized the results on the 1:2 plane and the representations, both separately and as a whole, for the patients (Figure 3) and variables (cols). This revealed that the three groups were generated by similar characteristics within the same profile (patients or variables), as reflected by the degree of the color scale. In addition, there were factors of patients which were closely correlated and which summarized this axis (correlation quite close to 0.97).

3.3. Decision Tree Classification Analysis

As part of our search for optimal profiles, we also explored Decision Tree (DT) classification analysis [32,33,34] as another way to support the main analysis (Cluster). This analysis was performed to detect any relevant information or classifications hidden among the optimal clusters that could improve the final clustering result generated by the three groups of final profiles.
The Decision Tree (DT) method from R’s rpart package [21] is a supervised learning technique. It is similar to Support Vector Analysis (SVM), where both supervisory algorithms are based on the classification (or regression) of previously labeled data. The idea was to find the relationships between them using a function, which associated the input variables with the output variables to generate a decision tree able classify or predict the observations of individuals in the future [35,36].
DT differ from unsupervised procedures, such as Cluster analysis, CA, and PCA, which describe and find data structures that either define groups based on their similarity (distance) or simplify them to maintain their essential characteristics, and thus increase the quality of the data by improving the performance of these algorithms.
Decision Trees [37,38,39] are one of the most powerful algorithmic approaches currently used in the vast majority of Machine Learning research fields, especially in the fields of healthcare and data mining or data science, which search for different behaviors that can define efficient and more accurate sets of patterns, that are applicable to the general population.
This classification algorithm uses simple decision rules or dichotomous constraints, which are usually composed of two groups or classes as a part of its classificatory analysis. It maintains a hierarchical order in its application and builds a tree structure to represent it visually as a final result. This aids decision making by describing the data without determining the final resolution of them [8,40].
However, this process only keeps the attributes or categories that matter in decision making and discards those attributes that do not contribute to the accuracy of the tree. This information is thus of great value and can be used to reduce the data to the most important variables before applying another technique.
Given their accuracy and security, decision tree models are now widespread and are frequently used in many research fields since they are a simple, self-explanatory, and easily scalable technique. They can handle a wide variety of different input data variables with a minimal computational cost. Moreover, they are able to process datasets with missing values, which results in high predictive power with little computational effort. In fact, they can even handle noisy data using internal mechanisms (such as “pruning”) that reduce the depth of the tree in such a way as to obtain the best generalization. They can also support different functions depending on the objective, either for classification, regression, or clustering. Finally, they are also an additional technique for the variable (or category) selection procedure in the final screening of attributes (or variables).
Nevertheless, the DT method has certain limitations. Firstly, when a high-depth factor is applied, it may present problems of overlearning, which complicates the algorithmic process on the class categories. Secondly, it does not detect correlations between the variables, since each decision node is obtained independently, without taking into account the rest of the nodes or tests.
However, the Decision Tree (DT) technique was used in our study as an additional exploratory classification support in the search for patient profiles and to detect any changes in the data that might improve the final definition of the profile groups. That said, various cases (classes) were explored. Although only two of them were of clinical interest, they could provide valuable information pertaining to hospital admissions and smoking, offer an extra result, or simply confirm the existing data.
This procedure was started by setting a seed (“seed = 500”), which was subsequently varied (seed = 1000 or 1500) in order to detect significant changes. This function (seed), which is a number (or vector), was used to initialize the pseudo-random number generator, which helped to make the results of the models reproducible elsewhere. Furthermore, the dataset was divided into two parts, where 70% were training data and 30% were test data for the subsequent validation of the results obtained.
In this context and in view of the results of the first decision tree on smoking data, it was observed that 82% of the patients were smokers. Classification thus started with the gender variable (sex labeled as (1) female or (0) male) in the first section of the decision flow and ended with a spirometric test and an advanced age cut-off. This tree (whose visual presentation was quite blurry) concluded by showing the following results: (i) 12% of female smokers over 73 years of age (6%) had a predisposition to severity of 6%; and (ii) 88% of male smokers without spirometry performed (27%) had a predisposition to severity of 60%.
Unfortunately, this mapping problem is very common when the tree has a greater depth. This makes its decision rules more complicated, which leads to a very unclear tree structure.
As in the previous case, the results of the second decision tree on the hospital admission data showed that 35% of the patients underwent hospital admission. The classification started with the exacerbation readmission variable (labeled (1) Yes or (0) No) in the first leg of the decision flow and ended with a three-month mortality event check.
Even though the visual representation of this other tree was not very clean, it reflected conclusive results that should be taken into consideration, such as the following: (i) out of 73% of the patients not readmitted for exacerbation, only 2% were predisposed to die; and (ii) 27% of the patients who were readmitted died within 90 days of readmission for exacerbation because of the severity of the disease itself and repeated hospital readmissions without improvement after the recommended therapies.

4. Discussion

Based on the results obtained, cluster analysis was found to be a good choice since it is one of the best methods for data representation when it is a question of dividing groups that are homogeneous within themselves and which are heterogeneous among themselves. In our study, this technique obtained an optimal final result of three groups of clinical profiles, which were both different from each other and similar with related characteristics. This achieved our initial objective, which was to identify patterns of related profiles in order to classify the patients into groups with the same behavior. In this way, their properties or characteristics could be studied in detail, with a view to improving their clinical treatment and care and to extrapolating the results and conclusions of our study to the general population.
For this purpose, different techniques were analyzed for the description and visual representation of the data collected. In a parallel way, different metrics were also implemented to compare the results obtained, and at the same time, to select the best strategy to obtain high-quality information. The goal was to implement different dimensional reduction techniques without the loss of valuable data. We also wished to identify an effective algorithm for the creation of different groups in order to obtain optimal clinical profiles.
Correspondence Analysis (CA) and Decision Analysis (DA) were also used as an additional part of the exploration because they provided a visual complement that profiled and adjusted the information obtained. They also allowed us to confirm that the optimal final result of the grouping was the right one, which verified our initial hypothesis.
As a result, the cluster analysis in our study obtained an optimal number of clusters of several grouped profiles of patients with similar characteristics. All of them have closely interrelated variables with sufficient relevance to be considered in each clinical profile. More specifically, these three group profiles are based on the following details, which point to three levels of severity in the individual.
Profile 1 represents a group of patients who were smokers with relevant spirometric data and problems with diet, blood pressure, and heart rate values, which could also cause long-term hospital stays (i.e., variables FVCP, SPIROMETRY_PA, SMOKING_HABIT, WKG, DUR_ADM, SBP, DBP, FEVFVC, HR and FEV1P). All of this aggravated and worsened the conditions of the patients, who were weakened by other pathologies in addition to the main one.
Profile 2 represents a group of hypertense patients with spirometry performed and varying spirometric data. These patients possibly needed ventilatory support due to comorbidities, including a history of a cerebral vascular event, weight-related dietary problems, and long-term hospital stays conducive to a short-term negative outcome (i.e., variables SBP, DBP, HR, ESPIROMETRY_PA_, VS, RR, WKG, BMI, DUR_ADM, DEATH_90DAYS, CVD and CCVSDM). The recommendation was for them to adopt healthy habits that could improve their quality of life until the final stage of life and lessen the general wear-and-tear caused by other pathologies in addition to the central one.
Profile 3 represents a group of fairly severe patients with various comorbidities and pathologies (i.e., cerebrovascular event and congestive heart failure), with variable spirometric and anthropometric values due to their advanced age. These patients, who are probably hypertensive, as well as smokers, have ventilatory support needs because of cardiovascular comorbidities, (i.e., variables FEVFVC, FEV1P, AGE, CVD, CHF, VD, CCVSDM, FVCP, SPIROMETRY_PA, SMOKING_HABIT, DBP, HR and VS). Their condition rapidly worsened because of the main pathology and therapies. This means that other alternatives should be reviewed to maintain a good quality of life.
However, these results are not clinically conclusive since all the algorithms have certain limitations and may need to be complemented with other techniques if these models are to be reproduced in different settings with the same approach.
The results of this study provide new insights into the relevant factors (i.e., final clinical variables) for this type of patient. However, since they are separated into three levels of severity, efforts and knowledge should be focused on devising clinical procedures and treatments that will improve their diagnoses and their quality of life over time.

5. Conclusions

This research study showed that it is possible to use a set of classification techniques to analyze situations characterized by a sufficiently large volume of data. In this particular case, the cluster groups showed very similar clinical characteristics. This signified that the development of other pathologies (i.e., chronic diseases) with irregular increases in vital signs and frequent hospital admissions was caused by the severity of the disease itself and the advanced age of these patients, combined with poor health habits and their COPD exacerbations. This meant that the clinical condition of these patients was severe and would progressively deteriorate over time.
In this sense, our research study showed the usefulness of computational methods of supervised and unsupervised learning to facilitate the search for patterns in high-dimensional datasets coming from multicenter studies or Big Data repositories. In such contexts, it is almost always necessary to apply a classification procedure by clustering, which can add value to the clinical variables collected by professionals. However, before applying any algorithm, it is necessary to perform a filtering mechanism to select the most relevant variables or data in order to give validity to the decision.
Moreover, in many cases, this additional information is needed to make the correct diagnoses and, at the same time, personalize the procedures and treatments that do not require so many clinical resources. This leaves more time for people to improve their health habits, which provide a better quality of life.

Author Contributions

Methodology, N.B.-A., M.Á.M.-A. and A.M.-G.; software, M.Á.M.-A. and A.M.-G.; formal analysis, N.B.-A., M.Á.M.-A. and A.M.-G.; investigation, N.B.-A., M.Á.M.-A. and A.M.-G.; data curation, N.B.-A.; writing—original draft, N.B.-A. and M.Á.M.-A.; writing—review and editing, A.M.-G.; supervision, A.M.-G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The research data have not been shared. The data are not publicly available because these clinical data may compromise confidentiality and might reveal the identity or location of the participants. Additionally, the public availability of these data would be in violation of the Spanish Organic Law 15/1999 of the protection of personal data (consolidated text 5 March 2011) and the European Law (EU) 2016/679 from European Parliament and European Council of 27 of April 2016 about Data Protection (RGPD), without a corresponding anonymization procedure. Therefore, the data presented in this study are only available on request from the corresponding author. If the publication is accepted, after masking all the sensible data from this final clinical dataset, the data could be made available to the journal for publication or could be published in an official repository.

Acknowledgments

The authors thank all the people who participated directly or indirectly in the contribution of this analytical, exploratory work.

Conflicts of Interest

The authors declare no conflicts of interest.

Glossary

Definition of variables
AGEAge (years)
SEX (Male/Female)Sex (Male/Female)
SMOKING_HABITSmoking habit
DUR_ADMDuration of admission to the hospital (days)
HEIGHTHeight (meters)
WKGWeight in Kilograms
BMIBody Mass Index
SBPSystolic Blood Pressure (mmHg)
DBPDiastolic Blood Pressure (mmHg)
TEMPTemperature (°C)
RRRespiratory Rate (resp./min)
HRHeart Rate (beats/min)
FEV1Forced Expiratory Volume in the first second
FEV1PFEV1 spirometry in % of theoretical
FVCForced Vital Capacity
FVCPFVC spirometry in % of theoretical
FEVFVCFEV1/FVC ratio with spirometry performed on admission or discharge
SPIROMETRY_PASpirometry performed on admission or discharge
ADMAdmissions for any reason after 90 days
VSVentilatory support at any time of admission
EXACER_90DAYSExacerbation of COPD after 90 days
ReADM_EXACERReadmission for exacerbation of COPD
DEATH_90DAYSDeath after 90 days
EXITUSExitus throughout the admission period
CHFCongestive Heart Failure
CCVSDMCardiovascular Comorbidity
DMDiabetes Mellitus
VDVascular Disease
CVDCerebrovascular Disease
PVDPeripheral Vascular Disease
MIMyocardial Infarction
NEPHNephropathy
STSolid Tumor
MEMalleolar Edema
Acronyms
PCAPrincipal Component Analysis
PMMPredictive Mean Matching
COPDChronic Obstructive Pulmonary Disease
MICEMultiple Imputation by Chained Equations
AIArtificial Intelligence
MLMachine Learning
SVMSupport Vector Machine
CACorrespondence Analysis
DTDecision Tree Classification Analysis
KNNK-Nearest Neighbor
DBSCANDensity-Based Spatial Clustering of Applications with Noise
CHCalinski–Harabasz Index
PAMPartitioning around Medoids
FAFactorial Analysis
DADiscriminant Analysis

References

  1. López-Campos, J.L.; Almagro, P.; Gómez, J.T.; Chiner, E.; Palacios, L.; Hernández, C.; Navarro, M.D.; Molina, J.; Rigau, D.; Soler-Cataluña, J.J.; et al. Actualización de la Guía Española de la EPOC (GesEPOC): Comorbilidades, automanejo y cuidados paliativos. Arch. Bronconeumol. 2022, 58, 334–344. [Google Scholar] [CrossRef]
  2. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. Unsupervised learning. JMLR 2011, 12, 2825–2830. Available online: https://scikit-learn.org/stable/modules/clustering.html (accessed on 1 October 2023).
  3. Mirzal, A. Statistical Analysis of Microarray Data Clustering using NMF, Spectral Clustering, Kmeans and GMM. IEEE/ACM Trans. Comput. Biol. Bioinform. 2022, 19, 1173–1192. [Google Scholar] [CrossRef]
  4. Fratello, M.; Cattelani, L.; Federico, A.; Pavel, A.; Scala, G.; Serra, A.; Greco, D. Unsupervised Algorithms for Microarray Sample Stratification. Microarray Data Anal. Methods Mol. Biol. 2022, 2401, 121–146. [Google Scholar]
  5. Tobón, Á.; Rueda, J.; Cáceres, D.H.; Mejía, G.I.; Zapata, E.M.; Montes, F.; Ospina, A.; Fadul, S.; Paniagua, L.; Robledo, J. Adverse treatment outcomes in multidrug resistant tuberculosis go beyond the microbe-drug interaction: Results of a multiple correspondence analysis. Biomedica 2020, 40, 616–625. [Google Scholar] [CrossRef] [PubMed]
  6. Rokach, L.; Maimon, O. Minería de datos con árboles de decisión. Teoría y Aplicaciones. Ser. Percepción Máquinas Intel. Artificial. Chapters 1, 6 10 2007, 69, 264. [Google Scholar]
  7. Rajaguru, H.; Sannasi Chakravarthy, S.R. Analysis of Decision Tree and K-Nearest Neighbor Algorithm in the Classification of Breast Cancer. Asian Pac. J. Cancer Prev. 2019, 20, 3777–3781. [Google Scholar] [CrossRef]
  8. Orellana Alvear, J. Árboles de decisión y Random Forest. Árboles de Decisión—Parte I. 2018. Available online: https://bookdown.org/content/2031/arboles-de-decision-parte-i.html (accessed on 31 March 2023).
  9. Granville, V. How to Automatically Determine the Number of Clusters in Your Data—and More. 2019. Available online: https://www.datasciencecentral.com/profiles/blogs/how-to-automatically-determine-the-number-of-clusters-in-your-dat (accessed on 1 December 2023).
  10. Wickham, H.; Grolemund, G. Visualización de Datos Usando el Paquete “ggplot2”. 2023. Available online: https://es.r4ds.hadley.nz/03-visualize.html (accessed on 1 December 2023).
  11. DeSarbo, W.; Jedidi, K.; Cool, K.; Schendel, D. Simultaneous multidimensional unfolding and cluster analysis: An investigation of strategic groups. Mark. Lett. 1991, 2, 129–146. [Google Scholar] [CrossRef]
  12. Vichi, M.; Saporta, G. Clustering and disjoint principal component analysis. Comput. Stat. Data Anal. 2009, 53, 3194–3208. [Google Scholar]
  13. Freitas, A.; Macedo, E.; Vichi, M. An empirical comparison of two approaches for CDPCA in high-dimensional data. Stat. Methods Appl. 2021, 30, 1007–1031. [Google Scholar]
  14. Boukichou-Abdelkader, N.; Montero-Alonso, M.Á.; Muñoz-García, A. Different Routes or Methods of Application for Dimensionality Reduction in Multicenter Studies Databases. Mathematics 2022, 10, 696. [Google Scholar] [CrossRef]
  15. Amat Rodrigo, J. Clustering y Heatmaps: Aprendizaje no Supervisado. 2017. Available online: https://www.cienciadedatos.net/documentos/37_clustering_y_heatmaps (accessed on 25 March 2023).
  16. Sancho Caparrini, F. Algoritmos de Clustering. Algunos Representantes. Density-Based Spatial Clustering of Applications with Noise (DBSCAN). 2023. Available online: http://www.cs.us.es/~fsancho/Blog/posts/Clustering (accessed on 23 December 2023).
  17. Liu, W.; Yuan, K.; Ye, D. Reducing microarray data via nonnegative matrix factorization for visualization and clustering analysis. J. Biomed. Inform. 2008, 41, 602–606. [Google Scholar] [CrossRef] [PubMed]
  18. Niño-Ramírez, S.; Jaramillo-Arroyave, D.; Ardila, O.; Guevara-Casallas, L.G. Reducing the heterogeneity in hepatocellular carcinoma. A cluster analysis based on clinical variables in patients treated at a quaternary care hospital. Rev. Gastroenterol. Mex. (Engl. Ed.) 2021, 86, 356–362. [Google Scholar] [CrossRef] [PubMed]
  19. Zheng, C.H.; Huang, D.S.; Zhang, L.; Kong, X.Z. Tumor clustering using nonnegative matrix factorization with gene selection. IEEE Trans. Inf. Technol. Biomed. 2009, 13, 599–607. [Google Scholar] [CrossRef] [PubMed]
  20. Boutros, P.C.; Okey, A.B. Unsupervised pattern recognition: An introduction to the whys and wherefores of clustering microarray data. Brief. Bioinform. 2005, 6, 331–343. [Google Scholar] [CrossRef]
  21. Therneau, T.; Atkinson, B.; Ripley, B. Rpart: Recursive Partitioning and Regression Trees. R package Version 4.1.23. 2023. Available online: https://CRAN.R-project.org/package=rpart (accessed on 29 December 2023).
  22. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2023; Available online: https://www.R-project.org/ (accessed on 30 October 2023).
  23. López Cano, E. Visualización Avanzada para Análisis de Correspondencias con R. 2019. Available online: https://emilopezcano.github.io/seminario_urjc_2019 (accessed on 30 November 2023).
  24. Lowie, T.; Callens, J.; Maris, J.; Ribbens, S.; Pardon, B. Decision tree analysis for pathogen identification based on circumstantial factors in outbreaks of bovine respiratory disease in calves. Prev. Vet. Med. 2021, 196, 105469. [Google Scholar] [CrossRef]
  25. Vega-Pons, S.; Ruiz-Shulcloper, J. Una encuesta de algoritmos de conjunto de agrupación en clústeres. Rev. Int. Reconoc. Patrones Intel. Artif. 2011, 25, 337–372. [Google Scholar]
  26. Ortiz-Gonçalves, B.; Perea-Pérez, B.; Labajo González, E.; Albarrán Juan, E.; Santiago-Sáez, A. Tipologías de los madrileños ante la etapa final de la vida mediante un análisis de clusters [Typologies of Madrid’s citizens (Spain) at the end-of-life: Cluster analysis]. Spanish. Gac Sanit. 2018, 32, 346–351. [Google Scholar] [CrossRef]
  27. Maechler, M.; Rousseeuw, P.; Struyf, A.; Hubert, M.; Hornik, K.; Studer, M.; Roudier, P.; Gonzalez, J.; Kozlowski, K.; Schubert, E.; et al. Cluster: Finding Groups in Data. Cluster Analysis Extended Rousseeuw et al. R package Version 2.1.6. 2023. Available online: https://CRAN.R-project.org/package=cluster (accessed on 2 December 2023).
  28. Kassambara, A.; Mundt, F. Factoextra: Extract and Visualize the Results of Multivariate Data Analyses. R package Version 1.0.7. 2020. Available online: https://CRAN.R-project.org/package=factoextra (accessed on 1 March 2023).
  29. Nenadic, O.; Greenacre, M. Correspondence Analysis in R, with two-and three-dimensional graphics: The ca package. J. Stat. Softw. 2007, 20, 1–13. [Google Scholar]
  30. De La Fuente Fernández, S. Análisis de Correspondencias Simples y Múltiples. Fac. Ciencias Económicas y Empresariales. UAM. 2011. Available online: https://www.fuenterrebollo.com/Economicas/ECONOMETRIA/REDUCIR-DIMENSION/CORRESPONDENCIAS/correspondencias.pdf (accessed on 30 November 2023).
  31. Rangel, J.; Perea, J.; De-Pablos-Heredero, C.; Espinosa-García, J.A.; Mujica, P.T.; Feijoo, M.; Barba, C.; García, A. Structural and Technological Characterization of Tropical Smallholder Farms of Dual-Purpose Cattle in Mexico. Animals 2020, 10, 86. [Google Scholar] [CrossRef]
  32. Navarro-Mateu, F.; Garriga-Puerto, A.; Sánchez-Sánchez, J.A. Análisis de las alternativas terapéuticas del trastorno de pánico en atención primaria mediante un árbol de decisión [Tree decision analysis of the therapeutic alternatives for Panic Disorders in Primary Care]. Aten. Primaria 2010, 42, 86–94. [Google Scholar] [CrossRef]
  33. Bosco Mendoza Vega, J. Árboles de Decisión con R. Clasificación. 2018. Available online: https://rpubs.com/jboscomendoza/arboles_decision_clasificacion (accessed on 31 March 2023).
  34. Wang, L.; Zhu, L.; Jiang, J.; Wang, L.; Ni, W. Decision tree analysis for evaluating disease activity in patients with rheumatoid arthritis. J. Int. Med. Res. 2021, 49, 3000605211053232. [Google Scholar] [CrossRef]
  35. Martínez De Lejarza, I.; Esparducer. Árboles de Clasificación y Regression. 2014. Available online: https://www.uv.es/mlejarza/actuariales/tam/arbolesdecision.pdf (accessed on 31 March 2023).
  36. Im, E.O.; Yi, J.S.; Chee, W. A decision tree analysis on multiple factors related to menopausal symptoms. Menopause 2021, 28, 772–786. [Google Scholar] [CrossRef]
  37. Franchuk, V.V.; Mikhaylichenko, B.V.; Franchuk, M.V. Primenenie metoda dereva resheniĭ v sudebno-meditsinskoĭ ékspertnoĭ praktike pri analize ‘vrachebnykh del’ [Application of the decision tree method in forensic-medical practice in the analysis of ‘doctors proceedings’]. Sud. Meditsinskaia Ekspertiza 2020, 63, 9–14. [Google Scholar] [CrossRef] [PubMed]
  38. Karacan, I.; Sennaroglu, B.; Vayvay, O. Analysis of life expectancy across countries using a decision tree. East. Mediterr. Health J. 2020, 26, 143–151. [Google Scholar] [CrossRef] [PubMed]
  39. Gheondea-Eladi, A. Patient decision aids: A content analysis based on a decision tree structure. BMC Med. Inf. Decis. Mak. 2019, 19, 137. [Google Scholar] [CrossRef] [PubMed]
  40. Martínez Heras, J. Decision Trees and Random Forests. Supervised Learning with Python. Classification models with Machine Learning, 2018. Update. 2023. Available online: https://github.com/jmartinezheras/2018-MachineLearning-Lectures-ESA/tree/master/3_DecisionTrees-RandomForests (accessed on 31 October 2023).
Figure 1. Dendrogram or hierarchical tree for 3 cluster groups.
Figure 1. Dendrogram or hierarchical tree for 3 cluster groups.
Computation 12 00055 g001
Figure 2. Silhouette method. Three groups represents the optimal solution according to maximum mean value of the indices.
Figure 2. Silhouette method. Three groups represents the optimal solution according to maximum mean value of the indices.
Computation 12 00055 g002
Figure 3. Description of plan 1:2 (patients) by groups separated by CA.
Figure 3. Description of plan 1:2 (patients) by groups separated by CA.
Computation 12 00055 g003
Table 1. Correlational description of final optimal vs. variable clusters.
Table 1. Correlational description of final optimal vs. variable clusters.
VariablesCluster 1Cluster 2Cluster 3
CCVSDM0.734−2.122.22
DUR_ADM2.34−2.45−0.203
AGE−1.48−1.415.15
SPIROMETRY_PA5.042.51−13.6
VD0.227−1.742.52
HR−12.815.5−2.54
FEV1P−11.9−0.99223.9
FEVFVC−24.3−1.0346.9
RR−1.282.18−1.31
FVCP9.46−0.63−16.5
SMOKING_HABIT2.96−0.112−5.32
CVD−0.334−2.354.6
CHF−0.114−1.382.54
BMI1.84−31.66
DEATH_90DAYS1.69−2.370.873
WKG2.51−3.771.72
VS−0.9362.2−1.98
DBP−26.730.8−2.62
SBP−40.344.4−0.136
Table 2. Decomposition of total inertia using CA. Eigenvalues for Dim.1–Dim.31.
Table 2. Decomposition of total inertia using CA. Eigenvalues for Dim.1–Dim.31.
Dim.1Dim.2Dim.3Dim.4Dim.5Dim.6Dim.7Dim.8Dim.9
Variance0.0180.0100.0080.0060.0050.0040.0030.0030.002
% of var.24.13214.05311.0388.3587.0374.8323.7133.4853.145
Cumulative % of var.24.13238.18549.22357.58164.61869.44973.16276.64779.792
Dim.10Dim.11Dim.12Dim.13Dim.14Dim.15Dim.16Dim.17Dim.18
Variance0.0020.0010.0010.0010.0010.0010.0010.0010.001
% of var.2.4151.8961.7231.6691.6081.5201.4911.3701.307
Cumulative % of var.82.20684.10385.82687.49589.10490.62492.11593.48594.792
Dim.19Dim.20Dim.21Dim.22Dim.23Dim.24Dim.25Dim.26Dim.27
Variance0.0010.0010.0010.0000.0000.0000.0000.0000.000
% of var.1.2041.0400.8130.6210.3310.2900.2730.2390.211
Cumulative % of var.95.99697.03697.84998.47098.80199.09199.36499.60399.814
Dim.28Dim.29Dim.30Dim.31
Variance0.0000.0000.0000.000
% of var.0.0840.0730.0270.002
Cumulative % of var.99.89899.97199.998100.000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Boukichou-Abdelkader, N.; Montero-Alonso, M.Á.; Muñoz-García, A. Group Classification for the Search and Identification of Related Patterns Using a Variety of Multivariate Techniques. Computation 2024, 12, 55. https://doi.org/10.3390/computation12030055

AMA Style

Boukichou-Abdelkader N, Montero-Alonso MÁ, Muñoz-García A. Group Classification for the Search and Identification of Related Patterns Using a Variety of Multivariate Techniques. Computation. 2024; 12(3):55. https://doi.org/10.3390/computation12030055

Chicago/Turabian Style

Boukichou-Abdelkader, Nisa, Miguel Ángel Montero-Alonso, and Alberto Muñoz-García. 2024. "Group Classification for the Search and Identification of Related Patterns Using a Variety of Multivariate Techniques" Computation 12, no. 3: 55. https://doi.org/10.3390/computation12030055

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop