Next Article in Journal
Multimodal Emotional Classification Based on Meaningful Learning
Previous Article in Journal
Hierarchical Co-Attention Selection Network for Interpretable Fake News Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Performance of International Students and Students with Disabilities: Early Prediction and Feature Selection through Educational Data Mining

by
Thao-Trang Huynh-Cam
1,2,
Long-Sheng Chen
1,* and
Khai-Vinh Huynh
3
1
Department of Information Management, Chaoyang University of Technology, Taichung 413310, Taiwan
2
Foreign Languages and Informatics Center, Dong Thap University, Cao Lanh City 81118, Vietnam
3
Quality Assurance Office, Dong Thap University, Cao Lanh City 81118, Vietnam
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2022, 6(3), 94; https://doi.org/10.3390/bdcc6030094
Submission received: 15 July 2022 / Revised: 29 August 2022 / Accepted: 5 September 2022 / Published: 7 September 2022

Abstract

:
The learning performance of international students and students with disabilities has increasingly attracted many theoretical and practical researchers. However, previous studies used questionnaires, surveys, and/or interviews to investigate factors affecting students’ learning performance. These methods cannot help universities to provide on-time support to excellent and poor students. Thus, this study utilized Multilayer Perceptron (MLP), Support Vector Machine (SVM), Random Forest (RF), and Decision Tree (DT) algorithms to build prediction models for the academic performance of international students, students with disabilities, and local students based on students’ admission profiles and their first-semester Grade Point Average results. The real samples included 4036 freshmen of a Taiwanese technical and vocational university. The experimental results showed that for international students, three models: SVM (100%), MLP (100%), and DT (100%) were significantly superior to RF (96.6%); for students with disabilities, SVM (100%) outperformed RF (98.0%), MLP (96.0%), and DT (94.0%); for local students, RF (98.6%) outperformed DT (95.2%) MLP (94.9%), and SVM (91.9%). The most important features were [numbers of required credits], [main source of living expenses], [department], [father occupations], [mother occupations], [numbers of elective credits], [parent average income per month], and [father education]. The outcomes of this study may assist academic communities in proposing preventive measures at the early stages to attract more international students and enhance school competitive advantages.

1. Introduction

The learning performance of minority students, such as international students and students with disabilities, has increasingly attracted considerable theoretical and practical attention from researchers and educational teams since it is currently considered to be one of the crucial criteria for assessing campuses’ quality [1]. Early prediction of minority students’ performance and the determining of important features affecting their learning performance are indispensable to higher education institutions (HEIs). Students’ big data analysis can assist academic communities to foresee students’ learning conditions, provide on-time support, and propose preventive measures at the early stages before students start the first semester. Thus, HEIs will attract more international students and enhance their competitive advantages in global educational environments.
Studying abroad benefits both international students and host campuses. For students, the extended stays abroad are considered valuable opportunities for acquiring and improving various skills, such as foreign language competence and intercultural skills [2,3]. Vice versa, international students contribute significantly to the financial resources, cross-culture environments, and high-quality human resources (i.e., research and teaching assistants) of host universities. HEIs’ staff can learn various cultures from different countries; schools can have more opportunities for further cooperation with global campuses from alumni recommendations [2,4].
In recent years, compared to Asian countries, Taiwan has offered many advantageous supports in high-quality international academic environments. Thus, thousands of international students have studied in Taiwan during the past decade [2,5,6]. Annually, the numbers of international students in Taiwan, including degree and non-degree level (i.e., exchange and language-study), have increased considerably [2,4,5,7]. Particularly, the populations of international degree-level students increased from 6380 in 2001 to 21,005 in 2007 [5,8]. The numbers of overseas students in Taiwan rose considerably, from approximately 63,000 in 2012 to 130,000 in the 2019 academic year [6]. The incoming international students remain unchanged owing to the COVID-19 pandemic.
Nevertheless, according to Education in Taiwan 2016–2017 [9], there is a huge increasing number of students with special needs in various levels between 2001–2015 in Taiwan. In particular, the rising number of students with disabilities is from 3689 to 15,559 in preschool, from 35,721 to 42,022 in primary, from 20,993 to 28,228 in junior high school, from 6952 to 23,577 in senior and vocational high school, and from 2961 to 12,376 in higher education. The government budget for supporting students with disabilities has increased from NTD$5.579 billion in 2001 to NTD$9.903 billion in 2015.
Owing to the digitalization of academic processes, large data repositories provide a vast educational data (ED) for learning and analyzing how students learn [10]. Data mining (DM) is also called knowledge discovery in databases (KDD). DM is concerned with data analysis by using different techniques/software to extract meaningful information and knowledge from the raw data sources and to identify the relationship among the dataset [11,12,13,14]. The DM process comprises data cleaning, data integration, data selection, data transformation, data mining, and knowledge representation [15]. DM techniques including machine learning algorithms and artificial intelligence (AI) have been widely and successfully applied to predict and to identify the important features associated with student academic performance [16].
Educational data mining (EDM), an application of DM in education sectors, has recently become popular and inspired many researchers. EDM can be used to understand students’ learning conditions, students’ behaviors, and their subject interests in order to improve teaching supports, and to make decisions in educational systems [17,18], since most ED are collected from students’ interactions and behaviors [12,19]. EDM’s models are capable of forecasting student knowledge and future performance [11]. However, from available research, no published papers focused on predicting and determining factors which impact greatly on the learning performance of international and special-need freshmen in Taiwanese vocational and technical universities.
Therefore, this study utilized EDM techniques to build prediction models for the academic performance of international students, students with disabilities, and local students based on students’ admission profiles and their first-semester grade point average results collected from the school database system of a Taiwanese vocational and technical university. The study may benefit stakeholders in the educational sectors. Teachers can adjust their pedagogical strategies and teaching methods when teaching international and/or special-needs students. Policymakers can plan, design, and implement institutional policies to improve the international and special-needs students’ performance, propose preventive measures at the early stages before students start the first semester in order to attract more international students, and enhance campuses’ competitive advantages in the global education markets. The government can establish educational policies from the analysis of cross-institutional data [12,19]. Parents can solve the students’ problems, especially financial problems and/or guide them properly.

2. Background and Related Works

2.1. Educational Data Mining (EDM)

EDM is data mining in the educational sector. EDM is the process by which the collected raw data from the educational environment based on hypothesis formation and/or educational questions are pre-processed and imported into DM instruments/software to produce models/patterns for interpretation and evaluation [10,12]. The results of interpretation and evaluation are then used to refine the educational environment and hypothesis formation. Correspondingly, Siemens and Baker [20] suggest that EDM is a practice of DM methods for studying big datasets in order to get insights from students and educational systems. EDM is “not only to turn data into knowledge, but also to filter mined knowledge for decision-making about how to modify the educational environment to improve student’s learning” [21]. It has been applied to evaluate the pedagogical support of a specific learning tool and recommend potential improvements [21]. The goals of EDM include (1) predicting student’s learning behavior; (2) exploring or upgrading domain models; (3) analyzing the effects of various instructional support types; and (4) advancing scientific knowledge [22,23]. To add value, EDM must assist in exploiting the multiple levels of meaningful hierarchy in ED, especially predicting student learning performance and their future achievement for long-term development.

2.2. Student Learning Performance Prediction

Prediction is to infer a target attribute (predicted variable) from a combination of other aspects of data (predictor variable) [10,11]. Prediction requires labels for output variables. In EDM, prediction has been popularly used for forecasting student performance and for detecting students’ behaviors [10,11]. The category of output variable for prediction can be either categorical or continuous.
Student learning performance prediction (SLPP) has been deeply concerned by HEIs, since early SLPP can lead to careful strategic intervention plans before students reach the final semester [24] and even prevent students at-risk/dropouts by providing them with additional assistance or tutoring on time [25]. Therefore, SLPP has increasingly attracted a diversity of researchers and educational teams in educational sectors. In order to gain a high-value contribution, various prediction techniques and feature selection methods are also utilized.
The main stream of SLPP methods are supervised learning, since “students’ academic learning performance” is the class label. Prediction models are implemented by algorithms on specific computational software/tools. Popular algorithms employed for SLPP include artificial neural network (ANN), decision tree (DT), support vector machine (SVM), random forest (RF), k-nearest neighbor (KNN), and Naïve Bayes [26,27]. DT is popularly and successfully applied for prediction and classification in the domain of machine learning due to its simple use, easy understanding, and high prediction accuracy [26,27,28,29] through a flowchart-like tree structure and IF-THEN rules [30,31,32]. In other words, DT is a model-based approach using a tree-shaped graph. It is a root-to-leaf route which represents classification rules [25,33]. In its flowchart-like tree structure, each inner node represents a test on the feature/variable, each branch represents the test result, and each leaf represents a class label [25]. Compared to other methods, DT can extract readable knowledge rules, which is helpful for university-side decision-making references [34]. However, one of the biggest drawbacks of the DT model is overfitting, which can be solved by the presence of RF [25,35,36]. SVM is good for handling small datasets, while ANN can solve the nonlinear and complex relationship between various input and output variables [37,38].
The SLPP features are mainly grouped into popular categories: demographic, academic performance, internal assessment, communication, behavioral, psychological, and family/personal background [38]. For instance, Huynh-Cam et al. [24] measured the impact of family background features on first-year undergraduates using DT and RF. Matzavela and Alepis [31] assessed the effects of gender, grade, parent education, parent income, whether one is the first child or not, and whether a student is working or not on university student performance using DT; whereas authors in [30] evaluated university students final GPA using J48 DT. Niyogisubizo et al. [39] examined the impact of the features: access, tests, tests grade, exam, project, project grade, assignments, result points, result grade, graduate, year, and academic year on student dropouts in higher education using ANN and RF.

2.3. Learning Performance Prediction of Minority Students

International students contribute significantly to host HEIs’ diversity, revenue, investment, research, and teaching [2,4,40]. Thus, it is important for the host HEIs’ heads to be aware of factors associated with international students’ learning performance in order to offer in-time support, prevent at-risk dropout, maintain international students’ retention, and attract a larger number of international students. Therefore, this issue has been one of significant concern of policy makers, practitioners in HEIs and researchers in the past decades. In Taiwanese universities, there were three main groups of studies focusing on top-ranking factors associated with international student performance. Group one included simplified Chinese characters, Taiwanese government scholarships, and high-quality Mandarin Studies programs [5]; group two consisted of stressors, living support, and adjustment [2]; and group three comprised social adjustment experiences to campus life [7]. In contrast, for graduate students in the United States, the important factors included gender, age, native region, native language, undergraduate GPA, proportion of time studying alone, teaching and learning methods, and length of study time [40]. In German high schools, socioeconomic background, grades, and course choice in English positively affected studying abroad during high school [3].
Inclusive education for students with special educational needs in regular classrooms has currently become an increasing trend [41]. It is also one of the required criteria for assessing campuses’ quality in many countries since it is claimed that students with disabilities should have equal rights to receive education in mainstream schools [41,42]. Therefore, the academic performance of students with special educational needs has attracted many researchers. For instance, Griffiths et al. [43] described the co-development of a six-phase tripartite model for a supportive framework for nursing students with special educational needs through an individual student pathway. Huang et al. [41] analyzed the academic peer influence of students with disabilities in the classroom of Chinese middle schools using students’ midterm scores in three compulsory subjects: Chinese, math, and English. Hersh [44] introduced an evaluation framework for ICT-based learning technologies for students with special needs by a systematic approach. Zainudin et al. [45] used a questionnaire to investigate the effects of family support, facility support, and lecture support on special-needs students’ academic performance in e-learning during the COVID-19 pandemic.

2.4. Research Gaps

From the available published studies, none of them employed EDM approaches/techniques to predict the academic performance of international students and students with disabilities. In addition, among these limited studies, many researchers used questionnaires and/or interviews to investigate factors associated with international and special-needs students’ learning performance. These methods cannot help universities provide on-time support to excellent and poor students.
There are a wide variety of SLPP techniques (Section 2.2). However, this study applied four well-known classification techniques: multilayer perceptron (MLP), decision tree (DT), support vector machine (SVM), and random forest (RF) algorithms to build prediction models owing to their critical features. We employed DT owing to its wide application, ease of use, and simplicity. We used MLP and SVM due to the nature of our small datasets, which did not require more complex algorithms. RF can help solve the overfitting/overtraining problems.

3. Experimental Methodology

Figure 1 summarizes the five-step experimental process in this study. Step 1 reports data collection methods and describes the research datasets. Step 2 describes a two-step data pre-processing process: data cleaning and data normalization. Step 3 explains a three-step model implementation: data split, feature selection, and model prediction building. Step 4 evaluates model performance and Step 5 concludes extracted knowledge from EDM and the important features affecting minority students’ academic performance. The following subsections will explain these five steps in detail.

3.1. Step 1. Dataset Collection and Description

The raw dataset used to conduct the experimental study was collected directly from the school database system of a Taiwanese technical and vocational university during the first semester of the 2020–2021 academic year. The original dataset contained 7736 enrolled records of first-year students after official admission. For each student, there were 22 categories of profile information: [department], [gender], [address], [admission status], [aboriginal], [child of new residents], [family children ranking], [parent average income per month], [on-campus accommodation], [main source of living expenses], [student loan], [tuition waiver], [father live or not], [father’s occupations], [father’s education], [mother live or not], [mother’s occupations], [mother’s education], [numbers of required credits], [numbers of elective credits], [sick leave], and [personal leave] were used as input variables. The first-semester grade point average (GPA) results in the students’ bachelor program was used as an output variable. These students’ identity remained anonymous for ethical reasons.

3.2. Step 2. Data Pre-Processing

Data pre-processing is a DM technique which is used to convert the original data, raw or primary data into a useful and efficient dataset by addressing DM algorithms [21,46,47]. This phase consists of two steps: data cleaning and data normalization.
Step 2.1: Data cleaning: In this step, irrelevant attributes and missing-value samples were removed. In addition, all category features were encoded and transferred to binary and/or numeric features.
At first, we eliminated some irrelevant attributes, e.g., student name, student ID, birth date, and birthplace. Then from 22 initial features (Section 3.1), we removed four features: [aboriginal], [child of new residents] [mother alive or not], and [father live or not] since our previous study [24] illustrated that they hardly affected students’ learning performance. The feature [address] is inconsistent among the three target groups of students because its corresponding geographic regions are only used for Taiwanese students, e.g., North, South, East, West, and Island. In addition, we deleted the feature [admission status] because we only focused on degree students, not exchange students, transfer students and non-degree students. After deleting irrelevant attributes, out of 22 initial features, a total number of 16 features were finally selected for model building, as listed in Table 1.
Features 1–15 were input (independent) variables and Feature 16 was the output (dependent) variable for prediction models. Nevertheless, Feature 10 [student loan] and Feature 11 [tuition waiver] were not applied for Group 1 (international students), since the school did not offer loan and tuition waiver to these students at the research time.
After data cleaning, out of 7736 original examples, a total of 4036 (52.2%) students were finally selected for experimental study. These students were divided into three target groups: international students (Group 1), students with disabilities (Group 2), and local students (Group 3). Figure 2 displays the numbers of employed students in each group. The remaining 3700 (47.8%) students, who had missing values, dropped out, and/or were suspended before implementing the experiment, were excluded from this study.
Finally, we encoded and transferred all category features to binary and/or numeric features as described in Table 1. We also constructed categorical target variables based on the original numeric parameter university GPA scores which are in compliance with the Taiwanese grading system (100-points scale). The target variable (Feature 16) has five categories: excellent (90–100 points), very good (80–89 points), good (70–79 points), average (60–69 points), and poor (0–59 points).
Step 2.2: Data normalization. In this step, the data was normalized in accordance with Equation (1), which was applied in our previous study [24].
X m o n = X X m i n X m a x X m i n
where Xmax is the maximum value, Xmin is the minimum value, and Xmon is the normalized value.

3.3. Model Implementation

There are three steps in this phase. In Step 3.1, the input data were randomly divided into training and testing datasets ten times with percentages of 80% and 20%, respectively. Step 3.2 was for feature selection in order to overcome overfitting problems and better prediction performance [37,39,48].
Step 3.3 was to build prediction models. In this step, we employed four supervised classification algorithms: SVM, MLP, RF, and DT in the Python language which were widely used in machine learning. The software tool chosen for building prediction models in this study was Jupyter, an open-source software project affiliated with the 501c3 NumFOCUS Foundation. The software was developed openly by the Jupyter/IPython Project and was hosted in public GitHub repositories under the IPython GitHub organization (https://github.com/ipython, accessed on 1 February 2022) and the Jupyter GitHub organization (https://github.com/jupyter, accessed on 1 February 2022). Jupyter Project is developed by a team of contributors. Contributors are individuals who have contributed code, documentation, designs or other work to one or more the Project repositories. The Python and Jupyter packages are available at scikit-learn: machine learning in Python—scikit-learn 1.0.2 documentation and https://www.dataquest.io/blog/jupyter-notebook-tutorial (accessed on 1 February 2022). Each model was experimented on ten times with ten different training-testing datasets. The mean value and standard deviation of the 10-time experiment for each model were taken and used for benchmarking prediction performance among SVM, MLP, RF, and DT models. After selecting the prediction model with the best performance, the Gini index was applied to select important features associated with students’ learning performance, especially the minority students: international students and students with disabilities. Furthermore, in order to achieve the highest performance accuracy, we classified the output variable into three cases (Figure 3).
  • Case 1: Origin included 5 origin classes: Excellent (EX), Very Good (VG), Good (G), Average (AVG), and Poor. This case was used for investigating if the models predict the minority or not. Figure 4 graphically displays the numbers of employed samples in each group.
  • Case 2: Minority focus and combination (MFC) focused on two minority classes: EX and Poor. The majority classes VG and G were removed. Since the classes “AVG” and “Poor” were very few, we combined the two minority AVG and Poor classes into the Poor class. Therefore, the new combined Poor class (AVG+Poor) included six samples in Group 1, 12 samples in Group 2, and 750 samples in Group 3 (Figure 5a). However, after combining, the imbalanced data problem was present in each group. As shown in Figure 5a, the Poor (AVG+Poor) class remained the minority class in Group 1 (nPoor = 6; nEX = 18); whereas it became the majority class in Group 2 (nPoor = 12; EX = 1) and Group 3 (nPoor = 750; nEX = 127). Therefore, we proposed a resampling method: random oversampling the minority (Case 3) to solve the imbalanced data problem in each group (Figure 5b).
  • Case 3: Random oversampling (ROM) was to randomly oversample the minority class in each group by duplicating or generating new minority class instances [49,50]: “EX” and “Poor” classes. As shown in Figure 5b, the numbers of samples in each group are approximately balanced. In the works of Chen et al. [49] and Chang et al. [50], they indicated that oversampling is one of effective solutions for tackling class imbalance problems. Therefore, we employed ROM to deal with class imbalance problems in this study.

3.4. Model Evaluation

In order to evaluate the effectiveness of our proposed method (ROM), we employed three methods.
Method 1: we benchmarked the prediction accuracy of Case 3 (ROM), Case 2 (MFC), and Case 1 (Origin). However, relying only on accuracy could lead to misinterpretation when the classification involves imbalanced data [49,51], since the models can predict the majority class and ignore the minority class [39]. Therefore, we employed various evaluation metrics.
Method 2: we compared the accuracy (ACC), precision (PR), recall (Rec), the area under the curve (AUC), and F1-score (F1), which were employed by the authors in [39,52,53,54]. We also used confusion matrix results to explain the classification models. The parameters of the confusion matrix were:
  • True Positive (TP): instances, which are actually positive, are classified as positive.
  • False Positive (FP): instances, which are actually negative, are classified as positive.
  • False Negative (FN): instances, which are actually positive, are classified as negative.
  • True Negative (TN): instances, which are actually negative, are classified as negative.
In this study, Positive instances were EX class and Negative instances were Poor class. The formula of all metrics used in the evaluation stage are described below:
ACC = True   Positive   ( TP ) + True   negative   ( TN ) True   Positive   ( TP ) + True   negative   ( TN ) + False   Negative   ( FN ) + False   Positive   ( FP )
ACC is used to find the portion of correctly classified values.
PR = TP TP   +   FP
PR is used to calculate the model’s ability to classify positive values correctly.
Rec = TP TP   +   FN
Rec is used to calculate the models’ ability to predict positive values.
F 1 = 2     Rec     PR Rec   +   PR
F1-score is the harmonic mean of Recall and Precision, which is used when we need to take both PR and Rec into account.
Method 3: we utilized the area under the curve (AUC), which is frequently used to measure the prediction performance of a classification method for all classification thresholds [39]. The AUC values range between 0.5 and 1.0, where the 1.0 value indicated the excellent performance and the 0.5 value was for the poor performance of a specific model.

4. Experimental Results

After pre-processing, the dataset was imported to Jupyter software to implement MLP, RF and DT, and SVM models. About parameter settings in RF, the number of trees in the forest is set to 100. With regard to decision trees, pruning confidence value (CF) affects the way of estimating the error rate, thereby affecting the severity of pruning in order to avoid overfitting of the model. In this study, pruning CF was set to 25%. In MLP, the learning rate is set to 0.3, and the training stop condition is set to the number of learning iterations to 1000. At this time, the RMSE (Root-Mean-Square Error) has been flattened, meaning that the network has converged. All optimal parameter settings of SVM could be obtained automatically by using the grid search technique.

4.1. Results of Case 1: Origin

In this case, we used accuracy, precision, recall, and F1-score values to evaluate the prediction performance. Table 2 summarizes all results for three cases. From this table, we can find that all accuracies of the four classifiers, SVM, MLP, RF, and DT are very low. The same situation could also be found in PR, Rec, and F1. In other words, we cannot identify the minority classes “Excellent” and “Poor”. If we cannot identify these minority classes, the classification results will be meaningless. Therefore, we implement Case 2, which is minority focus and combination (MFC). In MFC, we only focused on two minority classes, EX and Poor.

4.2. Results of Case 2: Minority Focus and Combination (MFC)

Table 3 summarizes the classification results of Case 2. In this case, we only focus on two classes, EX and Poor. From this table, we find that the accuracy in three groups was improved. And, PR, Rec, and F1 are all acceptable. In Group 1 and 2, SVM can fully predict 100% EX and Poor examples. However, in Group 3 (local students), RF can have the best performance. Table 4 shows the confusion matrix results. It is obvious that RF can identify the Poor class very well (0.931), but it poorly classifies the EX class (0.412) for local students. Therefore, we implemented random oversampling in Case 3.

4.3. Results of Case 3: Random Oversampling the Minority (ROM)

Table 5 lists the classification results of Case 3. Compared to Case 1 and Case 2, it is clear that the class imbalance problem has been tackled. In Group 1 (international students), SVM, MLP, and DT could achieve 100% performance in both accuracy and F1. For Group 2 (students with disabilities), SVM can fully predict both Poor and EX examples. In Group 3 (local students), RF outperforms the other 3 methods in ACC, PR, Rec, and F1. Consequently, we will use the results of Case 3 to select the important features associated with students’ learning performance.
Moreover, as shown in Table 6, the AUC results for RF, SVM, MLP, and DT models for Group 1 are 1.00, 1.00, 1.00, and 1.00; for Group 2 they are 1.00, 1.00, 0.83, and 1.00; and for Group 3 they are 1.00, 1.00, 0.96, and 0.95. An outperformed AUC score obtained from our proposed models showcases that the classification performance is better and acceptable. In Table 7, the outcomes of confusion matrices show that for Group 1 and Group 2, the True Positive Rate (TPR) value are all 1.00. For Group 3, the TPR value is 1.00 and the True Negative Rate (TNR) value is 0.934. It can be indicated all four models (SVM, RF, MLP, and DT) have excellent performance in Case 3.

5. Discussion

Most of the attention of resampling methods for imbalanced classification is put on oversampling the minority class. However, some techniques have been developed for under-sampling the majority class. In order to provide an in-depth discussion about the used re-sampling (ROM) method in the analysis, we implement the random under-sampling technique. Table 8 lists the comparison of under-sampling, ROM and MFC. In this table, MFC represents the original data without implementing re-sampling techniques. From this table, we can reach three conclusions. First, oversampling (ROM) outperforms under-sampling, no matter the accuracy and F1. Second, a skewed class distribution might not necessarily cause class imbalance problems. As can be seen from Table 4, Group 1 and 2 do not have a class imbalance problem. Group 3 has a class imbalance problem, since the classifier gains a poor accuracy rate for minority class, but a higher accuracy rate for majority class. Third, if we implement re-sampling methods for those data without having class imbalance problems (Group 1 and 2), the classification performance cannot be improved. Therefore, it is recommended that future researchers not only use the re-sampling technique just by looking at the distribution class, but also look at the classification performance, whether there is a very high detection rate for majority class examples, or a very low classification rate for minority class examples.
In practice, the prediction models built in Case 3 are more meaningful than the models of Case 1 and Case 2. Therefore, we focused on the results of Case 3: ROM. We used the Gini index of the SVM algorithm to select important features associated with Group 1 learning performance and that of the RF algorithm used for the learning performance of Group 2 and Group 3. Figure 6 graphically displays the ranking of 15 input features and Table 9 lists the seven top-ranking important features associated with learning performance of three groups of students.
For international students, the seven top-ranking features include [numbers of required credits], [main source of living expenses], [department], [father occupations], [numbers of elective credits], [parent average income per month], and [father’s education]. It is clear that financial issues seem very important to international freshmen students since they live far away from their homeland. In addition, due to the Chinese language barrier and different cultures, it is very difficult for international students to find part-time jobs when they first arrive in Taiwan. Thus, campuses should offer them financial supports, e.g., scholarships and tuition waiver since Scholarships and tuition waiver are considered one of the contributing features for expanding international enrollment [5,55]. Apart from finances, numbers of required and elective credits contribute to both excellent and poor academic performance. As newcomers in the first semester, students find that learning and teaching methods, academic requirements, materials, and language quite different from those in their home countries. They do not know how many credits are suitable for their study. The unique relation with schools is through the international affairs department and tutors in academic departments. However, in practice, some departments receive international students for the first time, thus, they do not know how to assist international students’ learning. In addition, intensive courses in the Mandarin language are offered to international students. Nevertheless, in the first semester, students’ Chinese level is not adequate for long lectures and challenging assignments. Some departments offer teaching assistants and/or language supporters; whereas the other departments do not. Therefore, students studying in the departments with language supporters may have excellent performance, and vice versa, students studying in the departments without language assistance may have poor academic performance. Thus, schools should offer orientation/guided sessions and language supporters at the first semester. Academic staff should select understandable and comprehensive materials to ease students’ learning process. Teaching methods should be adaptable to the students’ needs. In contrast, the four factors [on-campus accommodation], [gender], [sick leave], and [personal leave] had the least effect on their learning performance.
For students with disabilities, the seven top features are [father occupations], [mother education], [father education], [numbers of required credits], [mother occupation], [department], and [numbers of elective credits]. As students need special assistance, families play an essential role in both their daily and their academic lives. Universities should involve families in student learning. For instance, they could allow family members to enter the classes just in case students need helps since the family realize the students’ specific health status. By contrast, the four features: [tuition waiver], [personal leave], [main source of living expenses], and [student loan] had the lowest impact.
For local students, the seven top-ranking features consist of [numbers of required credits], [sick leave], [department], [personal leave], [mother occupations], [numbers of elective credits], and [father occupations]. Sick and personal leaves seem more important than the other features. Leaving school may not be the main reason for laziness, but students may have difficulties in their learning process as newcomers. In the first semester, students experience new academic lives, thus they need time for adjustment. In addition, family background is important to the local students’ learning performance. Nevertheless, the four features: [mother education], [student loan], [on-campus accommodation], and [tuition waiver] hardly affected their academic performance.

6. Conclusions

The present study successfully built prediction models for the academic performance of international students and students with special needs at a Taiwanese vocational and technical university using the SVM, MLP, RF, and DT algorithms. The findings tackled the limitations of our previous study in [24]. The experimental results showed that for international students, three models: SVM (100%), MLP (100%), and DT (100%) were significantly superior to RF (96.6%); for students with disabilities, SVM (100%) outperformed RF (98.0%), MLP (96.0%), and DT (94.0%); for local students, RF (98.6%) outperformed DT (95.2%) MLP (94.9%), and SVM (91.9%). The most important features were [numbers of required credits], [main source of living expenses], [department], [father occupations], [mother occupations], [numbers of elective credits], [parent average income per month], and [father education]. The outcome of this study is expected to benefit stakeholders in similar educational contexts when offering studying programs to international students and students with special needs.
Teachers can adapt more flexible teaching methods and offer understandable and comprehensive materials to facilitate students’ learning. Orientation/guided sessions and tutoring should be offered to minority students. Availability and accessibility of scholarships should be offered to international students in order to expand international enrollment and maintaining student retention. The government can establish educational policies from the analysis of cross-institutional data. Parents can solve the students’ problems, especially financial problems and guide them properly.
The major contributions of this study could be summarized as follows.
(1). To provide a predictive model for early warning for the academic performance of students with special needs and international students in universities.
Previous studies used variable data obtained during the semesters, such as quizzes, homework, absenteeism, etc. to predict academic performance. Our study only uses family factors, departments, and numbers of credits to predict minority students’ performance before students start the first semester. This early prediction can support HEIs in providing additional assistances and proposing preventive measures prior to the start of the semesters
(2). To determine key factors associated with the academic performance of students with special needs and international students in Taiwanese technical and vocational universities.
In Taiwan, students with special needs have a high dropout rate and generally poor academic performance. However, in accordance with the special education laws, these students have educational funds provided by the Taiwanese government to assist their studies. Our study allowed Taiwanese HEIs to identify the students who really need help in order to enhance the counseling for students with special needs.
(3). To realize the differences between students with excellent academic performance (international students) and students with poor academic performance (students with special needs).
In Taiwanese technical and vocational universities, international students need outstanding academic performance to win scholarships based on a stricter screening mechanism; therefore, their academic performance is excellent. Students with special needs, on the other hand, are generally less effective in learning because of congenital physical and psychological limitations. This study assists technical and vocational HEIs to distinguish the important factors associated with the academic performance of outstanding students (international students) and poor students (students with special needs). This study also suggests some measures for educational teams to expand international enrollment and maintain student retention. For future research, other EDM approaches can be used to predict the minority students’ performance and compare them to the outcomes of this study. Other influencing features, such as admission criteria (i.e., entrance exam scores and high school GPA) and educational background should be considered and recommended for future research in order to better predict applicants’ future academic performance before admission.

Author Contributions

Conceptualization, T.-T.H.-C. and L.-S.C.; methodology, T.-T.H.-C.; software, T.-T.H.-C.; validation, T.-T.H.-C. and K.-V.H.; formal analysis, T.-T.H.-C.; writing—original draft preparation, T.-T.H.-C.; writing—review and editing, L.-S.C.; visualization, K.-V.H.; supervision, L.-S.C.; project administration, L.-S.C.; funding acquisition, L.-S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Science and Technology Council, Taiwan (Grant No. MOST 111-2410-H-324-006).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Shahiri, A.M.; Husain, W. A review on predicting student’s performance using data mining techniques. Procedia Comput. Sci. 2015, 72, 414–422. [Google Scholar] [CrossRef]
  2. Hwang, K.P.; Wang, M.K.; Sodanine, S. The effects of stressors, living support, and adjustment on learning performance of international students in Taiwan. Soc. Behav. Personal. Int. J. 2011, 39, 333–344. [Google Scholar] [CrossRef]
  3. Hübner, N.; Trautwein, U.; Nagengast, B. Should I stay or should I go? Predictors and effects of studying abroad during high school. Learn. Instr. 2021, 71, 101398. [Google Scholar] [CrossRef]
  4. Chou, C.P.; Roberts, A.; Ching, G.S. A study on the international students’ perception and norms in Taiwan. Int. J. Res. Stud. Educ. 2012, 1, 71–84. [Google Scholar] [CrossRef]
  5. Roberts, A.; Chou, P.; Ching, G. Contemporary trends in East Asian higher education: Dispositions of international students in a Taiwan university. High. Educ. 2010, 59, 149–166. [Google Scholar] [CrossRef]
  6. Ministry of Education, Taiwan (R.O.C.). International Students Come Together to Taiwan, Where the One-Hundred Thirty Thousand Overseas Student Mark Has Been Passed. Available online: https://english.moe.gov.tw/cp-117-25416-d8868-1.html (accessed on 31 May 2022).
  7. Lewis, A.; Ching, G.S.; Su, Y.N. A case study on the international students’ social adaptability in Taiwan: A qualitative study. Int. J. Res. Stud. Psychol. 2013, 2, 13–24. [Google Scholar] [CrossRef]
  8. International Students in Taiwan at an All-Time High. Available online: http://www.taipeitimes.com/News/taiwan/archives/2008/02/12/2003400913 (accessed on 31 May 2022).
  9. Education in Taiwan 2016–2017: Ministry of Education-Republic of China. Available online: https://stats.moe.gov.tw/bookcase/Education_in_Taiwan/105/index.html (accessed on 31 May 2022).
  10. Romero, C.; Ventura, S. Data mining in education. Wiley Interdis. Rev. Data Min. Know. Disco. 2013, 3, 12–27. [Google Scholar] [CrossRef]
  11. Baker, R.S.J.D. Data mining for education. Inter. Ency. Educ. 2010, 7, 112–118. [Google Scholar]
  12. Nguyen, A.; Gardner, L.; Sheridan, D. Data analytics in higher education: An integrated view. J. Inf. Syst. Educ. 2020, 31, 61. [Google Scholar]
  13. Chakrabarti, S.; Ester, M.; Fayyad, U.; Gehrke, J.; Han, J.; Morishita, S.; Piatetsky-Shapiro, G.; Wang, W. Data mining curriculum: A proposal (Version 1.0). Inten. Work. Group ACM SIGKDD Curricu. Commit. 2006, 140, 1–10. [Google Scholar]
  14. Asif, R.; Merceron, A.; Ali, S.A.; Haider, N.G. Analyzing undergraduate students’ performance using educational data mining. Comp. Educ. 2017, 113, 177–194. [Google Scholar] [CrossRef]
  15. Alapont, J.; Bella-Sanjuán, A.; Ferri, C.; Hernández-Orallo, J.; Llopis-Llopis, J.D.; Ramírez-Quintana, M.J. Specialised tools for automating data mining for hospital management. In Proceedings of the First East European Conference on Health Care Modelling and Computation, Craiova, Romania, 31 August–2 September 2005. [Google Scholar]
  16. Tarik, A.; Aissa, H.; Yousef, F. Artificial Intelligence and Machine Learning to Predict Student Performance during the COVID-19. Procedia Comp. Sci. 2021, 184, 835–840. [Google Scholar] [CrossRef] [PubMed]
  17. Umadevi, B.; Dhanalakshmi, R. A comprehensive survey of students performance using various data mining techniques. Intern. J. Sci. Res. 2017, 6, 2233–2238. [Google Scholar]
  18. Zorić, A.B. Benefits of educational data mining. J. Int. Bus. Res. Mark. 2020, 6, 12–16. [Google Scholar] [CrossRef]
  19. Greller, W.; Drachsler, H. Translating learning into numbers: A generic framework for learning analytics. Educ. Technol. Soc. 2012, 15, 42–57. [Google Scholar]
  20. Siemens, G.; Baker, R.S.D. Learning analytics and educational data mining: Towards communication and collaboration. In Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, Vancouver, BC, Canada, 29 April–2 May 2012. [Google Scholar]
  21. Romero, C.; Ventura, S. Educational data mining: A survey from 1995 to 2005. Expert Syst. Appl. 2007, 33, 135–146. [Google Scholar] [CrossRef]
  22. Baker, R.S.; Yacef, K. The state of educational data mining in 2009: A review and future visions. J. Educ. Data Min. 2009, 1, 3–17. [Google Scholar]
  23. Upadhyay, N.; Katiyar, V. A survey on the classification techniques in educational data mining. Int. J. Comput. Appl. Technol. Res. 2014, 3, 725–728. [Google Scholar]
  24. Huynh-Cam, T.T.; Chen, L.S.; Le, H. Using decision trees and random forest algorithms to predict and determine factors contributing to first-year university students’ learning performance. Algorithms 2021, 14, 318. [Google Scholar] [CrossRef]
  25. Tomasevic, N.; Gvozdenovic, N.; Vranes, S. An overview and comparison of supervised data mining techniques for student exam performance prediction. Comp. Educ. 2020, 143, 103676. [Google Scholar] [CrossRef]
  26. Namoun, A.; Alshanqiti, A. Predicting student performance using data mining and learning analytics techniques: A systematic literature review. Appl. Sci. 2021, 11, 237. [Google Scholar] [CrossRef]
  27. Albreiki, B.; Zaki, N.; Alashwal, H. A systematic literature review of student’ performance prediction using machine learning techniques. Educ. Sci. 2021, 11, 552. [Google Scholar] [CrossRef]
  28. Kabakchieva, D. Student performance prediction by using data mining classification algorithms. Int. J. Comput. Sci. Manag. Res. 2012, 1, 686–690. [Google Scholar]
  29. Chen, W.K.; Chen, L.S.; Pan, Y.T. A text mining-based framework to discover the important factors in text reviews for predicting the views of live streaming. Appl. Soft Comp. 2021, 111, 107704. [Google Scholar] [CrossRef]
  30. Al-Barrak, M.A.; Al-Razgan, M. Predicting students final GPA using decision trees: A case study. Int. J. Inf. Educ. Technol. 2016, 6, 528. [Google Scholar] [CrossRef]
  31. Matzavela, V.; Alepis, E. Decision tree learning through a predictive model for student academic performance in intelligent m-learning environments. Comput. Educ. Artif. Intell. 2021, 2, 100035. [Google Scholar] [CrossRef]
  32. Khosravi, H.; Shum, S.B.; Chen, G.; Conati, C.; Tsai, Y.-S.; Kay, J.; Knight, S.; Martinez-Maldonado, R.; Sadip, S.; Gašević, D. Explainable Artificial Intelligence in education. Comput. Educ. Artif. Intell. 2022, 3, 100074. [Google Scholar] [CrossRef]
  33. Chauvin, Y.; Rumelhart, D.E. Backpropagation: Theory, Architectures, and Applications; Psychology Press: London, UK, 2013. [Google Scholar]
  34. Hamoud, A.; Hashim, A.S.; Awadh, W.A. Predicting student performance in higher education institutions using decision tree analysis. Int. J. Interact. Multimed. Artif. 2018, 5, 26–31. [Google Scholar] [CrossRef] [Green Version]
  35. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  36. Abubakar, Y.; Ahmad, N.B.H. Prediction of students’ performance in e-learning environment using random forest. Int. J. Innov. Comput. 2017, 7, 1–5. [Google Scholar]
  37. Chen, K.Y.; Chen, L.S.; Chen, M.C.; Lee, C.L. Using SVM based method for equipment fault detection in a thermal power plant. Comput. Ind. 2011, 62, 42–50. [Google Scholar] [CrossRef]
  38. Baashar, Y.; Alkawsi, G.; Ali, N.A.; Alhussian, H.; Bahbouh, H.T. Predicting student’s performance using machine learning methods: A systematic literature review. In Proceedings of the 2021 International Conference on Computer & Information Sciences (ICCOINS), Kuching, Malaysia, 13–15 July 2021. [Google Scholar]
  39. Niyogisubizo, J.; Liao, L.; Nziyumva, E.; Murwanashyaka, E.; Nshimyumukiza, P.C. Predicting student’s dropout in university classes using two-layer ensemble machine learning approach: A novel stacked generalization. Comput. Educ. Artif. Intell. 2022, 3, 100066. [Google Scholar] [CrossRef]
  40. Ren, J.; Hagedorn, L.S. International graduate students’ academic performance: What are the influencing factors? J. Int. Stud. 2012, 2, 135–143. [Google Scholar] [CrossRef]
  41. Huang, B.; Lu, H.; Zhu, R. Disabled peers and student performance: Quasi-experimental evidence from China. Econ. Educ. Rev. 2021, 82, 102121. [Google Scholar] [CrossRef]
  42. Balestra, S.; Eugster, B.; Liebert, H. Peers with special needs: Effects and policies. Rev. Econ. Stat. 2022, 104, 602–618. [Google Scholar] [CrossRef]
  43. Griffiths, L.; Worth, P.; Scullard, Z.; Gilbert, D. Supporting disabled students in practice: A tripartite approach. Nurse Educ. Pract. 2010, 10, 132–137. [Google Scholar] [CrossRef] [PubMed]
  44. Hersh, M. Evaluation framework for ICT-based learning technologies for disabled people. Comput. Educ. 2014, 78, 30–47. [Google Scholar] [CrossRef]
  45. Zainudin, N.; Abu, N.H.; Mansor, M.F.; Malaysia, K.; Perlis, K. Exploring “Backbone” Factors associated to Academic Performance among Disability Student’s in E-learning: An Empirical Analysis during Pandemic COVID-19. In Proceedings of the 3rd Conference on Technology & Operations Management, Sintok, Malaysia, 24–25 August 2021. [Google Scholar]
  46. Koutri, M.; Avouris, N.; Daskalaki, S. A survey on web usage mining techniques for web-based adaptive hypermedia systems. In Adaptable and Adaptive Hypermedia Systems; Idea Group Inc. (IGI): Calgary, AB, Canada, 2005; pp. 125–149. [Google Scholar]
  47. Zorrilla, M.E.; Menasalvas, E.; Marin, D.; Mora, E.; Segovia, J. Web usage mining project for improving web-based learning sites. In Computer Aided Systems Theory–EUROCAST 2005; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
  48. Chang, J.R.; Liang, H.Y.; Chen, L.S.; Chang, C.W. Novel feature selection approaches for improving the performance of sentiment classification. J. Ambient. Intell. Humaniz. Comput. 2020, 1–14. [Google Scholar] [CrossRef]
  49. Chen, M.Y.; Chang, J.R.; Chen, L.S.; Chuang, Y.J. Identifying the key success factors of movie projects in crowdfunding. Multi. Tools Appl. 2022, 81, 27711–27736. [Google Scholar] [CrossRef]
  50. Chang, J.R.; Chen, L.S.; Lin, L.W. A Novel Cluster based Over-sampling Approach for Classifying Imbalanced Sentiment Data. IAENG Int. J. Comput. Sci. 2021, 48, 1118–1128. [Google Scholar]
  51. Rodríguez-Hernández, C.F.; Musso, M.; Kyndt, E.; Cascallar, E. Artificial neural networks in academic performance prediction: Systematic implementation and predictor evaluation. Comput. Educ. Artif. Intell. 2021, 2, 100018. [Google Scholar] [CrossRef]
  52. Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar]
  53. Iqbal, A.; Aftab, S. A classification framework for software defect prediction using multi-filter feature selection technique and MLP. Int. J. Mod. Educ. Comput. Sci. 2020, 12, 18–25. [Google Scholar] [CrossRef]
  54. Muschelli, J. ROC and AUC with a binary predictor: A potentially misleading metric. J. Classif. 2020, 37, 696–708. [Google Scholar] [CrossRef]
  55. Ching, G.S.; Wang, W.L.; Wen, T.S. Study abroad tales: Experiences of international students in Taiwan. Int. J. Res. Stud. Psychol. 2016, 5, 63–79. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Experimental process.
Figure 1. Experimental process.
Bdcc 06 00094 g001
Figure 2. Numbers of employed students in each group.
Figure 2. Numbers of employed students in each group.
Bdcc 06 00094 g002
Figure 3. Class distribution in three cases. Note: EX = class “Excellent”, VG = class “Very good”, G = class “Good”, AVG = class “Average”.
Figure 3. Class distribution in three cases. Note: EX = class “Excellent”, VG = class “Very good”, G = class “Good”, AVG = class “Average”.
Bdcc 06 00094 g003
Figure 4. Numbers of samples in each group in Case 1: Origin.
Figure 4. Numbers of samples in each group in Case 1: Origin.
Bdcc 06 00094 g004
Figure 5. Numbers of samples in each group: (a) Case 2: MFC; (b) Case 3: ROM.
Figure 5. Numbers of samples in each group: (a) Case 2: MFC; (b) Case 3: ROM.
Bdcc 06 00094 g005
Figure 6. Gini feature importance: (a) Group 1: from SVM; (b) Group 2: from SVM; (c) Group 3: from RF.
Figure 6. Gini feature importance: (a) Group 1: from SVM; (b) Group 2: from SVM; (c) Group 3: from RF.
Bdcc 06 00094 g006aBdcc 06 00094 g006b
Table 1. The employed input and output data description and transformation.
Table 1. The employed input and output data description and transformation.
No.Feature NameFeature Description and Transferred ValuesNo.Feature NameFeature Description and Transferred Values
1Department1 = TCA, 2 = TCJ, 3 = CK,
4 = TCL, 5 = TDJ, 6 = TDN,
7 = TD4, 8 = TD5, 9 = TD6,
10 = TD7, 11 = TC6, 12 = TC7,
13 = TC8, 14 = TC9, 15 = TE1,
16 = TE2, 17 = TE3, 18 = TE4,
19 = TE5, 20 = TQ1, 21 = TF1,
22 = TJ2, 23 = TJ4, 24 = TF2,
25 = TF3, 26 = TF4, 27 = TJ9
9Main source of living
expenses
1 = Parents
2 = Family and friends support
3 = Self-earning
4 = Grants in- or outside the school
5 = Income from full-time job
6 = Family provided
7 = Income from part-time job
8 = Scholarships
9 = Student loans
2Gender1 = Male, 2 = Female10Student loan1 = Yes, 0 = No
3Numbers of
required credits
0–2311Tuition waiver1 = Yes, 0 = No
4Numbers of
elective
credits
1–1412Father’s
occupations
1 = Military
2 = Education
3 = Public
4 = Service
5 = Industry
6 = Business
7 = Agriculture
8 = Others
5Sick leave0–3613Father’s
education
1 = Junior high school and below
2 = High school
3 = Bachelor
4 = Master
5 = Specialist
6 = PhD
6Personal leave0–3314Mother’s
occupations
1 = Military
2 = Education
3 = Public
4 = Service
5 = Industry
6 = Business
7 = Agriculture
8 = Others
7Parent Average
income per month
1 = 25,000 NTD,
2 = 40,000 NTD,
3 = 60,000 NTD,
4 = 80,000~100,000 NTD,
5 = Above 100,000 NTD
15Mother’s
education
1 = Junior high school and below
2 = High school
3 = Bachelor
4 = Master
5 = Specialist
6 = PhD
8On-campus
accommodation
1 = Yes, 0 = No16Grade Point Average (GPA)1 = Excellent (90–100 points),
2 = Very Good (80–89 points),
3 = Good (70–79 points),
4 = Average (60–69 points),
5 = Poor (0–59 points)
Note: Feature 10 and 11 were not applied for Group 1 (international students).
Table 2. Classification results of Case 1 (Origin).
Table 2. Classification results of Case 1 (Origin).
PerformanceAccuracy (%)PrecisionRecallF1-Score
MeanSDMeanSDMeanSDMeanSD
MethodsGroup 1: International students
SVM43.000.000.230.000.260.000.240.00
MLP35.604.060.260.020.260.020.260.02
RF40.003.500.300.020.300.020.300.02
DT40.004.830.380.070.470.110.410.08
Group 2: Students with disabilities
SVM24.0010.750.180.230.240.140.160.13
MLP36.008.430.330.140.300.080.300.10
RF40.0014.140.430.200.410.190.380.16
DT30.006.670.220.100.250.110.581.20
Group 3: Local students
SVM51.000.000.460.000.260.000.240.00
MLP48.801.230.380.030.310.010.320.01
RF53.101.970.460.040.320.010.330.01
DT45.201.230.350.010.350.020.350.01
Table 3. Classification results of Case 2 (Minority focus and combination (MFC)).
Table 3. Classification results of Case 2 (Minority focus and combination (MFC)).
PerformanceAccuracy (%)PrecisionRecallF1-Score
MeanSDMeanSDMeanSDMeanSD
MethodsGroup 1: International students
SVM80.0026.670.700.390.800.260.730.35
MLP92.0013.980.940.100.930.120.920.14
RF90.0017.000.940.110.910.150.890.88
DT94.009.660.950.080.940.100.940.10
Group 2: Students with disabilities
SVM100.000.001.000.001.000.001.000.00
MLP96.3011.100.940.170.930.220.930.20
RF96.3011.100.940.170.930.220.930.20
DT96.3011.100.940.170.930.220.930.20
Group 3: Local students
SVM88.103.540.840.070.710.070.740.07
MLP87.505.870.810.070.750.060.770.07
RF92.102.020.890.040.810.070.840.06
DT85.603.840.750.080.770.080.760.08
Table 4. Confusion matrix results of Case 2.
Table 4. Confusion matrix results of Case 2.
True Positive RateFalse Negative RateFalse Positive RateTrue Negative Rate
(a) Group 1: International students (SVM)1001
(b) Group 2: Students with disabilities (SVM)1001
(c) Group 3: Local students (RF)0.4120.5880.0690.931
Note: Positive: “EX”; Negative: “POOR”.
Table 5. Classification results of Case 3 (Random oversampling the minority (ROM)).
Table 5. Classification results of Case 3 (Random oversampling the minority (ROM)).
PerformanceAccuracy (%)PrecisionRecallF1
MeanSDMeanSDMeanSDMeanSD
MethodsGroup 1: International students
SVM100.000.001.000.001.000.001.000.00
MLP100.000.001.000.001.000.001.000.00
RF96.607.170.970.060.970.060.970.07
DT100.000.001.000.001.000.001.000.00
Group 2: Students with disabilities
SVM100.000.001.000.001.000.001.000.00
MLP96.0012.650.980.080.970.100.960.13
RF98.006.320.980.050.980.050.980.06
DT94.0013.500.960.090.950.110.940.14
Group 3: Local students
SVM91.901.450.920.020.920.010.920.02
MLP94.900.740.950.010.950.010.950.01
RF98.600.520.990.010.990.010.990.01
DT95.200.790.950.010.960.010.950.01
Table 6. AUC results of Case 3.
Table 6. AUC results of Case 3.
RFSVMMLPDT
(a) Group 1: International students1.001.001.001.00
(b) Group 2: Students with disabilities1.001.000.831.00
(c) Group 3: Local students1.001.000.960.95
Table 7. Classification results of Case 3.
Table 7. Classification results of Case 3.
True Positive RateFalse Negative RateFalse Positive RateTrue Negative Rate
(a) Group 1: International students1.00001.00
(b) Group 2: Students with disabilities1.00001.00
(c) Group 3: Local students1.0000.0660.934
Note: Positive: “EX”; Negative: “POOR”.
Table 8. Comparison of re-sampling techniques (under-sampling and ROM).
Table 8. Comparison of re-sampling techniques (under-sampling and ROM).
PerformanceUnder-samplingROMMFC
AccuracyF1AccuracyF1AccuracyF1
MethodsGroup 1: International students
SVM86.630.86100.001.0080.000.73
MLP83.300.82100.001.0092.000.92
RF96.700.9796.600.9790.000.89
DT93.400.93100.001.0094.000.94
Group 2: Students with disabilities
SVM0.000.0100.001.00100.001.00
MLP20.000.2096.000.9696.300.93
RF40.000.4098.000.9896.300.93
DT0.000.0094.000.9496.300.93
Group 3: Local students
SVM86.600.8691.900.9288.100.74
MLP85.300.8594.900.9587.500.77
RF87.700.8798.600.9992.100.84
DT77.400.7695.200.9585.600.76
Table 9. The seven top-ranking important features associated with learning performance of three groups of students.
Table 9. The seven top-ranking important features associated with learning performance of three groups of students.
Group 1: International StudentsGroup 2: Students with DisabilitiesGroup 3: Local Students
1 No. of required credits1 Father occupations1 No. of required credits
2 Department2 Department2 Sick leave
3 Main source of living expenses3 Mother education3 Department
4 Father occupations4 No. of required credits4 Personal leave
5 Parent average income per month5 No. of elective credits5 Mother occupations
6 Numbers of elective credits 6 Father education6 Numbers of elective credits
7 Father’s education7 Mother occupations7 Father occupations
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huynh-Cam, T.-T.; Chen, L.-S.; Huynh, K.-V. Learning Performance of International Students and Students with Disabilities: Early Prediction and Feature Selection through Educational Data Mining. Big Data Cogn. Comput. 2022, 6, 94. https://doi.org/10.3390/bdcc6030094

AMA Style

Huynh-Cam T-T, Chen L-S, Huynh K-V. Learning Performance of International Students and Students with Disabilities: Early Prediction and Feature Selection through Educational Data Mining. Big Data and Cognitive Computing. 2022; 6(3):94. https://doi.org/10.3390/bdcc6030094

Chicago/Turabian Style

Huynh-Cam, Thao-Trang, Long-Sheng Chen, and Khai-Vinh Huynh. 2022. "Learning Performance of International Students and Students with Disabilities: Early Prediction and Feature Selection through Educational Data Mining" Big Data and Cognitive Computing 6, no. 3: 94. https://doi.org/10.3390/bdcc6030094

Article Metrics

Back to TopTop