Next Article in Journal
An Empirical Analysis on DPRK: Will Grain Yield Influence Foreign Policy Tendency?
Previous Article in Journal
An Investigation of the Influencing Factors of Chinese WeChat Users’ Environmental Information-Sharing Behavior Based on an Integrated Model of UGT, NAM, and TPB
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Physical Fatigue Detection Using Entropy Analysis of Heart Rate Signals

1
School of Architecture and Built Environment, Deakin University, Geelong 3220, Australia
2
System Administrator, Dibrugarh University, Assam 786004, India
3
School of Architecture, Islamic Azad University Taft, Taft 8991985495, Iran
4
Institute for Intelligent Systems Research and Innovation (IISRI), Locked Bag 20000, Deakin University, Geelong 3220, Australia
5
Institute for Physical Activity and Nutrition (IPAN), School of Exercise and Nutrition Sciences, Deakin University, Geelong 3220, Australia
*
Author to whom correspondence should be addressed.
Sustainability 2020, 12(7), 2714; https://doi.org/10.3390/su12072714
Submission received: 21 January 2020 / Revised: 19 March 2020 / Accepted: 21 March 2020 / Published: 30 March 2020

Abstract

:
Physical fatigue is one of the most important and highly prevalent occupational hazards in different industries. This research adopts a new analytical framework to detect workers’ physical fatigue using heart rate measurements. First, desired features are extracted from the heart signals using different entropies and statistical measures. Then, a feature selection method is used to rank features according to their role in classification. Finally, using some of the frequently used classification algorithms, physical fatigue is detected. The experimental results show that the proposed method has excellent performance in recognizing the physical fatigue. The achieved accuracy, sensitivity, and specificity rates for fatigue detection are 90.36%, 82.26%, and 96.2%, respectively. The proposed method provides an efficient tool for accurate and real-time monitoring of physical fatigue and aids to enhance workers’ safety and prevent accidents. It can be useful to develop warning systems against high levels of physical fatigue and design better resting times to improve workers’ safety. This research ultimately aids to improve social sustainability through minimizing work accidents and injuries arising from fatigue.

1. Introduction

The safer and healthier a workplace is, the fewer probabilities of diseases, accidents, injuries, and low performance [1]. Each day, an average of 6000 people die as a result of work-related accidents or diseases, totaling more than 2.2 million work-related deaths a year [2]. Of these, about 350,000 deaths are from workplace accidents, and more than 1.7 million are from work-related diseases [2]. Studies and estimates by many countries and the International Labor Organization (ILO) have shown that the economic costs of work-related illness and injury would be equivalent to a range of 1.8%–6% of gross domestic product (GDP) [3].
Physical fatigue is the ephemeral inability of muscles to maintain ideal physical performance, and results from prolonged activity [4,5,6]. It is one of the most important and highly prevalent occupational hazards in different industries [5,7]. Fatigue can occur due to excessive physical workload and results in a temporary reduction in the capability of physical activities [8]. Fatigue is one of the major sources for reduction in productivity, poor quality of work, injuries, and accidents in workplaces [9,10]. According to the World Health Organization, in recent years, the effects of physical fatigue on human health have been accumulating. Although “tiredness” has been reported as a construction accident risk by several studies, less attention was paid to this risk [5]. Fatigue affects workers’ health and safety and needs to be measured and controlled to prevent severe injuries.
In recent years, several studies have been conducted to assess physical workload and fatigue in construction projects using various subjective and objective methods. In most of the previous studies, subjective methods, including interviews [5] and questionnaires [6,11], have been used. The subjective methods used in previous studies do not provide broad insights into the adverse effects of physical fatigue and have several disadvantages. For example, they may suffer from subjective bias, and some people may answer in such a way to support the researcher’s hypothesis [6,11,12]. Therefore, a significant shift in the methods and tools adopted in previous research is needed. Measurement of the physiological responses which are less susceptible to participant bias represents a viable alternative and objective method to measure the impacts of physical fatigue. The use of physiological measurements, in addition to questionnaires, is the fundamental step change that this research area requires to establish a defensible evidence base for producing safety guidelines regarding the impact of physical fatigue on human health.
The previous studies harnessed physiological signals to estimate fatigue level, since it results from physical overexertion and is associated with physiological symptoms [8]. In the previous studies, heart rate, oxygen consumption, muscle activity, skin temperature, electrodermal (EDA) activity, blood pressure, and inertial measurement units were used to assess physical fatigue [13,14,15,16,17,18,19]. Chang et al. [20] investigated the relationship between the heart rate and fatigue and suggested heart rate as an indicator to predict the extent of strains or hazards which construction workers encounter [20]. Aryal et al. [21] used heart rate, skin temperature, and an electroencephalogram (EEG) sensor to assess the level of fatigue. The classification accuracy based on features extracted from the average of skin temperature and heart rate data was 82% [21]. Maman et al. [22] used inertial measurement units (IMUs) and developed a data-driven approach to assess workers’ fatigue in simulated manufacturing tasks. Important features from the five sensor locations were selected, and they achieved high accuracy in estimating the Ratings of Perceived Exertions (RPE) scale [22]. Jebelli et al. [18] used EDA, skin temperature (ST), and a photoplethysmogram (PPG) to assess workers’ physical states under different workloads. The heart rate variability, percentage heart rate, and EDA level showed a clear difference in light and medium tasks [18]. Zhang et al. [23] assessed the feasibility of using jerk, the time derivative of acceleration, as an indicator of physical exertion as fatigue develops over the course of a demanding task. They concluded that jerk is a useful indicator of physical exertion and experience level [23].
According to the reviewed literature, among the previous physiological symptoms, heart rate is the most widely used method for estimating a worker’s fatigue level [24,25,26]. It can be used as an effective means of determining the physiological strain of subjects in applied field situations [24]. Heart rate has also been used in other fields to assess fatigue [27,28]. Based on the previous studies, it has been proved that there is an association between physical fatigue and various heart rate metrics, such as heart rate, heart rate variability, and heart rate reserve (%HRR) [15,19,21,24]. According to [29], one of the common symptoms of fast or slow heart rate is fatigue.
Therefore, the aim of this study is to fill the gap in the literature by adopting a novel analysis method to monitor workers’ physical fatigue using physiological measurements. To study the patterns of physiological signals while subjects perform different tasks, the authors extracted features using different entropies and statistical measures. The features that resulted in the highest fatigue prediction accuracy were selected. Then, different classification algorithms were used to create a suitable model that could be used for the prediction of physical fatigue in new cases.
The proposed method will help to recognize physical fatigue more accurately. Having accurate and real-time monitoring of physical fatigue may enhance workers’ safety and helps to prevent accidents. The proposed method may be useful for developing warning systems against high levels of physical fatigue and for designing better resting times to improve workers’ safety.
This research is organized as follows. First, in Section 2, the method is described. The data collection method, physically fatiguing tasks, entropy-based nonlinear features, statistical measures, information gain (IG) as our feature selection algorithm, and different classification algorithms are reviewed in this section. Then, our proposed method is explained in detail. In Section 3, experimental results are discussed. Finally, the conclusion and future work are discussed in Section 4.

2. Method

In this research, an integral design framework is used to detect physical fatigue. In the following, the method of data collection is first introduced. Then, the physically fatiguing tasks, different methods for entropy calculation, different methods of statistical measures, feature selection algorithm, and classification methods are introduced. Finally, our proposed method is described in detail.

2.1. Data Collection (Participants)

The data used in this research were collected earlier by Maman et al. using a wearable sensor (Polar CR800X). Their protocol consisted of three physically demanding tasks [22]. Five males and 3 females of different ages ranging from 18 to 62 years from the local community were engaged for a duration of 3.5 months. Table 1 presents more detailed demographic information of the participants. Out of eight participants, two were from manufacturing industries and the rest were students with exposure to different physical activities. The Review Board of the Buffalo University approved the procedures of the experiments. A huge sample size could not be generated because of the time constraints of the participants.

2.2. Physically Fatiguing Tasks

The prime goal of the study is to detect physical fatigue based on the data obtained from the wearable sensors. RPE [27] was used to label the data. Fatigue occurrence was detected using RPE analysis and a binary decision rule as follows [22]:
Y = { 0 ,   R P E < 15 1 ,   R P E       15 } .
In [22], participants did three experimental sessions. In each session, one physically fatiguing task was performed for three hours divided into one-hour periods. Between the one-hour periods, there was one minute of rest for the collection of subjective ratings using the Borg RPE scale [30]. The physically fatiguing tasks were categorized as:
  • Part Assembly Task (ASM): A simulation was applied for fine motor control; it comprised part assembly operation in this task. Participants, with the aid of visual instructions, built sub-assemblies by utilizing Erector Assembly Kits while executing the task. The working posture was a stationary standing position for the three one-hour periods while carrying out the task.
  • Supply Pickup and Insertion Task (SPI): This task required unscrewing and fastening the bolts by bending forward and moving with supplies to a bolt box. In several manufacturing units, this uncomfortable and common posture is used for a certain amount of time. Hence, this physical task was selected. Scientific literature confirmed that repeated lifting and bending generate high probabilities of disorders of the lower back.
  • Manual Material Handling (MMH): Bodily fatigue is generally caused by continuous walking, as reported by 45% of the workers of the industry [22]. The whole task involved selecting packages with various weights, viz. 26, 18, or 10 kg, shifting them to a two-wheeled trolley, moving it to another section, and stacking the weights at the destined location. The palletization of the package was done on the basis of the orders received. One minute was the median time for shifting one package in one cycle. Out of three scenarios, the participants completed two sets. During the three-hour period, each scenario comprised of shifting 18 packages, summing up to a total of 108 packages as a whole.

2.3. Entropy-Based Nonlinear Features

There are many published studies that have demonstrated the advantages of entropy over conventional methods [31]. Using entropies instead of or alongside statistical measures can improve the performance of diagnosing systems. Pandian and Srinivasa proved that entropy techniques are useful tools in diagnosing patients with heart disorders [32]. Farahabadi et al. used entropy measures and observed that the entropy in heart patients is higher compared to the healthy ones [33]. Entropy as a nonlinear analysis can also give additional information about the heart rate variability [34]. In this research, the aim is to detect fatigue using heart rate signals. Previous studies has achieved good performance and results using entropy measures.
There are different methods for calculating the entropy of data. In this section, some of these methods that are used in this research are reviewed.
  • Shannon Entropy: In Information Theory, Shannon entropy is considered as a basic entropy technique [35]. It calculates the average minimum number of bits needed to encode the string of symbols. It also takes into account the frequencies of symbols. The equation is:
    H ( X ) = i = 0 N 1 p i l o g 2 p i ,
    where for a given symbol, pi is the probability of occurrence.
    For each symbol, the minimum average number of bits measured as
    Number   of   bits = H ( X ) .
    Tsallis and Rényi entropies are the simplified versions of Shannon entropy.
  • Tsallis Entropy (Ets): Constantino Tsallis introduced the Tsallis entropy concept in 1998 [36]. It is identical to Havrda–Charvát structural α-entropy. This non-extensive and non-additive entropy generalizes the Boltzmann–Gibbs theory. Consequences and predictions can be derived using this entropy when applied in social, artificial, and natural complex systems. The Tsallis entropy can be applied in various applications, including the black hole entropy calculation uniqueness theorem, the Klein, Gordon, and Dirac equations, the thermo-statistics of overdamped motion of interacting particles, and anomalous diffusion.
    E ts = 1 α 1 ( 1 n 1 x     E n α ) ,       f o r   α     1 ,
    where the probability of occurrence, termed as En and pn, is the feature value of the feature p that has the range of values from p1 to px.
  • Rényi Entropy (Er): In information theory, the Hartley, Shannon, collision, and min entropies are generalized by the Rényi entropy. Entropies measure the randomness, uncertainty, and diversity of a system. Alfréd Rényi proposed this entropy [37]. This entropy conceptualized the generalized dimensions with reference to fractal dimension measurement. It is vital in statistics and ecology for index diversity. It is utilized for entanglement estimation in quantum information. The min entropy is applied in theoretical computer science for randomness extractors. The formula is given in the below equation for Rényi entropy.
    E r = 1 α 1 l o g ( n 1 x E n α ) ,       f o r   α     1
  • Log-Energy Entropy: The selection of the entropies must be based on the occasions applied in information theory. The log energy entropy (concentration in normalized entropy) measures the complexity intensity of the signals [35]. Let us consider s as the signal and si as the coefficient of s on the orthonormal basis. The additive cost function is termed as entropy E so that E(0) = 0 and E(s) = i E ( s i ). With this convention, the log energy entropy can be defined:
    log(0) = 0: E2(si) = log(si2).
    Hence,
    E 2 ( s ) = i l o g ( s i 2 )
  • Kraskov Entropy: The differential statistical entropy of random variable X with d dimensions in the continuous time domain and with unknown density function f(x) is defined as [38]
    H ( x ) = d x   μ ( x ) l o g μ ( x ) .
The probability distribution function for the distance to the k-nearest neighbor samples and xi can be applied to obtain the density function dx. Kraskov entropy can be obtained by measuring and estimating k-nearest neighbors (k-NN) entropy. It can be defined as follows:
H ^ ( X ) = ψ ( k ) + ψ ( N ) + l o g ( C d ) + d N i = 0 N l o g ε ( i ) ,
where Cd symbolizes the volume of a d-dimensional unit ball that depends on the sample space, (x1, x2,……, xn) are the d-dimensional n random samples of the variable X, ε ( i ) is the distance between the k-NN sample points in the d-dimensional sample space and sample xi, and ϕ ( t ) represents the digamma function. Non-linear features based on Kraskov entropy may be applicable in EEG signal analysis.

2.4. Statistical Measures

In addition to mean, variance, and standard deviation, we used two more tests named kurtosis and skewness.
  • Skewness: Skewness is a measure of the asymmetry of the probability distribution of a random variable about its mean. The D’Agostino test allows the testing of whether a given distribution is symmetric. When the distribution’s shape is observed, one or more peaks (modes) may be found. The volume of data at the left side makes the right tail longer, and the distribution is termed as skewed right or positively skewed. On the contrary, the distribution is called negatively skewed or skewed left if the peak is towards the right, which makes the left tail longer. Skewness can be defined as the quantification of difference in distribution from the normal distribution. Skewness is a number and has no unit. The moment coefficient of skewness of a set of data is defined as below [39]:
    Skewness: g1 = m3/m23/2,
    where m3 = ∑(x x ¯ )3/n, m2 = ∑(x x ¯ )2/n, n is the sample size, x ¯ is the mean, m2 is the variance, and m3 is the third moment of the dataset. The skewness may be termed as g1 = average of z3, where z = (x x ¯ )/σ, (σ is the standard deviation). A zero value of skewness makes the data perfectly symmetrical, which is very rare to find in real-world scenarios. Bulmer [40] suggested the following rules:
    • If the skewness is <–1 or >+1, then it is highly skewed distribution.
    • If the skewness is between +1/2 and +1 or between −1 and −1/2, it is moderately skewed distribution.
    • If the skewness is between −1/2 and +1/2, it is approximately symmetric distribution.
  • Kurtosis: Kurtosis, which involves the fourth moment, is another common measure of shape. The kurtosis is more effective than skewness in the case of outliers. Both tails enhance the kurtosis in the case of symmetric distribution. The kurtosis has no unit and is measured as a number like skewness. Balanda et al. [41] mentioned the tails and that “increasing kurtosis is associated with the movement of probability mass from the shoulders of distribution into its center and tails”. Westfall et al. [42] proposed that the central peak was not the reason for kurtosis, but rather the tail. If we change the exponent from 3 to 4 in the coefficient of skewness, the moment of the coefficient of the kurtosis formula is derived.
    Kurtosis a4 = m4/m22 and excess kurtosis g2 = a4 − 3,
    where:
    m 4   =   ( x     x ¯ ) 4 / n and m 2   =   ( x     x ¯ ) 2 / n .
The excess kurtosis of the normal distribution is zero. m4 is the fourth moment of the dataset, n is the sample size, m2 is the variance, and x ¯ is the mean.
The kurtosis can also be termed as the average of z4, where z = (x x ¯ )/σ, where σ is the standard deviation. The average value of z4 is always greater than or equal to 1, whereas the average value of z is always zero. It should be noted that identical observation for kurtosis can be tested with the Anscombe–Glynn test.

2.5. Information Gain (IG) as Our Feature Selection Algorithm

IG is an entropy-based feature selection technique extensively used in machine learning applications. IG estimates the amount of information that a feature provides about a class [43]. Unrelated features do not convey any information, while key features give us maximal information. IG results in entropy reduction. Entropy is defined in information theory. More information content results in higher entropy. Entropy is the common way of measuring impurity and can be defined as:
Entropy = i p i l o g 2 p i ,
where p i is the probability for class i. In a two-class problem, a good training set does not have all of the examples belonging to the same class for the entropy of a group. A total 50% in either class for the entropy of a group makes a good training set for learning. The vital attributes need to be extracted from the training feature vectors for differentiating between the learning classes. The ordering of attributes in the nodes of a decision tree may be utilized by the extraction of key features.

2.6. Classifier Algorithms

In this section, the different classification methods that are used in this research are reviewed briefly.
  • k-nearest neighbors (KNN): Both regression and classification problems can be tackled by the KNN algorithm, which is a supervised machine learning classifier [44]. The advantage of this algorithm is that it is easy to implement and simple. KNN assumes that similar things exist in close proximity. KNN classifies the unknown labels based on similarity measures. Pattern recognition and statistical estimation utilize this non-parametric method. Euclidean, Manhattan, and Minkowski distances can be used for continuous variables, whereas in the case of categorical variables, Hamming distance can be applied. The choice of k is dependent on the data. A larger value of k reduces the noise but makes less distinction between the class boundaries. Irrelevant features and noisy data degrade the performance of the algorithm. Hence, evolutionary algorithms may be utilized to select key features for improving classification results [45].
  • Naïve Bayes: The Bayes theorem of probability is a building block of the Naïve Bayes classification technique [46]. The classifier believes the presence of a feature to be unrelated to the existence of any other feature. It is useful for huge datasets and can perform better than any other sophisticated classifiers. It is easy to implement and performs nicely in the case of categorical variables as well as multi-class classification. In the training dataset, if a category is not present but is available in the test dataset, the model assigns a zero probability, which is termed as “Zero Frequency”. To overcome this situation, smoothing techniques like Laplace estimation may be used. In the real-life scenario, it is simply unlikely to have independent predictors as assumed by Naïve Bayes. This classifier is used in various applications, including sentiment analysis, spam filtering, text classification, and recommendation systems [45].
  • Decision Tree: The decision tree is a pattern recognition and machine learning technique that can be applied to represent decisions visually [47]. It can be utilized in both classification and regression and is a tree-like model of decision making. This flowchart-like structure has three types of nodes. The nodes are decision, chance, and end nodes, represented by squares, circles, and triangles, respectively. It is commonly utilized in operations management and research. It can be used for measuring conditional probabilities as well. The decision tree is related to the influence diagram. It generates understandable rules and can perform the classification task without rigorous computation. It can provide the vital fields necessary for classification and regression. The disadvantages include that it is unstable, relatively inaccurate, and biased for attributes with more levels. If many outcomes are linked and most of the values are uncertain, then the calculations can be very complex in the case of this classifier [45].
  • Random Forest: Various individual decisions trees working as an ensemble constitutes the random forest classification technique [48]. The individual tree spits out a class prediction and the class with the maximum votes becomes the model’s prediction. The basic idea of the random forest is that many uncorrelated models operating as a unit can perform better than the individual constituent models. It corrects the overfitting problem of a decision tree that exhibits the training set. Linear models, such as Naïve Bayes and multinomial logistic regression, can be applied as base estimators instead of using decision trees. The random forest algorithms can be designed as kernel methods, which are easier to analyze and interpret [45].
  • Rule Induction: Rule induction extracts the formal rules from a set of observations [49]. Based on the extracted rules, patterns in the data can be represented to derive a scientific model. It is the extraction of statistically significant if–then rules. Some of the prime rule induction paradigms are Boolean decomposition, inductive rule programming, rough set rules, version spaces, horn clause induction, hypothesis testing algorithms, decision rule algorithms, and association rule learning algorithms. CN2, Progol, Rulex, and Charade are some of the rule induction algorithms [45].
  • Neural Network: A neural network is a set of algorithms that try to extract the hidden patterns and relationships in datasets through a process related to human brain functions [50]. It can adapt the changes in input criteria without changing the output and generate the best possible results. The concept of a neural network is based on artificial intelligence. A neuron in this network is a mathematical function that classifies and collects the information based on precise architecture. It contains layers with interconnected nodes, which are a perceptron. The perceptron feeds the signal to an activation function. The activation function may be non-linear in nature. Hidden layers fine-tune the input weighting and extrapolate relevant features that have predictive power related to the output. Applications of the neural network include pattern recognition, medical diagnosis, financial applications, sequence recognition, game playing, object recognition, non-linear system identification, and control [45].
  • Linear Regression: The foundation of linear regression is the linear relationship between one or more predictors and a target variable [51]. In this method, the assumption of linearity of the continuous variables is considered. Simple and multiple are the two types of this classifier. The finding of a relationship between two continuous variables is defined as simple linear regression. One is the dependent or response variable while the other is the independent or predictor variable. If one variable can be precisely expressed by the other, then the relationship is termed as deterministic. Statistical relations are not always accurate for relationship determination between two variables, such as height and weight relationships. When there are multiple input variables, the regression method is termed as multiple. The basic idea is to generate a line that best fits the data. The overall prediction error should be as small as possible in the best fit line. The distance of a point from the regression line is the error. Least squares techniques are often used for fitting linear regression models. They may be fitted by some other approaches, like minimizing a penalized type of least squares cost function, as in lasso and ridge regression, or minimizing the lack of fit, such as least absolute deviation regression. In linear regression, all observations are subject to normal errors and constant variance in linear regression, and the number of outliers and high-leverage points should be minimized [45]. Then, the relationship between the continuous variables and the dependent variables can be analyzed.
  • Logistic Regression: When the target variable is dichotomous in nature, we can apply logistic regression [52]. This predictive analysis describes data and finds the relationship between one target variable and one or more nominal, ordinal, ratio-level, or interval-independent variables. There should be no outliers in the data and no high correlations among predictors [45].
  • Linear Discriminant Analysis (LDA): LDA is a dimensionality reduction method that reduces the number of variables (dimensions) in a dataset while retaining useful information [53]. LDA is the linear classification method used when there are more than two classes. LDA works on the assumption that the data are Gaussian and each feature has the same variance. LDA calculates the variable and mean of the data for each class with these assumptions. LDA can make predictions based on probability, and the output class has the highest probability. There are many variations and extensions of this method. Flexible Discriminant Analysis (FDA), Regularized Discriminant Analysis (RDA), and Quadratic Discriminant Analysis (QDA) are such variations and extensions. It has many real-world applications, such as medical data analysis, face recognition, and customer identification [45].

2.7. Proposed Method

In our proposed method, at first, we extracted 10 features in different conditions, such as different categories and window lengths, using different entropy methods, such as Shannon, Tsallis, Rényi, Log-Energy, and Kraskov, and statistical measures like mean, variance, standard deviation, kurtosis, and skewness. The window type was fixed. The different categories are ASM, MMH, and SPI. Meanwhile, the different window lengths used in this research were 25, 50, and 125 s. Then, the extracted data were divided into training, validation, and test sets according to ten-fold cross-validation. The training, validation, and test data were 0.8, 0.1, and 0.1 of the data in each fold. The classification algorithms were trained by the trained section of data. Overfitting was controlled by the validation data. Finally, the trained classification algorithms were tested based on the test data. The performance of the designed system was calculated by the mean of all folds. The flowchart of our proposed method is illustrated in Figure 1.

3. Experimental Results

To test the proposed method on the dataset collected from the eight subjects, we used three hours of heart rate signals of each subject. During these three hours, each subject may be fatigued at different points of time. In other words, a subject may feel fatigued several times during the three-hour test. The window size depends on the fundamental frequency, intensity, and changes of the signal [32]. Some of window sizes were tested to see which ones have more accurate results than others. The recorded heart rate signals were divided by window lengths of 25, 50, and 125 s. The overall number of data extracted from these samples is calculated as 8     3     3600 window   length . If the window length is 25, 50, or 125 s, we will have 3456, 1728, or 691 samples, respectively. This process will increase the number of samples and aids in making the application of machine learning algorithms in the manuscript strong. The labels of each of these samples are fatigued and non-fatigued. As the subjects’ states change and they may be fatigued or non-fatigued during the test, our data are balanced. For example, in the MMH category, when the window size is 25, there are 42.27% fatigued and 57.73% non-fatigued samples.
The nine best classification algorithms were tested on the extracted data in different sample rates for different categories of ASM, MMH, and SPI tasks. The window lengths that were used were 25, 50, and 125. Ten-fold cross-validation was used on the extracted data to evaluate the performance of the algorithms. We used the accuracy, sensitivity, specificity, and area under the curve (AUC) of the algorithms for their performance comparison [54,55,56]. The accuracy of these algorithms is shown in Table 2. In this table, changing the color from green to red shows the gradient of the accuracy of the algorithms. The green color shows a lower accuracy level and the red color shows a higher accuracy level. According to this table, SPI accuracy is low, while ASM has a good accuracy rate.
For a better comparison between different algorithms in various conditions, the accuracy of algorithms is also illustrated in Figure 2. It is clear in this figure that the neural network has the best performance in almost all tasks. We achieved the best accuracy (90.36%) in the ASM category and with a sampling rate of 125. In this category, the neural network, rule induction, linear discriminant analysis (LDA), random forest, and k-nearest neighbor with k equal to 22 had the accuracies of 90.36%, 87.76%, 87.63%, 86.6%, and 86.57%, respectively. We tested different values of K in the KNN algorithm. K = 22 and K = 25 showed the best performance. Linear regression’s performance is not very good because it is suitable for continuous domains, not discrete ones. In addition, Naïve Bayes had the worst performance in all conditions because it is only suitable for high-dimensional feature matrices (high number of features). In each category, when the sampling rate was decreased, the accuracy of algorithms also decreased. Overall, ASM had better performance with respect to the other categories.
As shown in Figure 2, ASM (125) has the best accuracy for almost all of the classifiers. ASM (50) and ASM (25) are in the second and third ranks, respectively. As shown, MMH and SPI have the second and third best performance after ASM. In each category, the greater the window length is, the better performance is achieved. However, the effect of the categories is more than that of the window length, i.e., the performance of MMH (125) is less than that of ASM (25).
The differences between some methods, such as KNN and LDA in ASM (25), are very small. Therefore, one method cannot gain superiority over another method in practice.
In Table 3, the sensitivity of the investigated algorithms is shown. Sensitivity, also called the true positive rate or recall, estimates the proportion of actual positives that are correctly recognized as such [57]. Similar to Table 2, changing the color from green to red shows the gradient of the sensitivity of the algorithms. The green color shows lower sensitivity, while the red color shows higher sensitivity. According to the results shown in this table, Naïve Bayes has the best sensitivity rate. The reason that this algorithm does not have a good accuracy rate while its sensitivity is high is that it is biased towards labeling the data as fatigue. This leads to a reduction in specificity and accuracy of the algorithm. The neural network and rule induction algorithms are in the second and third ranks, respectively. Meanwhile, logistic regression, linear regression, and KNN (K = 22) have the lowest sensitivity. To achieve the best performance of KNN, we changed K (number of nearest neighbors) from 1 to 50. When K was 22 and 25, the algorithm achieved the best accuracy rate. So, in this research, we used these two values for K in the KNN algorithm. As we only have 10 features, while the number of our records is 300, the neural network has better performance with respect to the KNN algorithm [58]. KNN sensitivity is low because the number of records with target features labeled as fatigue is less than non-fatigue records.
Table 4 shows the specificity of the different algorithms used in this research. Specificity is also called the true negative rate and estimates the proportion of actual negatives that are correctly recognized as such [54]. Again, the neural network has the best specificity rate, followed by rule induction, KNN (K = 22), KNN (K = 25), and LDA. KNN’s specificity is high because the number of records labeled as fatigue is much more than non-fatigue records. So, it tends to classify records as non-fatigue. Meanwhile, according to this table, Naïve Bayes has the worst specificity rate. According to this table, SPI specificity is low, while ASM has a good specificity rate. In Table 5, the AUC can be seen for different algorithms.
The AUC measures the tradeoff between true positive rate and false positive rate. As the threshold varies over all possible values, the AUC is an evaluation of the classifier. It is a metric that checks the quality of the value generated by a classifier and then compares it to a threshold. It does not test the quality of a specific threshold value [58]. According to Table 5, ASM (125) has the best AUC in almost all of the classifiers. According to the AUC values, LDA, decision tree, and neural network have the best performance, respectively.
The ranking of features in different tasks is shown in Table 6, Table 7 and Table 8. In these tables, when the weight of a feature is high, it shows that this feature is more valuable in the classification of that record. Different feature selection algorithms have been suggested. We used one of the best algorithms, namely information gain [56,59,60]. According to the results shown in these tables, when window length was changed, different weights were assigned to features. As these tables show, in comparison with statistical measures, entropies commonly have better ranks in the classification of fatigue. For example, Shannon entropy is the best feature in six cases out of nine. Log-Energy entropy also has a high rank. Among the statistical measures, the standard deviation has a good rank. It has one first-place rank and seven second-place ranks according to these tables. Among statistical tests, mean has the worst rank, and among entropies, Tsallis entropy has the worst rank. Skewness and kurtosis usually do not have good ranks.

Discussion

In our research, the physiological data collected by Maman et al. [22] were used to detect workers’ physical fatigue. They used five different sensors’ features in their research. However, in this research, only the data extracted from the heart sensor were used. Many studies have demonstrated the advantages of entropy over conventional methods [28]. In contrast to Maman et al. [22], whose research which used only statistical measures, entropies are used in our proposed method. According to the achieved results, entropies are better than statistical measures for fatigue detection. In addition, ten-fold cross-validation was used in the proposed method to estimate the performance of the system. Maman et al. [22] used six out of eight cases for training the algorithms and the rest for testing. The most important weakness of the work of Maman et al. is that the accuracy of their proposed algorithm is not reported due to the small sample size. In addition, in contrast to Maman’s proposed method, in this research, we applied some of the frequently used classification algorithms. Using these powerful feature extraction and classification algorithms, we achieved the same results using the data prepared by only one sensor, while they used data of five sensors.
The results show that the extracted features of data have a significant influence on the accuracy of algorithms for fatigue detection. Some features such as standard deviation, Shannon entropy, and Log-Energy entropy have a high rank in feature selection algorithms. So, they can help classification algorithms a lot. An interesting point is that statistical measures have lower ranks with respect to entropies. Only the standard deviation has a good rank among statistical measures. Meanwhile, Tsallis entropy has the lowest rank between different entropies. Another interesting point is that in the ASM category, the algorithms can classify fatigue better than in MMH, and in the MMH category, classification can be done more precisely than in the SPI category. In addition, the greater the window length is, the better performance of the algorithms is achieved. Among the investigated algorithms, the neural network, rule induction, LAD, and KNN have good performance in fatigue detection.
Based on the achieved results, it is concluded that using entropies instead of or alongside statistical measures can improve the performance of diagnosing systems, such as fatigue detection. Moreover, using heart rate as the only indicator for fatigue assessment provided two benefits toward real-time detection of fatigue in the workplace. First, it is not required for the workers to wear several sensors. Second, since a lower amount of data is required to be processed, time and cost savings are achieved by using heart rate as the only sensor for fatigue detection.
The current findings are, however, subject to certain limitations. Specifically, the sample size used by the source publication [22] was small and, therefore, individual variation between participants (e.g., age, height, body mass) can skew the results. In future research, we suggest collecting more data to account for individual variation and to add greater generalizability to the more accurate predictions generated by our new analysis method.

4. Conclusions and Future Work

In this research, for fatigue detection, some important features were extracted from heart signals using entropies and statistical measures. The frequently used classification algorithms were applied to test the performance of the proposed method. Neural networks and rule-based induction could achieve excellent results in our tests. Overall, entropies extracted better features for classification in comparison with statistical measures.
It was shown that the proposed method provides an efficient tool for accurate and real-time monitoring of physical fatigue, aids in enhancing workers’ safety, and prevent accidents. The proposed method can be useful for developing warning systems against high levels of physical fatigue and designing better resting times to improve workers’ safety. Future research can utilize the proposed method in a user-friendly application to conduct this analysis so that project managers can look at workers’ fatigue in real time with a reliable accuracy.
Collecting more data can result in more accurate results. In future research, a larger dataset can be collected to improve accuracy and increase generalizability. As features play a vital role in the performance of an algorithm, extracting new features using other entropies may help to improve our proposed method’s accuracy. According to the achieved results, neural networks showed outstanding performance for fatigue detection. Therefore, deep learning algorithms provide a new horizon in this direction for future research.

Author Contributions

F.N., S.N., A.K., and B.A. contributed to the conception of the work. S.H., F.N., B.A., and A.K. contributed to statistical analysis and data interpretation. F.N., M.M., A.K., and M.T.D. contributed to drafting of the work. F.N. was the study supervisor. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fernandez, M.D.; Quintana, S.; Ballesteros, J.A.; Chavarria, N. Are workers in the construction sector overexposed to noise? Noise Vib. Worldw. 2010, 41, 11–14. [Google Scholar] [CrossRef]
  2. International Labor Organization. Facts on Safety at Work; International Labor Organization: Geneva, Switzerland, 2005. [Google Scholar]
  3. Takala, J.; Hämäläinen, P.; Saarela, K.L.; Yun, L.Y.; Manickam, K.; Jin, T.W.; Heng, P.; Tjong, C.; Kheng, L.G.; Lim, S. Global estimates of the burden of injury and illness at work in 2012. J. Occup. Environ. Hyg. 2014, 11, 326–337. [Google Scholar] [CrossRef] [PubMed]
  4. Hagberg, M. Muscular endurance and surface electromyogram in isometric and dynamic exercise. J. Appl. Physiol. 1981, 51, 1–7. [Google Scholar] [CrossRef] [PubMed]
  5. Chan, M. Fatigue: The most critical accident risk in oil and gas construction. Constr. Manag. Econ. 2011, 29, 341–353. [Google Scholar] [CrossRef]
  6. Fang, D.; Jiang, Z.; Zhang, M.; Wang, H. An experimental method to study the effect of fatigue on construction workers’ safety performance. Saf. Sci. 2015, 73, 80–91. [Google Scholar] [CrossRef]
  7. Nasirzadeh, F.; Soltanmohammadlou, N.; Sadeghi, S.; Khosravi, A. Occupational health-related risk factors in mining industry. In Proceedings of the 2nd International Conference on Civil, Architectural and Environmental Engineering(ICCAEE), Auckland, New Zealand, 9–11 December 2019; pp. 1–6. [Google Scholar]
  8. Sharpe, M. A report–chronic fatigue syndrome: Guidelines for research. J. R. Soc. Med. 1991, 84, 118–121. [Google Scholar] [CrossRef] [Green Version]
  9. Nojedehi, P.; Nasirzadeh, F. A hybrid simulation approach to model and improve construction labor productivity. KSCE J. Civ. Eng. 2017, 21, 1516–1524. [Google Scholar] [CrossRef]
  10. Nasirzadeh, F.; Mir, M.; Rostanmezhad, M.; Khosravi, A.; Carmichael, D. Monitoring labour fatigue using physiological measurements. In Proceedings of the 2nd International Conference on Civil, Architectural and Environmental Engineering(ICCAEE), Auckland, New Zealand, 9–11 December 2019; pp. 12–16. [Google Scholar]
  11. Techera, U.; Hallowell, M.; Littlejohn, R.; Rajendran, S. Measuring and Predicting Fatigue in Construction: Empirical Field Study. J. Constr. Eng. Manag. 2018, 144, 04018062. [Google Scholar] [CrossRef]
  12. Park, S.H.; Lee, P.J. Effects of floor impact noise on psychophysiological responses. Build. Environ. 2017, 116, 173–181. [Google Scholar] [CrossRef] [Green Version]
  13. Abdelhamid, T.S.; Everett, J.G. Physiological demands during construction work. J. Constr. Eng. Manag. 2002, 128, 427–437. [Google Scholar] [CrossRef]
  14. Gatti, U.C.; Schneider, S.; Migliaccio, G.C. Physiological condition monitoring of construction workers. Autom. Constr. 2014, 44, 227–233. [Google Scholar] [CrossRef]
  15. Hwang, S.; Lee, S. Wristband-type wearable health devices to measure construction workers’ physical demands. Autom. Constr. 2017, 83, 330–340. [Google Scholar] [CrossRef]
  16. Hwang, S.; Seo, J.; Jebelli, H.; Lee, S. Feasibility analysis of heart rate monitoring of construction workers using a photoplethysmography (PPG) sensor embedded in a wristband-type activity tracker. Autom. Constr. 2016, 71, 372–381. [Google Scholar] [CrossRef]
  17. Hwang, S.; Seo, J.; Ryu, J.; Lee, S. Challenges and opportunities of understanding construction workers’ physical demands through field energy expenditure measurements using a wearable activity tracker. In Proceedings of the Construction Research Congress, San Juan, Puerto Rico, 31 May–2 June 2016; pp. 2730–2739. [Google Scholar]
  18. Jebelli, H.; Choi, B.; Kim, H.; Lee, S. Feasibility study of a wristband-type wearable sensor to understand construction workers’ physical and mental status. In Proceedings of the Construction Research Congress, New Orleans, LA, USA, 2–4 April 2018; pp. 367–377. [Google Scholar]
  19. Jebelli, H.; Choi, B.; Lee, S. Application of wearable biosensors to construction sites. II: Assessing workers’ physical demand. J. Constr. Eng. Manag. 2019, 145, 04019080. [Google Scholar] [CrossRef]
  20. Chang, F.-L.; Sun, Y.-M.; Chuang, K.-H.; Hsu, D.-J. Work fatigue and physiological symptoms in different occupations of high-elevation construction workers. Appl. Ergon. 2009, 40, 591–596. [Google Scholar] [CrossRef] [PubMed]
  21. Aryal, A.; Ghahramani, A.; Becerik-Gerber, B. Monitoring fatigue in construction workers using physiological measurements. Autom. Constr. 2017, 82, 154–165. [Google Scholar] [CrossRef]
  22. Sedighi Maman, Z.; Alamdar Yazdi, M.A.; Cavuoto, L.A.; Megahed, F.M. A data-driven approach to modeling physical fatigue in the workplace using wearable sensors. Appl. Ergon. 2017, 65, 515–529. [Google Scholar] [CrossRef]
  23. Zhang, L.; Diraneyya, M.M.; Ryu, J.; Haas, C.T.; Abdel-Rahman, E.M. Jerk as an indicator of physical exertion and fatigue. Autom. Constr. 2019, 104, 120–128. [Google Scholar] [CrossRef]
  24. Gatti, U.C.; Migliaccio, G.C.; Bogus, S.M.; Schneider, S. An exploratory study of the relationship between construction workforce physical strain and task level productivity. Constr. Manag. Econ. 2014, 32, 548–564. [Google Scholar] [CrossRef]
  25. Mital, A.; Foononi-Fard, H.; Brown, M.L. Physical fatigue in high and very high frequency manual materials handling: Perceived exertion and physiological indicators. Hum. Factors 1994, 36, 219–231. [Google Scholar] [CrossRef]
  26. Wong, D.P.-L.; Chung, J.W.-Y.; Chan, A.P.-C.; Wong, F.K.-W.; Yi, W. Comparing the physiological and perceptual responses of construction workers (bar benders and bar fixers) in a hot environment. Appl. Ergon. 2014, 45, 1705–1711. [Google Scholar] [CrossRef] [Green Version]
  27. Duan, Z.; Xu, J.; Ru, H.; Li, M. Classification of Driving Fatigue in High-Altitude Areas. Sustainability 2019, 11, 817. [Google Scholar] [CrossRef] [Green Version]
  28. Lee, J.; Kim, D.; Ryoo, H.Y.; Shin, B.-S. Sustainable Wearables: Wearable Technology for Enhancing the Quality of Human Life. Sustainability 2016, 8, 466. [Google Scholar] [CrossRef] [Green Version]
  29. Lilly, L.S.; Braunwald, E. Braunwald’s Heart Disease: A Textbook of Cardiovascular Medicin; Elsevier Health Sciences: Amsterdam, The Netherlands, 2012. [Google Scholar]
  30. Borg, G. Borg’s Perceived Exertion and Pain Scales; Human kinetics: Champaign, USA, 1998. [Google Scholar]
  31. Hughes, M.; McCarthy, J.; Bruillard, P.; Marsh, J.; Wickline, S.J.E. Entropy vs. energy waveform processing: A comparison based on the heat equation. Entropy 2015, 17, 3518–3551. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Srinivasa, M.G.; Pandian, P.S. Application of Entropy Techniques in Analyzing Heart Rate Variability using ECG Signals. Int. J. Recent Innov. Trends Comput. Commun. 2019, 7, 9–16. [Google Scholar]
  33. Farahabadi, E.; Farahabadi, A.; Rabbani, H.; Dehnavi, A.M.; Mahjoob, M.P. An entropy-based method for ischemia diagnosis using ECG signal in wavelet domain. In Proceedings of the IEEE 10th International Conference on Signal Processing, Beijing, China, 24–28 October 2010; pp. 195–198. [Google Scholar]
  34. Beckers, F.; Ramaekers, D.; Aubert, A.E. Approximate Entropy of Heart Rate Variability: Validation of Methods and Application in Heart Failure. Cardiovasc. Eng. Int. J. 2001, 1, 177–182. [Google Scholar] [CrossRef]
  35. Acharya, U.R.; Hagiwara, Y.; Koh, J.E.W.; Oh, S.L.; Tan, J.H.; Adam, M.; Tan, R.S. Entropies for automated detection of coronary artery disease using ECG signals: A review. Biocybern. Biomed. Eng. 2018, 38, 373–384. [Google Scholar] [CrossRef]
  36. Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  37. Rényi, A. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 20 June–30 July 1960; pp. 1–15. [Google Scholar]
  38. Torse, D.; Desai, V.; Khanai, R. A Review on Seizure Detection Systems with Emphasis on Multi-domain Feature Extraction and Classification using Machine Learning. Broad Res. Artif. Intell. Neurosci. 2017, 8, 109–129. [Google Scholar]
  39. Joanes, D.N.; Gill, C.A. Comparing measures of sample skewness and kurtosis. Statistician 1998, 47, 183–189. [Google Scholar] [CrossRef]
  40. Bulmer, M.G. Principles of Statistics; Courier Corporation: New York, NY, USA, 1979; Volume 3. [Google Scholar]
  41. Balanda, K.P.; Macgillivray, H.L. Kurtosis: A Critical Review. Am. Stat. 1988, 42, 111–119. [Google Scholar] [CrossRef]
  42. Westfall, P.H. Kurtosis as Peakedness, 1905–2014. R.I.P. Am. Stat. 2014, 68, 191–195. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Lee, C.; Lee, G.G. Information gain and divergence-based feature selection for machine learning-based text categorization. Inf. Process. Manag. 2006, 42, 155–165. [Google Scholar] [CrossRef]
  44. Keller, J.M.; Gray, M.R.; Givens, J.A. A fuzzy K-nearest neighbor algorithm. IEEE Trans. Syst. ManCybern. 1985, 15, 580–585. [Google Scholar] [CrossRef]
  45. Tan, P.-N.; Steinbach, M.; Kumar, V. Introduction to Data Mining; Pearson Education India: New Delhi, India, 2016. [Google Scholar]
  46. Rish, I. An empirical study of the naive Bayes classifier. In Proceedings of the IJCAI-01 Workshop on Empirical Methods in Artificial Intelligence, Sicily, Italy, 4–8 August 2001; pp. 41–46. [Google Scholar]
  47. Safavian, S.R.; Landgrebe, D. A survey of decision tree classifier methodology. IEEE Trans. Syst. ManCybern. 1991, 21, 660–674. [Google Scholar] [CrossRef] [Green Version]
  48. Liaw, A.; Wiener, M. Classification and regression by randomForest. R News 2002, 2, 18–22. [Google Scholar]
  49. Langley, P.; Simon, H.A. Applications of machine learning and rule induction. Commun. Acm 1995, 38, 54–64. [Google Scholar] [CrossRef]
  50. Hansen, L.K.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [Google Scholar] [CrossRef] [Green Version]
  51. Seber, G.A.; Lee, A.J. Linear Regression Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2012; Volume 329. [Google Scholar]
  52. Hosmer Jr, D.W.; Lemeshow, S.; Sturdivant, R.X. Applied Logistic Regression; John Wiley & Sons: Hoboken, NJ, USA, 2013; Volume 398. [Google Scholar]
  53. Balakrishnama, S.; Ganapathiraju, A. Linear discriminant analysis-a brief tutorial. Inst. Signal Inf. Process. Manag. 1998, 18, 1–8. [Google Scholar]
  54. Alizadehsani, R.; Abdar, M.; Roshanzamir, M.; Khosravi, A.; Kebria, P.M.; Khozeimeh, F.; Nahavandi, S.; Sarrafzadegan, N.; Acharya, U.R. Machine learning-based coronary artery disease diagnosis: A comprehensive review. Comput. Biol. Med. 2019, 111, 103346. [Google Scholar] [CrossRef]
  55. Alizadehsani, R.; Roshanzamir, M.; Abdar, M.; Beykikhoshk, A.; Khosravi, A.; Panahiazar, M.; Koohestani, A.; Khozeimeh, F.; Nahavandi, S.; Sarrafzadegan, N. A database for using machine learning and data mining techniques for coronary artery disease diagnosis. Sci. Data 2019, 6, 227. [Google Scholar] [CrossRef] [PubMed]
  56. Alizadehsani, R.; Habibi, J.; Alizadeh Sani, Z.; Mashayekhi, H.; Boghrati, R.; Ghandeharioun, A.; Khozeimeh, F.; Alizadeh-Sani, F. Diagnosing Coronary Artery Disease via Data Mining Algorithms by Considering Laboratory and Echocardiography Features. Res. Cardiovasc. Med. 2013, 2, 133–139. [Google Scholar] [CrossRef] [PubMed]
  57. Alizadehsani, R.; Zangooei, M.H.; Hosseini, M.J.; Habibi, J.; Khosravi, A.; Roshanzamir, M.; Khozeimeh, F.; Sarrafzadegan, N.; Nahavandi, S. Coronary artery disease detection using computational intelligence methods. Knowl. Based Syst. 2016, 109, 187–197. [Google Scholar] [CrossRef]
  58. Mitchell, T. Machine Learning; McGraw Hill: New York, NY, USA, 1997; Volume 45. [Google Scholar]
  59. Alizadehsani, R.; Habibi, J.; Hosseini, M.J.; Boghrati, R.; Ghandeharioun, A.; Bahadorian, B.; Sani, Z.A. Diagnosis of coronary artery disease using data mining techniques based on symptoms and ecg features. Eur. J. Sci. Res. 2012, 82, 542–553. [Google Scholar]
  60. Khozeimeh, F.; Alizadehsani, R.; Roshanzamir, M.; Khosravi, A.; Layegh, P.; Nahavandi, S. An expert system for selecting wart treatment method. Comput. Biol. Med. 2017, 81, 167–175. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the proposed algorithm. * Shannon, Tsallis, Rényi, Log-Energy, and Kraskov were used. ** Variance, standard deviation, mean, kurtosis, and skewness were used.
Figure 1. The flowchart of the proposed algorithm. * Shannon, Tsallis, Rényi, Log-Energy, and Kraskov were used. ** Variance, standard deviation, mean, kurtosis, and skewness were used.
Sustainability 12 02714 g001
Figure 2. Accuracies of different algorithms in various conditions.
Figure 2. Accuracies of different algorithms in various conditions.
Sustainability 12 02714 g002
Table 1. Detailed demographic information of the participants (adopted from [22]).
Table 1. Detailed demographic information of the participants (adopted from [22]).
ParticipantGenderAgeHeight (m)Weight (kg)RHR (bpm)
P1Female181.664842
P2Male211.8979.765
P3Female291.7770.262
P4Male621.7188.871
P5Male231.7169.367
P6Male591.673.867
P7Male301.7272.274
P8Female191.6262.562
Resting heart rate (RHR).
Table 2. Accuracies (%) of different algorithms in various conditions.
Table 2. Accuracies (%) of different algorithms in various conditions.
Sampling Rate
2550125
Category Algorithm NameASMMMHSPIASMMMHSPIASMMMHSPI
KNN (K = 22)84.6772.257.3286.6773.459.8286.5773.2758.67
KNN (K = 25)84.8472.0557.4987.0673.2359.6986.5673.7160.50
Naïve Bayes80.1952.551.7979.8357.4450.3979.8564.751.09
Decision tree82.7272.7958.1683.9472.7158.0985.3572.9558.44
Random forest83.7274.8960.1684.1774.7962.0986.675.9560.44
Rule induction85.6675.4461.1986.6775.6161.3487.7679.6771.97
Neural network87.4176.1662.7588.5876.3463.4590.3677.8169.37
Linear regression83.4374.5661.0583.7775.6561.5986.2477.2864.72
Logistic regression82.8772.7758.3182.6975.8258.7882.7877.7159.2
LDA84.1574.5860.7985.6375.5261.7287.6377.2866.01
Table 3. Sensitivity (%) of different algorithms in various conditions.
Table 3. Sensitivity (%) of different algorithms in various conditions.
Sampling Rate
2550125
Category Algorithm NameASMMMHSPIASMMMHSPIASMMMHSPI
KNN (K = 22)69.6248.7955.3361.2154.1858.1864.6553.256.93
KNN (K = 25)77.3855.0554.8664.9759.955.5167.859.258.92
Naïve Bayes89.1390.0980.6885.3586.679.785.7381.685.05
Decision tree60.6750.8650.3779.3645.3250.3679.8160.3956.36
Random forest62.5355.4955.6575.2850.3861.4368.9965.5258.52
Rule induction66.8872.8957.1270.8468.1759.5373.4662.9571.09
Neural network71.1263.3859.5677.7974.2862.2382.2672.866.15
Linear regression59.8853.1650.0759.5555.5158.9160.1666.5461.93
Logistic regression51.1245.9758.860.2557.2653.151.2662.858.12
LDA56.7553.0156.5659.8754.2455.662.0866.459.58
Table 4. Specificity (%) of different algorithms in various conditions.
Table 4. Specificity (%) of different algorithms in various conditions.
Sampling Rate
2550125
Category Algorithm NameASMMMHSPIASMMMHSPIASMMMHSPI
KNN (K = 22)86.1688.4473.1496.1388.0675.4389.6388.1370.56
KNN (K = 25)85.885.898095.8285.6869.9289.5388.569.44
Naïve Bayes54.638.4538.9252.0246.5239.1451.549.9739.81
Decision tree85.1281.4970.7887.5679.7664.5289.7974.6275.79
Random forest89.1183.4274.7289.3780.4975.3891.2980.5977.42
Rule induction88.7683.8775.9394.385.8669.6894.989.9172.59
Neural network89.9988.4179.4494.9883.8470.3396.287.0971.67
Linear regression85.2287.5183.3484.2286.2667.9688.9581.8168.93
Logistic regression86.9589.9779.6385.8480.4668.9685.7483.0779.07
LDA88.2287.685.4288.4882.5570.5794.4383.1484.81
Table 5. Area under the curve (AUC) of different algorithms in various conditions.
Table 5. Area under the curve (AUC) of different algorithms in various conditions.
Sampling Rate
2550125
Category Algorithm NameASMMMHSPIASMMMHSPIASMMMHSPI
KNN (K = 22)0.820.780.740.850.770.690.860.810.72
KNN (K = 25)0.790.650.720.840.790.760.870.840.81
Naïve Bayes0.770.790.710.810.80.740.790.80.79
Decision tree0.870.810.740.880.870.820.910.890.81
Random forest0.830.780.710.860.830.820.890.850.83
Rule induction0.840.810.760.840.810.790.890.830.8
Neural network0.830.790.790.880.840.780.880.870.86
Linear regression0.840.790.730.860.80.720.890.80.79
Logistic regression0.790.760.770.840.810.780.860.830.73
LDA0.860.830.810.890.890.830.920.890.88
Table 6. The rankings of features in different categories when window length is 25.
Table 6. The rankings of features in different categories when window length is 25.
25-ASM25-MMH25-SPI
Feature NameWeightFeature NameWeightFeature NameWeight
Shannon entropy1.0Shannon entropy1.0Shannon entropy1.0
Standard deviation0.9703Standard deviation0.9382Standard deviation0.9937
Log-Energy entropy0.9624Log-Energy entropy0.8563Log-Energy entropy0.9935
Kraskov entropy0.4774Rényi entropy0.3515Kraskov entropy0.3784
Variance0.4774Tsallis entropy0.3487Variance0.3784
Rényi entropy0.4211Kraskov entropy0.3387Skewness0.2665
Skewness0.0867Variance0.3387Rényi entropy0.1909
Kurtosis0.0818Skewness0.1176Tsallis entropy0.1693
Tsallis entropy0.0789Kurtosis0.0088Kurtosis0.0136
Mean0.0Mean0.0Mean0.0
Table 7. The rankings of features in different categories when window length is 50.
Table 7. The rankings of features in different categories when window length is 50.
50-ASM50-MMH50-SPI
Feature NameWeightFeature NameWeightFeature NameWeight
Standard deviation1.0Shannon entropy1.0Shannon entropy1.0
Shannon entropy0.9984Standard deviation0.9287Log-Energy entropy1.0
Log-Energy entropy0.9828Log-Energy entropy0.8926Standard deviation1.0
Kraskov entropy0.4981Kraskov entropy0.4087Kraskov entropy0.7753
Variance0.4981Variance0.4087Variance0.7753
Rényi entropy0.4185Tsallis entropy0.3964Rényi entropy0.5393
Skewness0.2212Rényi entropy0.3854Tsallis entropy0.4087
Kurtosis0.1799Skewness0.2306Skewness0.2253
Mean0.1687Kurtosis0.0309Kurtosis0.1042
Tsallis entropy0.0Mean0.0Mean0.0
Table 8. The rankings of features in different categories when window length is 125.
Table 8. The rankings of features in different categories when window length is 125.
125-ASM125-MMH125-PSI
Feature NameWeightFeature NameWeightFeature NameWeight
Log-Energy entropy1.0Log-Energy entropy1.0Shannon entropy1.0
Standard deviation 1.0Standard deviation1.0Standard deviation0.8883
Shannon entropy0.9942Shannon entropy0.9942Log-Energy entropy0.8118
Kraskov entropy0.4857Kraskov entropy0.4857Tsallis entropy0.4038
Variance0.4857Variance0.4857Rényi entropy0.3942
Rényi entropy0.3420Rényi entropy0.3420Kraskov entropy0.3695
Skewness0.2744Skewness0.2744Variance0.3695
Mean0.2594Mean0.2594Skewness0.2287
Kurtosis 0.2108Kurtosis0.2108Kurtosis0.0016
Tsallis entropy0.0Tsallis entropy0.0Mean0.0

Share and Cite

MDPI and ACS Style

Nasirzadeh, F.; Mir, M.; Hussain, S.; Tayarani Darbandy, M.; Khosravi, A.; Nahavandi, S.; Aisbett, B. Physical Fatigue Detection Using Entropy Analysis of Heart Rate Signals. Sustainability 2020, 12, 2714. https://doi.org/10.3390/su12072714

AMA Style

Nasirzadeh F, Mir M, Hussain S, Tayarani Darbandy M, Khosravi A, Nahavandi S, Aisbett B. Physical Fatigue Detection Using Entropy Analysis of Heart Rate Signals. Sustainability. 2020; 12(7):2714. https://doi.org/10.3390/su12072714

Chicago/Turabian Style

Nasirzadeh, Farnad, Mostafa Mir, Sadiq Hussain, Mohammad Tayarani Darbandy, Abbas Khosravi, Saeid Nahavandi, and Brad Aisbett. 2020. "Physical Fatigue Detection Using Entropy Analysis of Heart Rate Signals" Sustainability 12, no. 7: 2714. https://doi.org/10.3390/su12072714

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop