Next Article in Journal
An Open Source Framework Approach to Support Condition Monitoring and Maintenance
Next Article in Special Issue
Load Calculation of the Most Loaded Rolling Element for a Rolling Bearing with Internal Radial Clearance
Previous Article in Journal
Preventing the Interaction between Coronaviruses Spike Protein and Angiotensin I Converting Enzyme 2: An In Silico Mechanistic Case Study on Emodin as a Potential Model Compound
Previous Article in Special Issue
A Novel Stacked Auto Encoders Sparse Filter Rotating Component Comprehensive Diagnosis Network for Extracting Domain Invariant Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bearing Fault Diagnosis Based on Improved Convolutional Deep Belief Network

1
School of Rail Transportation, Soochow University, Suzhou 215000, China
2
Wuxi Metro Group Co, Ltd., Wuxi 214000, China
3
State Key Laboratory of Mechanical System and Vibration, Shanghai Jiao Tong University, Shanghai 200000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(18), 6359; https://doi.org/10.3390/app10186359
Submission received: 29 July 2020 / Revised: 7 September 2020 / Accepted: 9 September 2020 / Published: 12 September 2020
(This article belongs to the Special Issue Bearing Fault Detection and Diagnosis)

Abstract

:
Mechanical equipment fault detection is critical in industrial applications. Based on vibration signal processing and analysis, the traditional fault diagnosis method relies on rich professional knowledge and artificial experience. Achieving accurate feature extraction and fault diagnosis is difficult using such an approach. To learn the characteristics of features from data automatically, a deep learning method is used. A qualitative and quantitative method for rolling bearing faults diagnosis based on an improved convolutional deep belief network (CDBN) is proposed in this study. First, the original vibration signal is converted to the frequency signal with the fast Fourier transform to improve shallow inputs. Second, the Adam optimizer is introduced to accelerate model training and convergence speed. Finally, the model structure is optimized. A multi-layer feature fusion learning structure is put forward wherein the characterization capabilities of each layer can be fully used to improve the generalization ability of the model. In the experimental verification, a laboratory self-made bearing vibration signal dataset was used. The dataset included healthy bearings, nine single faults of different types and sizes, and three different types of composite fault signals. The results of load 0 kN and 1 kN both indicate that the proposed model has better diagnostic accuracy, with an average of 98.15% and 96.15%, compared with the traditional stacked autoencoder, artificial neural network, deep belief network, and standard CDBN. With improved diagnostic accuracy, the proposed model realizes reliable and effective qualitative and quantitative diagnosis of bearing faults.

1. Introduction

With the explosive progress of modern science and industry, machinery and equipment in fields such as aerospace, rail, and wind power are becoming faster, more automated, and meticulous than before. However, increasingly complex operating conditions inevitably lead to failures. Thus, monitoring and diagnosing the health of rotating machinery and equipment to ensure operational safety and reliability are immensely necessary and urgent; fortunately, intelligent diagnosis approaches, the combination of classification algorithms, and signal processing techniques have produced promising results [1].
Based on vibration signal processing, the traditional fault diagnosis methods extract fault components from the initial noise signal by using rich professional knowledge. At present, signal processing methods, such as empirical mode decomposition (EMD) [2], wavelet packet transform [3], and morphological filter [4], are commonly used for time-frequency analysis. For separating bearing fault ingredients in a wind turbine gearbox, Dong et al. [5] applied improved convolutional neural network with anti-interference to rolling bearing performance degradation assessment. Gong et al. [6] proposed a novel deep learning method called improved convolutional neural networks and support vector machines with data fusion for intelligent fault diagnosis. To realize bearing fault detection, Shi et al. [7] formed a quantity matrix of bearing vibration signal features on the basis of EMD and local mean decomposition. In recent years, machine learning methods, such as support vector machines [8] and artificial neural networks (ANNs) [9], have been gradually introduced into mechanical fault diagnosis. Such an introduction has enabled the minimization of human intervention and reliance on professional skills for development toward an intelligent direction in this field.
As an urgent field in machine learning research, deep learning has begun to appear in bearing failure in recent years, including stacked autoencoder (SAE) [10], deep belief networks (DBNs) [11], and convolutional neural networks (CNNs) [12]. Zhao et al. [13] summarized the emerging research on machine health monitoring based on deep learning, then discussed new trends in machine learning monitoring methods of deep learning. In considering ways of enhancing the robust features for rotating machinery, Shen et al. [14] proposed an automatic learning method with robust features based on contractive autoencoder. Chen et al. [12] proposed cyclic spectral coherence and convolutional neural networks for bearing fault diagnosis. Shao et al. [15] compressed the data with autoencoders and constructed a convolutional deep belief network (CDBN) to diagnose bearing faults. In view of compressed sensing and deep learning theory, Wen [16] studied a fault diagnosis method of bearing vibration signals that can extract features automatically.
In summary, traditional methods have several shortcomings, as follows:
  • Traditional vibration signal processing and analysis methods rely on certain professional skills;
  • Existing shallow machine learning methods rely on the accuracy of manual feature extraction;
  • Improper selection of parameters for standard deep learning models can easily result in failure to effectively converge, thus, diagnostic accuracy is difficult to guarantee;
  • Existing research on the quantitative diagnosis of bearing fault is relatively inadequate compared with that on qualitative diagnosis.
It can be concluded that the general deep models have challenges: (1) a suitable signal preprocessing method is needed to enhance features; (2) necessary measures need to be taken during the model training process to make it more stable; (3) the models should determine not only the fault type, but also the fault degree. In response to these problems, a CDBN has the superiority of fast calculations and feature extraction. Based on the standard CDBN model, this study introduces Adam’s optimization algorithm and optimizes the structure of the CDBN. An improved CDBN model is proposed to enhance diagnosis accuracy. Unlike image processing, in the case of fault diagnosis, the raw data will probably be noisy. The feature extraction capability of a standard CDBN is insufficient, which is the reason of conducting band-pass filter to pre-process the data. The main constructions of the proposed method are listed as follows:
  • A band pass filter has been introduced to the preprocessing step to filter out noise;
  • Qualitative and quantitative diagnoses of bearing faults can be effectively implemented;
  • Both single fault and compound faults can be effectively identified;
  • A comparative experiment under different operation loads further confirmed the reliability of the model.
The rest of the paper is organized as follows. Section 2 contains an introduction of the theoretical background of a CDBN and the structure of the proposed method. The experimental results are discussed in Section 3. Finally, in Section 4, conclusions are summarized.

2. Theoretical Background and Proposed Method

2.1. Restricted Boltzmann Machine

DBN is composed of several restricted Boltzmann machines (RBMs). The definition of the RBM model is based on the energy function. It is a basic component of DBNs and a key preprocessing unit in deep learning [17]. RMB is a basic single-layer machine learning network and has extensive applications. The formulas involved in the following text refer to the study of Chen and Li [17]. As shown in Figure 1, an RBM structure has two layers, namely, visible and hidden layers. A connection exists between each neuron node in the adjacent layer, and in each layer, the neuron nodes remain in a connectionless state with each other. The connection weight between two layers is W. RBM neurons are Booleans, implying that only two states exist, 0 and 1. State 1 indicates the activation of a neuron, and state 0 indicates suppression. In a way, for a given visible and hidden layer neuron in a binary state, the energy function of the RBM is as follows:
E ( v , h ; θ ) = i = 1 n a i v i i = 1 n j = 1 m w i j v i h j j = 1 m b j h j
where θ = { W i j , a i , b j } is the parameter set of RBM, w i j denotes the weight of the ith cell in the visible layer and the jth cell in the hidden layer, a i denotes the offset of the ith unit of the visible layer, b j denotes the offset of the jth cell of the hidden layer, v i is the state of the ith unit in the visible layer, and h j is the state of the jth unit in the hidden layer.
After the parameters are determined, the joint probability distribution of (v,h) is exported by Formula (1), where Z(θ) represents the normalized constant.
P ( v , h ; θ ) = 1 Z ( θ ) e E ( v , h ; θ )
Z ( θ ) = v , h e E ( v , h ; θ )
From the joint probability distribution, the edge probability and the conditional probability of visible and hidden neurons are driven by Formulas (4) and (5), respectively:
P ( v ; θ ) = h e E ( v , h ; θ ) v h e E ( v , h ; θ ) P ( v | h ; θ ) = e E ( v , h ; θ ) v e E ( v , h ; θ )
P ( h ; θ ) = v e E ( v , h ; θ ) v h e E ( v , h ; θ ) P ( h | v ; θ ) = e E ( v , h ; θ ) h e E ( v , h ; θ )
Given a visible neuron and a hidden neuron state, the probability of activating the jth hidden neuron and the ith visible neuron are respectively provided as follows, where σ = 1 / ( 1 + e x ) :
p ( h j = 1 | v ; θ ) = h k ¹ j p ( h j = 1 , h k ¹ j , v ; θ ) h p ( h , v ; θ ) = σ ( i = 1 n w i j v i + b j )
p ( v i = 1 | h ; θ ) = v k ¹ i p ( v i = 1 , v k ¹ i , h ; θ ) v p ( v , h ; θ ) = σ ( j = 1 m w i j h j + a i )
The contrast divergence algorithm [18] is used in RBM during model training. The task of training RBM is to find the optimal value of parameter θ so that the edge probability of the visible neurons in the distribution represented by RBM is maximized; that is, the goal is to maximize the log-likelihood function.
θ = arg max θ v ln P ( v ; θ )

2.2. Convolutional Deep Brief Network and Its Improvement

2.2.1. Convolutional Restricted Boltzmann Machine

Convolutional RBM (CRBM) is an improvement on the basis of the original RBM, and the structure of CRBM is similar to that of RBM. CRBM is an improved model composed of two random variable matrix layers, namely, the visible and hidden layers. An image with salient features of the local receptive field and weight sharing is regarded as the input layer of CRBM. The hidden layers are locally connected with visible layers, and their weights are shared by convolution.
The CRBM model, as shown in Figure 2, comprises the view layer V, the hidden layer H, and the pooling layer P, for a total of three layers. We assume that the size of the input layer matrix is N v × N v , the number of groups of matrices in the hidden layer is k, each group is a binary array with the size of N H × N H , and N H 2 K hidden layer units are present. Each group of hidden layers is associated with an N w × N w size filter.
Figure 3 shows the process of obtaining the hidden layer from the visible layer. The size of the convolution kernel is 3 × 3 , and the hidden layer units are divided into k sub-matrices. W 1 , W 2 , , W K connects the visible and hidden layers. Each hidden layer unit represents a specific feature extracted from a neighborhood unit of a visible layer. Moreover, b k indicates the value of each hidden layer unit bias, and C is the bias globally shared by all visible units. We obtain the energy number of CRBM as Equation (9):
E ( v , h ) = k = 1 k h k ( w k v ) k = 1 k b k i , j h i j k c i , j v i j
where * indicates convolution, v i j denotes the input values of the ith visible layer unit and the jth hidden layer unit, h k denotes the kth hidden layer, h i j k denotes the values of the ith visible layer unit and the jth unit of the kth hidden layer, and w k denotes the convolution kernel of the kth hidden unit.
As with standard RBM, the conditional probability distribution is given as follows:
p ( h j k = 1 | v ) = σ ( ( W ˜ k v ) j + b k )
p ( v i = 1 | h ) = σ ( ( W k h k ) i + c )
where σ = 1 / ( 1 + e x ) , W ˜ j k W N w j + 1 k .
A CRBM is a single-layered network. It can be considered as the basis of a CDBN. Stacking multiple CRBMs, with the hidden layer of the previous RBM as the visible layer of the next CRBM, constitutes a CDBN. In each training, the RBM of the lowest layer is trained, one layer at a time, until the top layer.

2.2.2. CDBN and Its Improvement

Proposed by Lee in 2009, a CDBN is a network model that consists of a CRBM. Multiple CRBMs are connected to form a CDBN. The outcome of the previous CRBM layer is regarded as the input of its subsequent layer. The model fitting ability is further improved by multi-layer linking.
The objective function of a CDBN is to maximize the log-likelihood function θ (·) in order to get the optimal parameters, and reconstruction error is used to evaluate the model. The reconstruction error refers to the difference between the original data after Gibbs sampling with the training sample as the initial state and the distribution of the RBM model. A smaller reconstruction error indicates better training. Reconstruction error is given by Equation (12), with X ^ i and X i indicating the real output and ideal output:
E r r o r =   1 N i = 1 N ( X ^ i ( W k , b k ) X i ) 2
Based on the standard CDBN model, an improved structure is proposed. The standard CDBN uses only single-layer output features but ignores the comprehensive utilization of the features of each layer. As a consequence, the classification results are not representative. An improvement of the model connection mode is proposed in this study. The outputs of a two-layer CRBM are combined into a vector and input to the softmax classifier by multi-level feature fusion to utilize features, further improving classification accuracy.
Model training can be summarized according to the following steps:
  • Forward propagation:
    (a)
    Use CD algorithm to pre-train W and b and determine the opening and closing of the corresponding hidden element;
    (b)
    Propagate upward layer by layer, calculate the excitation value of each hidden element, and use the sigmoid function to complete the standardization;
  • Backpropagation:
    (a)
    Use the minimum mean square error criterion for the backward error propagation algorithm and update the parameters of the network.
    (b)
    Update the weight and bias of the network with Adam optimizer.

2.3. Band-Pass Filter for Signal Preprocessing

A band-pass filter is a commonly used data processing method to filter out clutter. A band-pass filter can retain frequency components in a certain frequency range [19] while filtering out those in other ranges. Its bandwidth is determined by parameter λ, following positive correlation.
Assuming that the input image size is N × N, its passband bandwidth is given by Equation (13) [19]:
f 0 = λ N
Therefore, the filter function is given by Equation (14), with x and y denoting the angular frequency of a two-dimensional filter [20]:
f i l t e r ( x , y ) = ρ ( x , y ) e ( ρ ( x , y ) / f 0 ) 4
where ρ ( x , y ) = x 2 + y 2 .
Figure 4 shows the spectrogram of a band-pass filter. It can effectively remove high-frequency noise and low frequency-connected domain parts in the image, retaining the texture features in the image and increasing the suitability of the samples for CNNs to process.

2.4. Fault Diagnosis Model Based on Improved CDBN

Based on the CDBN model in 2.2.2 2.2, we further improved the model according to two aspects: changes to the preprocessing stage and the introduction of an optimizer. First, in the preprocessing stage, we introduced a band-pass filter to process the original samples after fast Fourier transform (FFT) and folding. The center frequency of the band-pass filter is set to 50% of the bandwidth, and the passband width is set to 40% of the bandwidth. The spectrogram of the band-pass filter is shown in Figure 4. Second, during the model training phase, the Adam optimizer [20] is introduced, enabling the model to adjust to different learning rates for each parameter. For frequently changing parameters, smaller steps are applied, whereas larger steps are better for sparse parameters. This algorithm is used to further improve the model’s convergence speed and to reduce the number of model errors.
Figure 5 is the flowchart of the improved CDBN-based bearing fault diagnosis model proposed in this study. The input vibration signal is at a length of 1024. After FFT, it is normalized to between 0 and signals of different fault types and degrees of failure. Moreover, band-pass filtering is conducted to enhance texture features and reduce the noise effect. The CDBN model is a combination of two CRBM layers. For the first CRBM layer, the convolution kernel size is 7 × 7, the number of link weight matrices is 9, each size matches the size of the convolution kernel, the pooling kernel size is 2 × 2, and the maximum pooling method is used. The output of the first layer with a size of 13 × 13 is a feature map, and it is used as the input of the second CRBM layer. The second hidden layer adopts a size of 9 × 9 with 16 link weight matrices. The size of the second convolution kernel is 5 × 5, and the size of the second pooling kernel is 2 × 2. The proposed model also uses the maximum pooling method. At the end of the model, softmax is used for classification. The extracted features are mapped to 13 different types, divided according to fault types or degrees. Unlike the standard CDBN, before classification, the outputs of the first layer and second layer use feature fusion to further increase the diagnostic accuracy. Table 1 displays the specific parameters of the network.

3. Experimental Validation

3.1. Dataset Description

To confirm the performance of the proposed model in bearing fault diagnosis, a laboratory-made bearing fault test platform was used to collect multiple faults and vibration signals at various levels.
As shown in Figure 6, the test platform included variable frequency motors, normal bearings, test bearings, loading systems, and acceleration sensors. The test bearing model was 6205-2RS SKF, and specific bearing parameters are listed in Table 2. Single-point or compound faults were set on the bearing surface by wire cutting. Figure 7 shows the real pictures of four bearings with four different health conditions. Types of faults included outer-race faults, ball faults, inner-race faults, inner-race compound ball faults, outer-race compound ball faults, and inner-race compound outer-race faults (IO). In this table, IB indicates that inner-race fault and ball faults exist at the same time on the test bearing, both under a fault degree of 0.2 mm. In the same way, IO indicates that inner-race fault and outer-race fault exist at the same time, and OB indicates that outer-race fault and ball fault exist at the same time. Specific data sets are described in Table 3 and Table 4; dataset1 was collected under 0 kN load and dataset2 was collected under 1 kN load. The motor speed was set to 961 rpm, signal sampling frequency was set as 10 kHz, and the entire experiment was under a load of zero condition.
The original vibration signal was intercepted into samples with a length of 1024. From each label, 300 samples were obtained. Two-thirds of the total samples were used as training data, and the remaining third was used as testing data.
The raw data of the bearing vibration signals is shown in Figure 8. It can be observed that the signal is noisy and lacks key information. The first step of the data preprocessing is to fold it into the size of 32 × 32 by FFT. Figure 9 shows the state of 13 sample types after FFT processing and folding. The second step of data preprocessing is filtering and normalization, for the purpose of filtering out noise. Figure 10 shows the samples after band-pass filtering preprocessing. The figures clearly show that the noise was reduced without weakening the fault signal. By contrast, the texture features of the samples were enhanced, which is more conducive for convolutional neural networks to extract features.

3.2. Diagnosis Results and Comparative Analysis

The comparison experiment was split into two parts. The proposed model was first compared with SAE, ANN, and DBN. Then, it was compared with standard CDBN.

3.2.1. Comparison with SAE, ANN, and DBN

The test results of the proposed model are shown in Figure 11b,d. The proposed model can achieve 100% accuracy for most types and degrees of faults. By contrast, misclassifications occur in the case of detecting an IO fault type with a small degree, as a result of the features of the IO fault type are less obvious than those of other types of faults. Operating on the collected experimental data set, the proposed improved CDBN model generally achieved an accuracy of 98.15% with 0 kN load and 96.15% with 1 kN load.
Table 5 and Table 6 illustrate the comparison results of the proposed improved CDBN model with SAE, ANN, and DBN. The length of the input signal was 1024. A single hidden layer with 100 hidden neurons was used in SAE; ANN also had the same structure as SAE; DBN had two hidden layers, whose first and second hidden layers each have 100 neurons. The listed results indicate that the proposed model had an obvious advantage over the other models. When the detection performance of the other models was considerably decreased, the proposed model was still able to maintain a high detection accuracy.

3.2.2. Comparison with Standard CDBN

The test results of the standard CDBN are illustrated in Figure 11a,c. The figure shows that the improved CDBN was able to achieve higher accuracy than the standard CDBN, especially in outer-race faults and ball faults. However, for IO fault types, namely, label 12, the classification accuracies of both models had no obvious optimization because of the difficulty in catching distinguishable features. In general, the proposed improved CDBN model had better performance compared with the standard CDBN model.
Figure 12 reveals the comparison of the second-layer reconstruction error between the standard and the improved CDBN models. Take dataset1 for example. During training, the reconstruction errors of both models decreased smoothly with the increase of the training batch, whereas the improved model had a much smaller reconstruction error. The same phenomenon occurred in the reconstruction error of the first layer of the two models, and the training accuracy rate increased slightly with the increase of epoch. The difference between final training accuracy and test accuracy was about 2%, which is not much different.
A commonly used method for high-dimensional data analysis is t-distribution stochastic neighbor embedding (tSNE). tSNE can reduce any dimension to less than three at will. As shown in Figure 13, an intuitive feature distribution was obtained by using tSNE to visualize the features of the standard and improved CDBN.
The results indicate the obvious optimization effect of the proposed method on feature processing. CDBN integrates convolution operation into origin DBN, which makes it have a stronger feature extraction capability. On the basis of the standard CDBN, a filter was introduced to filter out noise, so that the improved CDBN model has a more stable learning ability. The feature distribution of the original data is relatively loose, and mixed regions exist without obvious class boundaries. These phenomena are not conducive to the subsequent classification. For instance, features of labels 6 and 7 are mostly muddled; thus, misclassification occurred, as shown in Figure 11. By contrast, the feature distribution is clustered tightly with the improved CDBN, and it has easily observed boundaries and a wider distance, which are helpful for further feature extraction and classification.
Reconstruction error is a key indicator for measuring model learning ability. Different optimizers were employed for comparative analysis, and the reconstruction errors by different optimizers are shown in Figure 14. As shown in the figure, during the training, the reconstruction error of the CDBN model with Adam optimizer dropped smoothly and had the smallest reconstruction error.

4. Conclusions

An improved CDBN-based fault diagnosis model was proposed for the effective extraction and learning of quantitative and qualitative features of different bearing fault types and sizes. The proposed model was able to extract deep features of the dataset, effectively implementing the diagnosis of multiple degrees and types of bearing faults. First, a band-pass filter was used to preprocess the original signal to obtain optimized features. Simultaneously, the Adam optimizer was introduced to speed up training and improve the model convergence speed. Finally, in combination with the softmax two-layer connection, multi-layer feature fusion was used to fully enable the feature representation capabilities of each layer of the model. Compared with the standard CDBN, the improved CDBN model displayed higher accuracy. The improved model also ensured the characterization of the learned features and ultimately enabled qualitative and quantitative diagnosis of bearing faults. The experimental results show that the proposed method had reduced training error, a smoother decline, and a higher diagnostic accuracy than SAE, ANN, and DBN. The proposed model had a positive impact on the classification of single and compound fault types of the datasets. Future research will focus on dealing with unbalanced data or small data, since these cases are more meaningful for practical applications. In future investigations, comparative experiments of different optimizers, other networks (such as recurrent neural network and generative adversarial networks), and attention-based mechanisms will be conducted.

Author Contributions

Conceptualization, S.L. and J.X.; Methodology, S.J., C.S. and X.S.; Validation, S.L., J.X. and D.W.; Formal Analysis, S.L. and C.S.; Data Curation, X.S., D.W. and Z.Z.; Writing – Original Draft Preparation, S.L., C.S. and D.W.; Writing – Review & Editing, X.S. and Z.Z.; Supervision, C.S. and Z.Z.; Funding Acquisition, C.S., D.W. and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (No. 51875376, 51875375), as well as the Suzhou Science Foundation (No. SYG201802).

Conflicts of Interest

The authors have no conflict of interest to declare.

References

  1. Jiang, X.; Shen, C.; Shi, J.; Zhu, Z. Initial center frequency-guided VMD for fault diagnosis of rotating machines. J. Sound Vib. 2018, 435, 36–55. [Google Scholar] [CrossRef]
  2. Kumar, P.S.; Kumaraswamidhas, L.A.; Laha, S.K. Selecting effective intrinsic mode functions of empirical mode decomposition and variational mode decomposition using dynamic time warping algorithm for rolling element bearing fault diagnosis. Trans. Inst. Meas. Control 2019, 41, 1923–1932. [Google Scholar] [CrossRef]
  3. Huang, W.T.; Kong, F.Z.; Zhao, X.Z. Spur bevel gearbox fault diagnosis using wavelet packet transform and rough set theory. J. Intell. Manuf. 2018, 29, 1257–1271. [Google Scholar] [CrossRef]
  4. Kaplan, K.; Kaya, Y.; Kuncan, M.; Minaz, M.R.; Ertunç, H. An improved feature extraction method using texture analysis with LBP for bearing fault diagnosis. Appl. Soft Comput. 2020, 87, 106019. [Google Scholar] [CrossRef]
  5. Dong, S.J.; Wu, W.L.; He, K.; Mou, X.Y. Rolling bearing performance degradation assessment based on improved convolutional neural network with anti-interference. Measurement 2020, 151, 107219. [Google Scholar] [CrossRef]
  6. Gong, W.; Chen, H.; Zhang, Z.; Zhang, M.; Wang, R.; Guan, C.; Wang, Q. A Novel Deep Learning Method for Intelligent Fault Diagnosis of Rotating Machinery Based on Improved CNN-SVM and Multichannel Data Fusion. Sensors 2019, 19, 1693. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Shi, Q.J.; Guo, X.Z.H.; Liu, D.S.H. Bearing fault diagnosis based on feature fusion and support vector machine. J. Electron. Meas. Instrum. 2019, 33, 104–111. [Google Scholar]
  8. Hongfeng, T.; Chaochao, Z.; Liu, Y.; Yajun, T. Decision Tree SVM Fault Diagnosis Method of Photovoltaic Diode-Clamped Three-Level Inverter. U.S. Patent Application 10/234,495, 19 March 2019. [Google Scholar]
  9. Moosavian, A.; Jafari, S.M.; Khazaee, M.; Ahmadi, H. A Comparison Between ANN, SVM and Least Squares SVM: Application in Multi-Fault Diagnosis of Rolling Element Bearing. Int. J. Acoust. Vib. 2018, 23, 432–440. [Google Scholar]
  10. Meng, Z.; Zhan, X.; Li, J.; Pan, Z. An enhancement denoising autoencoder for rolling bearing fault diagnosis. Measurement 2018, 130, 448–454. [Google Scholar] [CrossRef] [Green Version]
  11. Liang, T.; Wu, S.; Duan, W.; Zhang, R. Bearing fault diagnosis based on improved ensemble learning and deep belief network. J. Phys. Conf. 2018, 1074, 012154. [Google Scholar] [CrossRef]
  12. Chen, Z.; Mauricio, A.; Li, W.; Gryllias, K. A Deep Learning method for bearing fault diagnosis based on Cyclic Spectral Coherence and Convolutional Neural Networks. Mech. Syst. Signal Process. 2020, 140, 106683. [Google Scholar] [CrossRef]
  13. Zhao, R.; Yan, R.; Chen, Z.; Mao, K.; Wang, P.; Gao, R.X. Deep learning and its applications to machine health monitoring. Mech. Syst. Signal Process. 2019, 115, 213–237. [Google Scholar] [CrossRef]
  14. Shen, C.; Qi, Y.; Wang, J.; Cai, G.; Zhu, Z. An automatic and robust features learning method for rotating machinery fault diagnosis based on contractive autoencoder. Eng. Appl. Artif. Intell. 2018, 76, 170–184. [Google Scholar] [CrossRef]
  15. Shao, H.; Jiang, H.; Zhang, H.; Liang, T. Electric locomotive bearing fault diagnosis using a novel convolutional deep belief network. IEEE Trans. Ind. Electron. 2018, 65, 2727–2736. [Google Scholar] [CrossRef]
  16. Wen, J.T.; Yan, C.H.; Sun, J.D.; Qiao, Y. Bearing fault diagnosis method based on compressed acquisition and deep learning. Chin. J. Sci. Instrum. 2018, 39, 171–179. [Google Scholar]
  17. Chen, Z.; Li, W. Multi sensor Feature Fusion for Bearing Fault Diagnosis Using Sparse Autoencoder and Deep Belief Network. IEEE Trans. Ind. Electron. 2017, 66, 1693–1702. [Google Scholar]
  18. Hinton, G.E. Training Products of Experts by Minimizing Contrastive Divergence. Neural Comput. 2014, 14, 1771–1800. [Google Scholar] [CrossRef] [PubMed]
  19. Cui, L.; Wang, X.; Wang, H.; Ma, J. Research on Remaining Useful Life Prediction of Rolling Element Bearings Based on Time-Varying Kalman Filter. IEEE Trans. Ind. Electron. 2020, 69, 2858–2867. [Google Scholar] [CrossRef]
  20. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Structure of restricted Boltzmann machine.
Figure 1. Structure of restricted Boltzmann machine.
Applsci 10 06359 g001
Figure 2. Structure diagram of a convolutional restricted Boltzmann machine (CRBM).
Figure 2. Structure diagram of a convolutional restricted Boltzmann machine (CRBM).
Applsci 10 06359 g002
Figure 3. The process of obtaining hidden layers from the visible layer based on CRBM.
Figure 3. The process of obtaining hidden layers from the visible layer based on CRBM.
Applsci 10 06359 g003
Figure 4. The spectrogram of a band-pass filter.
Figure 4. The spectrogram of a band-pass filter.
Applsci 10 06359 g004
Figure 5. Bearing fault diagnosis model based on an improved convolutional deep belief network (CDBN). FFT; fast Fourier transform.
Figure 5. Bearing fault diagnosis model based on an improved convolutional deep belief network (CDBN). FFT; fast Fourier transform.
Applsci 10 06359 g005
Figure 6. Bearing fault diagnosis model based on an improved CDBN.
Figure 6. Bearing fault diagnosis model based on an improved CDBN.
Applsci 10 06359 g006
Figure 7. The real pictures of four bearings with four different health conditions.
Figure 7. The real pictures of four bearings with four different health conditions.
Applsci 10 06359 g007
Figure 8. Raw data of bearing vibration signals of different labels.
Figure 8. Raw data of bearing vibration signals of different labels.
Applsci 10 06359 g008
Figure 9. Visualization of the data after FFT.
Figure 9. Visualization of the data after FFT.
Applsci 10 06359 g009
Figure 10. Visualization of data after filtering and normalization. The main purpose of this step is to filter out noise.
Figure 10. Visualization of data after filtering and normalization. The main purpose of this step is to filter out noise.
Applsci 10 06359 g010
Figure 11. (a) Testing result of dataset1 using the standard CDBN. (b) Testing result of dataset1 using the improved CDBN. (c) Testing result of dataset2 using the standard CDBN. (d) Testing result of dataset2 using the improved CDBN.
Figure 11. (a) Testing result of dataset1 using the standard CDBN. (b) Testing result of dataset1 using the improved CDBN. (c) Testing result of dataset2 using the standard CDBN. (d) Testing result of dataset2 using the improved CDBN.
Applsci 10 06359 g011
Figure 12. Comparison of second-layer reconstruction error when training (a) standard CDBN and (b) improved CDBN.
Figure 12. Comparison of second-layer reconstruction error when training (a) standard CDBN and (b) improved CDBN.
Applsci 10 06359 g012
Figure 13. Feature visualization of (a) dataset1 using the standard CDBN, (b) dataset1 using the improved CDBN, (c) dataset2 using the standard CDBN, and (d) dataset2 using the improved CDBN.
Figure 13. Feature visualization of (a) dataset1 using the standard CDBN, (b) dataset1 using the improved CDBN, (c) dataset2 using the standard CDBN, and (d) dataset2 using the improved CDBN.
Applsci 10 06359 g013
Figure 14. Comparison of second-layer reconstruction error when training the CDBN with different optimizers: (a) Adam, (b) Gradient Descent, (c) Nesterov, (d) RMSprop.
Figure 14. Comparison of second-layer reconstruction error when training the CDBN with different optimizers: (a) Adam, (b) Gradient Descent, (c) Nesterov, (d) RMSprop.
Applsci 10 06359 g014
Table 1. Specific parameters of the network.
Table 1. Specific parameters of the network.
Learning Rate Batch SizeEpochNumber of Layer
0.001200202
Size of Convolution
Kernel in the 1st Layer
Size of Convolution
Kernel in the 2nd Layer
Size of Pooling
Kernel in the 2nd Layer
Pooling Method
7 × 75 × 52 × 2Maximum pooling
Table 2. Specifications of the test bearings.
Table 2. Specifications of the test bearings.
Contact Angle Roller DiameterNumber of Rollers
7.938 mm9
Outer DiameterInternal Diameter Pitch Diameter
52 mm25 mm38.5 mm
Table 3. Description of compound faults for dataset1. IB: inner-race fault and ball faults exist at the same time on the test bearing; IO: inner-race fault and outer-race fault exist at the same time; OB: outer-race fault and ball fault exist at the same time.
Table 3. Description of compound faults for dataset1. IB: inner-race fault and ball faults exist at the same time on the test bearing; IO: inner-race fault and outer-race fault exist at the same time; OB: outer-race fault and ball fault exist at the same time.
Fault TypeLoadFault Degree
(mm)
Train SampleTest SampleLabel
Normal0 kN\2001001
Outer race0.22001002
0.32001003
0.62001004
Ball fault0.22001005
0.32001006
0.62001007
Inner race0.22001008
0.32001009
0.620010010
IB0.220010011
IO0.220010012
OB0.220010013
Table 4. Description of compound faults for dataset2.
Table 4. Description of compound faults for dataset2.
Fault TypeLoadFault Degree
(mm)
Train SampleTest SampleLabel
Normal1 kN\2001001
Outer race0.22001002
0.32001003
0.62001004
Ball fault0.22001005
0.32001006
0.62001007
Inner race0.22001008
0.32001009
0.620010010
IB0.220010011
IO0.220010012
OB0.220010013
Table 5. Comparison of testing results of dataset1 under load 0 kN by different models.
Table 5. Comparison of testing results of dataset1 under load 0 kN by different models.
Label12345678910111213Average Accuracy %
Model
SAE100833302082100810010010010010071.23
ANN100967659061100100100100100010076.31
DBN10010097989087967991100100659592.15
Proposed model1001001001001009896961001001008610098.15
Table 6. Comparison of testing results of dataset2 under load 1 kN by different models.
Table 6. Comparison of testing results of dataset2 under load 1 kN by different models.
Label12345678910111213Average Accuracy %
Model
SAE100100100100959940100010010010010087.23
ANN907479380501009810020100010065.31
DBN0969590968943917194778510079.00
Proposed model98100100100739990100100100948610096.15

Share and Cite

MDPI and ACS Style

Liu, S.; Xie, J.; Shen, C.; Shang, X.; Wang, D.; Zhu, Z. Bearing Fault Diagnosis Based on Improved Convolutional Deep Belief Network. Appl. Sci. 2020, 10, 6359. https://doi.org/10.3390/app10186359

AMA Style

Liu S, Xie J, Shen C, Shang X, Wang D, Zhu Z. Bearing Fault Diagnosis Based on Improved Convolutional Deep Belief Network. Applied Sciences. 2020; 10(18):6359. https://doi.org/10.3390/app10186359

Chicago/Turabian Style

Liu, Shuangjie, Jiaqi Xie, Changqing Shen, Xiaofeng Shang, Dong Wang, and Zhongkui Zhu. 2020. "Bearing Fault Diagnosis Based on Improved Convolutional Deep Belief Network" Applied Sciences 10, no. 18: 6359. https://doi.org/10.3390/app10186359

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop