Abstract

Intelligent fault diagnosis technology of the rotating machinery is an important way to guarantee the safety of industrial production. To enhance the accuracy of autonomous diagnosis using unlabelled mechanical faults data, a novel intelligent diagnosis algorithm has been developed for rotating machinery based on adaptive transfer density peak search clustering. Combined with the wavelet packet energy feature extraction algorithm, the proposed algorithm can enhance the computational accuracy and reduce the computational time consumption. The proposed adaptive transfer density peak search clustering algorithm can adaptively adjust the classification parameters and mark the categories of unlabelled experimental data. Results of bearing experimental analysis demonstrated that the proposed technique is suitable for machinery fault diagnosis using unlabelled data, compared with other traditional algorithms.

1. Introduction

Modern machinery is developing in the direction of high speed, precision, and efficiency. Intelligent fault diagnosis technology can ensure the safe operation of machines by automatically extracting the fault information implied in the monitoring data and intelligently identifying the health condition [1]. In the research of fault diagnosis of rotating machinery, vibration signal analysis is the most widely used method, which has less influence on the operation status and is easy to be collected. However, the complexity of vibration signal makes the research more difficult. Intelligent fault diagnosis technology is a comprehensive discipline developed in recent years and consists of three main components [2]: (a) signal acquisition, (b) feature extraction, and (c) fault identification. In recent years, some methods have been developed by vibration signals in the field of fault diagnosis. A deep transfer diagnosis method is proposed in [3] to identify the faults in a practical machine by using the diagnosis knowledge derived from the machine in a laboratory. In order to obtain better generalization capability, Wang et al. [4] combined integrated learning with differential probabilistic neural networks and further proposed a new method based on selective integrated learning with particle swarm optimization. In order to solve this issue of the separation of multiple faults, Hu et al. [1] developed a novel multifault detecting technique based on tensor factorization. Liu et al. [5] proposed a novel approach based on the classification to integrate all three aspects (impulsiveness, cyclostationarity, and health reference) for frequency band selection. The above-mentioned methods, however, suffer from low classification accuracy or cannot determine the target domain fault types for unlabelled data.

In the study of diagnosing for a small amount of unlabelled mixed data, transfer learning recently has been applied in the fields of mechanical fault diagnosis by researchers [6]. For example, transfer learning was utilized to detect machine tool chatter based on the detected features of a vibration signal via wavelet packet transform (WPT) and ensemble empirical modal decomposition (EEMD) [7]. Shao et al. proposed a method based on deep transfer autoencoder for intelligent diagnosis of bearing faults among different mechanical equipment [8].

In the research of source domain sample classification, this paper uses an improved algorithm based on the fast density peak search clustering algorithm. The density peak search clustering (DPS) [9] algorithm is a novel centroid clustering algorithm developed by Alex and Alessandro. Some scholars have also proposed the improved versions based on the idea of DPS and achieved good results. For example, to overcome the issue of high computational complexity, Xu and Jia [10] proposed a fast sparse search density peaks clustering (FSDPC) algorithm. In [11], a clustering approach based on DPS and a modified gravitational search algorithm (GSA) has been developed. D'Errico and Rodriguez [12] introduced an approach based on the optimized cluster center set selected by DPS to achieve the best clustering distribution. Cheng et al. [13] propose a dense member of local core-based density peaks clustering algorithm DLORE-DP, which is more effective, efficient, and robust than other algorithms.

For the purpose of fault diagnosis on small sample data sets, the adaptive transfer density peak search clustering (ATDPS) algorithm was proposed in this paper. The transfer learning was applied to the DPS algorithm to improve the classification accuracy and reduce the computational time consumption. The results of the source domain are transferred to the classification of data in the target domain with only a small amount of data according to the data distribution and the best classification pattern. The main contributions of the intelligent mechanical fault diagnosis algorithm in this paper are summarized as follows:(1)A feature weight analysis algorithm is proposed to improve the classification accuracy of the DPS algorithm combined with wavelet packet transform energy feature analysis.(2)An adaptive transfer density peak search clustering algorithm (ATDPS) is developed to adjust the parameters for the classification of sample data adaptively. The algorithm can select the optimal clustering classification results and transfer to the clustering analysis in the target domain.(3)An intelligent fault diagnosis model has been designed based on the proposed ATDPS and feature extraction algorithm.(4)The proposed intelligent fault diagnosis algorithm using transfer learning can not only classify but also accurately determine the fault status.

This work is organized as follows. DPS and the proposed ATDPS are both introduced in Section 2. Section 3 describes the WPT energy features extraction technique and intelligent fault diagnosis method. In Section 4, two benchmark case studies are carried out for the evaluation of the proposed method using experimental data sets, compared with affinity propagation clustering [14] and K-means [15]. Conclusions are drawn in Section 5.

2. Adaptive Transfer Density Peaks Clustering

2.1. The DPS Algorithm

The principle of the DPS algorithm is to obtain the best classification results and clustering centers based on the distribution of data points, relying only on the distance and density between the data points. The innovation of the algorithm lies in the analysis of the clustering centers: (a) Clustering centers are often surrounded by neighbors with lower local densities. (b) They are at a relatively large distance from any point with higher local density [16]. The calculation process of the algorithm is proposed as follows.

2.1.1. Calculating Local Density

Define the source domain data set with enough labeled samples: , where is the set of samples in the source domain, is the sample ’s label, the number of labels is , and the number of samples is . The density of point is calculated aswhere is the distance between point and and is the cutoff distance. is a function written bywhere is the local density of point and the value of is equal to the number of points that are closer than to point .

2.1.2. Calculate the Minimum Distance

According to the definition of clustering center, it is necessary to calculate a minimum distance , which is measured by the minimum distance between the data point and any other point whose local density is higher than that of point :

If point has the highest density among all data points, is defined as

2.1.3. Determine the Clustering Centers

It is difficult to determine the clustering centers only by density and distance . Besides, a solution is provided by the value of sorted in decreasing order. Each is given by

This algorithm still requires a manual setting of parameters to find the number of clustering centers. To build an intelligent fault diagnosis model for autonomous classification, a novel adaptive transfer density peaks search (ATDPS) algorithm is proposed in this paper.

2.2. The Feature Weight Analysis

The features obtained by extraction need to be reduced in dimensionality before they can be fed into the clustering algorithm. We proposed a feature self-weighting algorithm that is suitable for calculating the correlated weights for a large amount of data. The proposed algorithm only relies on the features themselves and achieves the weights. The steps of the proposed algorithm are as follows.(i)Set up a data set, assuming that the input contains kinds of fault categories, each category contains samples, the number of features is , set , and the feature set of the original data can be expressed as . A feature set of containing 4 states of data is shown in Figure 1.(ii)Similarization. To process data from different testbeds to improve portability, the feature set is similarized so that the numerical size and distribution of different data can be as close as possible.(iii)Calculate the self-similarity coefficient:

The self-similarity coefficient represents a measure of variability between the same features for different samples. The self-similarity matrix for each feature is obtained as

Thus, the self-similarity weight coefficient of each feature is

The weight size of each feature in the data set is obtained according to the self-weighting algorithm, as shown in Figure 2 for the feature data set given in Figure 1. The sensitive features with high weights are selected as classification transfer features, and these classification features form the similarity matrix of the sample set as the initial input for clustering.

2.3. The Proposed ATDPS Algorithm

The ATDPS algorithm is a major improvement on the DPS algorithm. It can adaptively adjust the value of in the iteration until the optimal clustering centers are found. The proposed ATDPS algorithm is described in detail below.

Determine with squared Euclidean distance:

The calculation of parameters and is changed in ATDPS. Initialize and , where is used to select top maximum entities and denotes round operator.

Calculate using (5), and is normalized with

If , the subscript should be selected as a centroid. Then, the remaining data points are divided into the nearest clustering centers to get the classification results. Here is a key parameter in the ATDPS algorithm. We will discuss the selection of in the following case studies.

In the experimental validation, it was found that the clustering effect of the algorithm can be difficult to determine as the number of iterations increases. A plot of the number of clusters with iterations for sample data with 4 classes is shown in Figure 3, where the clustering effect is shown in Figure 4.

In order to get the best and correct clustering results, a clustering effect evaluation index is established based on the clustering results:where is the average distance of the data point to other points in the same cluster as and is the minimum average distance from the data point to points in a different cluster of which is not a member. Update and , and calculate the average value of :

Finally, the best clustering result can be obtained based on the value of the maximum .

Transfer learning is used to solve the problem of classifying unlabelled target domain data and determining the categories in this paper. After the clustering analysis of the target domain, the coordinates and density of its clustering center and the source domain clustering centers are jointly analyzed to get the clustering result.

Suppose that the clustering center set of the source domain is , and the local density of clustering centers is . The clustering center set of the target domain is , and the local density of clustering centers is . Set up the distance parameter , density parameter , and comprehensive parameter ; they are calculated as follows:where and are normalized with

The comprehensive parameter is obtained as

The best clustering result of the target domain data can be obtained as long as the minimum is found. It is worth noting that, in the ATDPS algorithm, the target domain is clustered again as a new sample along with the source domain samples after the learning of the source domain data. Therefore, the target domain samples are classified into existing classes or form a new class. When the variety and number of source domain samples are sufficient, the categories of the small amount of unlabelled target domain data can be accurately classified.

3. The Intelligent Fault Diagnosis Algorithm of Machine

3.1. Wavelet Packet Transform

The wavelet packet transform (WPT) is a further extension and deepening of the wavelet transform with rigorous mathematical reasoning. The WPT inherits the characteristics of wavelet time-frequency localization while decomposing the high-frequency and low-frequency parts of the signal. The WPT divides the signal into multiple levels in the full-frequency band, maps the signal into these bands so that weak features are revealed, and provides a way to describe the signal more finely. For a vibration signal , its wavelet packet transform process and the corresponding decomposition bands are shown in Figure 5.

WPT has the two following functions: (1) WPT can accurately extract the fault signal characteristics. (2) WPT can remove the noise influence. The original signal and the signal after WPT satisfy the energy equivalence relationship as follows:where is the amplitude of node in layer after WPT. Therefore, the wavelet packet energy spectrum can be chosen as the sum of squares of signals in different frequency bands after its transform. The signal energy in different frequency band intervals is shown in the following equation:where is the length of the original signal.

3.2. The Proposed Intelligent Fault Diagnosis Approach

The proposed intelligent machine fault diagnosis approach practically involves four main stages, which are shown in Figure 6. First, the original signal is decomposed by WPT to extract all energy features. In the second stage, the self-weight analysis algorithm is used to downscale the energy features to obtain the sensitive feature set. In the third stage, the sensitive feature set is input into the clustering algorithm for unsupervised cluster analysis to establish the adaptive fault diagnosis model. Finally, the unsupervised fault diagnosis is performed on the target domain data to obtain the diagnosis results.

4. Experimental Verification

4.1. Bear Fault Diagnosis on MFS-MG Test Rig

To verify the advantages of the proposed potential energy features, this section uses the three groups' experiments to evaluate the ATDPS algorithm. In this case, vibration signals were acquired on the MFS-MG test rig shown in Figure 7. The sample set contains normal, outer ring fault, inner ring fault, and rolling element fault, and there are 50 sets of samples for each set.

The threshold value is the most important parameter mentioned in the ATDPS algorithm, and it should be primarily investigated. If the value of is too large, it will lower the discrimination of categories, and all data points would be classified as one cluster. However, too small value of will classify each data point as a cluster due to the fine distinction. In order to select an appropriate parameter , there are two investigations of different spans as illustrated in Figures 8(a) and 8(b). The range of is between and with step size of in the first evaluation, and in the second evaluation the value of varies from to with step size of . A higher accuracy rate is achieved when the value of is in the range of . Therefore, the value of is set to in our study.

In this experiment, 3 sets of experimental comparison experiments are used to verify the effectiveness of the algorithm. The classification methods are K-means clustering, AP clustering, and the proposed ATDPS. Firstly, the original feature set obtained by WPT feature extraction methods is analyzed for feature self-weighting. Then the sensitive feature set form after the self-weight analysis is put in the different clustering. The result of clustering is shown in Figure 9. The classification accuracy of the different algorithms is shown in Figure 10.

4.2. Bearing Fault Diagnosis Using CWRU Data Set

An intelligent fault diagnosis experiment is conducted on the data set from the Bearing Data Center of Case Western Reserve University (CWRU) [17]. The CWRU bearing vibration signals under three fault conditions were collected, that is, normal, outer ring fault, and inner ring fault. The original vibration signal is divided into 200 segments with 2048 samples.

In this section, K-means clustering, AP clustering, and the proposed ATDPS are used to classify the three-fault data. The accuracy of the clustering will illustrate the performance of the algorithms. After feature extraction, self-weight analysis, and (transfer) cluster analysis, the classification results of the three methods are shown in Figure 11. The classification accuracy of each class of faults is shown in Figure 12.

4.3. Discussion

The performance of the ATDPS algorithm in the two-bearing-fault data set is verified in different methods. The comparison of the fault classification of ATDPS, AP clustering, and K-means clustering is shown in Figure 13. From the comparison, it can be seen that ATDPS can achieve an accuracy of 98% in the unlabelled mixed MFS-MG bearings data differentiation and label the categories, which is higher than the other two algorithms. The algorithm also achieves 99.5% accuracy in the classification of unlabelled CWRU bearing fault data. In addition, compared with the other two algorithms, the ATDPS algorithm using transfer learning can not only classify but also accurately determine the fault status, while AP clustering and K-means clustering cannot achieve labeling the samples. The superior performance of ATDPS can make the algorithm play a greater role in practical mechanical fault diagnosis.

5. Conclusion

For the problem of fault diagnosis of unlabelled data in practical applications, this paper proposed an intelligent fault diagnosis algorithm for rotating machinery based on ATDPS. Transfer learning provided a new idea for the diagnosis of small sample unlabelled data. The accurate and intelligent determination of mechanical faults using nondeep learning algorithm is conducted in combination with the high-performance adaptive clustering algorithm. In this work, WPT is used to extract energy features, and an unsupervised intelligent fault diagnosis model is built autonomously from the source domain data through transfer learning and self-weighting analysis. The model can be used for classification of the unlabelled data in the target domain. Experimental results on different bearing test rigs demonstrated that the proposed algorithm can autonomously identify faults compared with K-means clustering and AP clustering. In particular, the capability of ATDPS in the diagnosis of unlabelled fault data is well suitable for practical industrial applications without prior knowledge.

Data Availability

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest regarding the publication of this paper.

Acknowledgments

The financial sponsorship from the projects of the National Natural Science Foundation of China (51875032) and Architecture under Grant X20159 is gratefully acknowledged.