Abstract

Aiming at the difficulty of extracting and classifying early bearing faults, a fault diagnosis method based on weighted average time-varying filtering empirical mode decomposition and improved eigenclass is proposed in this paper. Firstly, the bearing fault signal is decomposed into a series of intrinsic mode functions by the signal decomposition method, and the amplitude of the component is modulated by the weighted average method to enhance the fault impulse component. Then, the fractional Fourier transform is used to filter the reconstructed signal. Regarding classification issues, the eigenclass classifier is optimized by the IDE method that can be used for feature dimensionality reduction. Finally, the optimal features are selected and input into the IDE-EigenClass model. The experimental results show that the bearing fault diagnosis method proposed in this paper has higher accuracy and stability than the traditional PNN, SVM, BP, and other methods.

1. Introduction

In the process of bearing fault diagnosis, one difficult point is to extract the fault impulse component. Because the background noise is present in the process of equipment operation, the fault impulse component is likely to be drowned by the noise, resulting in the failure of diagnosis [13]. To solve this problem, Huang et al. proposed the empirical mode decomposition (EMD) for the first time, which can effectively reduce the noise in the signal. However, EMD has some problems such as end effect and mode aliasing [4]. Based on EMD, Huang et al. proposed the ensemble empirical mode decomposition (EEMD). By adding noise to the EMD decomposition process, the effect of mode aliasing has been significantly reduced [5]. In 2014, the variational mode decomposition (VMD) was proposed. VMD takes into full consideration the narrow-band property of the components, so the filtering frequency band is more centralized and the signal-to-noise ratio of the signal components is higher [6]. However, there remains the problem of mode aliasing. The time-varying filtering empirical mode decomposition (TVFEMD) is a new signal decomposition method proposed by Li et al. in 2017 [7]. The method provides an effective solution to the above problems by using B-spline approximation as a time-varying filter.

In addition, the noise can be removed by filtering the signal, as by the method of combining fast spectral kurtosis with VMD to successfully extract the features of early rolling bearing faults [8]. On the basis of the fast spectral kurtosis method, autocorrelation spectral kurtosis has been improved to effectively eliminate the interference of nonperiodic noise and improve the recognition accuracy [9]. In recent years, the fractional Fourier transform (FRFT) has been applied intensively in fault diagnosis. The fractional Fourier transform has good time-frequency focusing property for LFM signals [10]. In the early stage of the fault, since the fault impulse frequency changes steadily, the fractional Fourier transform proves effective in extracting the fault impulse components. Jiao and Ma used FRFT to extract the features of the bearing signal, and the effect was significant [11]. In addition, the advantages of FRFT can be exploited to the full under variable conditions. Jia and Huang [12] successfully diagnosed the fault of variable-speed bearing using FRFT and LMS noise reduction methods.

Machine learning algorithms are used mainly for classification issue. For example, Zhang et al. used the EMD-SVM combination method to classify faults [13]. In addition, Jung and Koh used the DWT-KNN to classify faults and achieved certain results [14]. However, there remain many problems in classification. For example, the accuracy is not high enough, and the operation is too time-consuming. Standardized variable distance (SVD) is an improved algorithm based on MDE and proposed in 2020 to solve the problem of long operation time [15]. However, its classification accuracy is heavily dependent on features. In the neural network algorithm, Ma et al. used the CSBP method to improve the fault recognition rate by optimizing the BP neural network [16]. In addition, the convolutional neural network (CNN) [17] has also been used to classify the bearing faults. Eigenclass is a new machine learning classification method proposed by Uğur in 2020. In this procedure, a feature matrix is constructed before the proximity based on the eigenvalues and is used to classify the samples [18]. This method ensures not only a high accuracy but also a high speed and efficiency when the sample size is not very large.

The above studies mainly focus on the vibration signal. Nevertheless, for some devices which are not easily embedded with the acceleration sensor, it would be a good choice to use the acoustic pressure sensor to collect the acoustic signal for analysis and processing. Its advantage of noncontact measurement makes it possible to diagnose the fault of some devices in which it would otherwise be difficult or even infeasible to embed the acceleration sensor [19]. However, there are few studies on fault diagnosis methods based on acoustic signals probably because of the relatively low signal-to-noise ratio of the acoustic signal. In addition, the generated noise still contains certain impulse noise, which is not necessarily periodic but can have a great effect on the kurtosis of the signal. Therefore, the impulse noise may lead to the wrong selection of sensitive components.

To solve the above problems, a fault diagnosis method based on weighted average time-varying filtering empirical mode decomposition and improved eigenclass is proposed in this paper. Firstly, the acoustic array sensor is used to collect data. Then, the signal is preprocessed to reduce noise interference. Finally, the features are input into IDE-Eigenclass model for fault classification. Compared with the existing bearing fault diagnosis methods, the innovative point of this proposal is that the acoustic array sensor is used to collect the bearing fault signal. This method aims to realize the multichannel acquisition, which can not only improve the accuracy of bearing sound signal but also realize the localization of the sound source of fault bearing. In the process of actual equipment operation, there are usually more than one bearing, usually multiple or even multiple groups of bearings. In such a scenario, failure to accurately locate whichever of the bearings will make it inconvenient to maintain the equipment. The acoustic array technology has made it possible to accurately locate the pitch angle and azimuth of the bearing as well as the faulty bearing. This paper is organized as follows. Section 2 mainly introduces the principle of signal preprocessing. Section 3 introduces the principle of classification for optimization of IDE-Eigenclass. Section 4 deals with the diagnosis process. Section 5 is concerned on the experimental verification. Finally, a summary with relevant discussion is made in Section 6.

2. Signal Preprocessing

2.1. Time-Varying Filtering Empirical Mode Decomposition

Time-varying filtering empirical mode decomposition (TVFEMD) is an improved method based on EMD. Because the cut-off frequency in those methods based on filtering is constant over time, they are not suitable for nonstationary signals. This method can solve the problem of mode mixing while maintaining the time-varying characteristics of different modes. The detailed introduction on TVFEMD has been covered in [20] and will not be repeated here.

2.2. Signal Reconstruction

A new signal reconstruction method is proposed in this paper.

Firstly, the kurtosis of each component is calculated. The kurtosis can reflect the degree of a signal’s transient state, thereby identifying the fault impulse characteristics. The expression is given as follows:

Next, the coefficient of correlation between each component and the original signal is calculated, with the coefficient of correlation between two signals defined aswhere and are the mean values of X and Y, respectively, and N is the size of the dataset.

The product of kurtosis and coefficient of correlation is termed as correlation kurtosis:

The mean correlation kurtosis is calculated aswhere represents the correlation-kurtosis value of each IMF component.

The signal is weighted and reconstructed by

The above method is the so-called weighted average time-varying filtering empirical mode decomposition.

This method is tantamount to modulating the amplitude at different frequency components of the signal. The kurtosis correlation coefficient index is proposed to modulate the amplitude of each component. By multiplying the index factor by each component, the effective component can be enhanced, while the noise component can be reduced in amplitude, thereby pinpointing the fault component. At the same time, this method retains all IMF components to ensure the intactness of useful information, thereby improving the accuracy of diagnosis.

2.3. Fractional Fourier Transform

The fractional Fourier transform is based on the Fourier transform. Its physical significance is shown in Figure 1.

As shown in Figure 1, a signal is transformed from the time domain into the frequency domain through the Fourier transform, a process equivalent to rotating the signal by 90°. In other words, the physical significance of the fractional Fourier transform is equivalent to the whole transformation process of the signal at 0° to 90°. The Fourier transform is applied to every internal angle to find the optimal rotary angle through two-dimensional peak search. The principle of selection for the optimal rotary angle is described as follows. The rotary angle is optimal when the projection of the desired fault impulse component on the optimal plane of rotation does not interfere with the projection of the noise. Since the projection of noise at any rotary angle is invariable, the only consideration is the projection of the fault impulse component.

3. Feature Optimization and Fault Classification

3.1. IDE Feature Optimization

In recent years, machine learning has been applied more and more frequently in the research on classification problems. Two key factors that affect the effectiveness of classification are the quality of feature selection and the classification method used. The foremost step of classification is to select features, whereby to use the corresponding features to represent a section of data. The quality and the number of features can have a great impact on the effectiveness of classification. In addition, the number of features should be moderate. Too few features are not enough to make a distinction between the types of samples, while too many features may lead to the disaster of dimensionality, resulting in a sharp decline in classification accuracy. Overemphasizing the distinction between features may lead to overtraining, which means training the wrong features in the original data as correct ones at certain probability. Anyway, either an extreme number of features or overemphasis on the distinction between features may lead to a sharp decline in classification accuracy.

The improved distance evaluation (IDE) technique is a feature optimization method. The ratio of the average distance between the same type of samples to the average distance between different types of samples is calculated to judge whether the feature can make a distinction between the types [21]. All features are solved and normalized according to the above method. The numerical results are sorted, and the optimal features are selected by setting a certain threshold. For example, a feature set contains class B samples, and its expression is given as follows:where represents the jth eigenvalue of the nth sample in the class B sample, represents the sample size of class B, J represents the eigenvalue, and B stands for the class. Therefore, the distinguishing degree of a feature to all classes can be calculated by the mathematical formula.

First, the average distance between the eigenvalues of the same samples is calculated as

Then, we calculate the average value :

Define the variance factor of as

Calculate the mean value of all samples under the same condition:

Calculate the average distance between different samples:

Define the variance factor as

A compensation coefficient is introduced as

The ratio between the two average distances is calculated and the compensation coefficient is allocated:

Finally, is normalized as

After normalization, the features are sorted in the size order, with the optimal features selected according to the threshold to form a feature matrix, which is then input into the machine learning model for training.

3.2. Eigenclass

According to the method, the feature matrix is first made up out of each set of features; next, the eigenvalues of the feature matrix in the training set and the test set are determined; finally, the distance between them is calculated.Step 1: divide the dataset A into the matrix of the training set and the test set .Step 2: if 0 is an element of the matrix, replace it with a very small positive number, such as 0.0001.Step 3: for each class r, divide the training set into the training matrix according to the number of classes.Step 4: each line element of the test set is composed of the feature matrix, so is each line element of each class in the training set. Then, the eigenvalues of the test matrix and all the training matrices are obtained byStep 5: calculate the quasi-distance between the training matrix and the test matrix by the following definition:Step 6: derive the quasi-distance between the test set and all the training sets of the same type, and then, construct a distance matrix and reorganize it in the ascending order.Step 7: calculate the r-mean value of each line element. The test set of the line with the minimum mean value belongs to the class represented by this line.

The steps of the eigenclass algorithm are shown in Table 1.

4. Procedure

According to the method proposed in this paper, a flowchart can be drawn as shown in Figure 2. The diagnostic steps in this paper are shown in Table 2.

5. Experimental Verification

5.1. Feature Extraction

To further verify the effectiveness of the method proposed in this paper, the NU1004 cylindrical roller bearing is used to test the outer ring fault, with bearing parameters shown in Table 3. The rolling bearing fault testbed is shown in Figure 3. The test platform consists of a rolling bearing, a DC motor, a support bench, a loading bolt, and other components. The speed of the bearing is adjusted by the motor, and the loading bolt is used to load the bearing. The test speed has reached 780 rpm. ReSpeaker Mic Array v2.0 is used as an acoustic sensor. The sampling frequency is 44.1 kHz. The device is shown in Figure 4. Placed at a horizontal distance of 20 cm from the bearing, the sensor can implement 4-channel acquisition. Two arrays of such sensors are placed in the axial and radial directions of the bearing, respectively, to implement a total of eight channels of acquisition.

According to the main parameters of the bearing, the theoretical fault characteristic frequency is 64.48 Hz at a test speed of 780 rpm. The collected sound signal is decomposed by TVFEMD to obtain 15 intrinsic mode functions. The time-domain waveform of the original signal and some IMF components are shown in Figure 5. The purpose of selecting components 6, 7, 12, and 13 is that the kurtosis correlation coefficient index of the four components exceeds the threshold value of 0.7 after normalization, an indication that these components are most sensitive. The kurtosis and the coefficient of correlation are calculated as shown in Figure 6.

Figure 5 indicates that IMF component 1 has a relatively large kurtosis value. By the kurtosis selection principle, it should contain the richest fault information. However, the envelope spectrum shows that it has a peak value near 100 Hz, which is quite discrepant from the theoretical fault characteristic frequency of 64.48 Hz. The discrepancy may be ascribed to the impulse noise contained in this component rather than in the fault impulse component. By calculating the coefficient of correlation, it is found that the correlation with the original signal is very low, and therefore, it can be concluded that IMF 1 is not an effective component.

By multiplying the kurtosis correlation coefficient index by the amplitude of each IMF, the amplitude of each component is modulated to increase the ratio of sensitive component to noise component, thus fulfilling the purpose of feature enhancement.

Then, the reconstructed signal is filtered by undergoing the fractional Fourier transform, and the result is obtained at 90° to 180°, where the optimal rotary angle is found, as shown in Figure 7, and the best order is 1.01.

At the order of 1.01, narrow-band filtering is carried out, with the results shown in Figure 8.

Then, the inverse fractional Fourier transform is performed, with the results shown in Figure 9.

Figure 9 indicates that the fault characteristic frequency is 64.66 Hz. Given its bandwidth of 1 Hz, its harmonics and other noise components are filtered out. Therefore, the envelope spectrum contains only the fault characteristic frequency. It follows that the proposed method proves effective.

5.2. Fault Classification

The fault data of normal bearing, inner ring bearing, outer ring bearing, and rolling element bearing have been collected. More details of the dataset is presented in Table 4.

After all data are denoised, 15 time-domain indexes and 3 entropy indexes are selected to represent a segment of data. The time-domain characteristics are shown in Table 5. Three entropy indexes, namely, sample entropy, permutation entropy, and fuzzy entropy, have also been selected.

Then, the IDE is used for feature optimization, and the result is shown in Figure 10.

The threshold is set to 0.6. As shown in Figure 10, the threshold values of features 12, 13, 15, and 17 are greater than 0.6, an indication that these four features can make a good distinction between the experimental data.

Then, 400 datasets are combined according to the ratio of 1 : 4 to form the training set and the test set. Finally, the eigenclass is used to classify the data. The results are shown in Figure 11.

Figure 11 indicates that the eigenclass has a satisfactory effect of classification on the test set, with only 3 label classification errors, including only one classification error for normal bearing, inner circle, and outer ring, respectively. The classification accuracy of 96.25% demonstrates that the eigenclass has a good effect on bearing fault classification. The effectiveness of this method is preliminarily verified.

In addition, this paper makes the comparison and verification in two aspects. Firstly, to verify whether the eigenclass has sufficiently accurate classification effect, the processed data are input into PS0-SVM, BP, PNN, and SVD. 10 training sessions are conducted, respectively, with the results of one testing shown in Figure 12. The comparison of testing accuracy for 10 times is shown in Figure 13.

As can be seen from Figure 12, the eigenclass achieves the highest testing accuracy within the least operation time. Therefore, it can be verified that eigenclass has a better classification effect than the traditional classification methods.

Secondly, to verify the effectiveness of the IDE method and the robustness of the eigenclass method, four features are randomly selected and input into the above five classification models, with the results shown in Table 6.

Table 4 indicates that the IDE method can improve the success rate of classification. In addition, no matter which four features are used, the eigenclass achieves the highest classification accuracy. Thus, the effectiveness of the proposed method is verified.

6. Conclusion

In this paper, a bearing fault diagnosis method has been proposed, based on weighted average time-varying filter modal decomposition and improved eigenclass. Through experimental verification, the following conclusions can be drawn:(1)The method of weighted reconstruction of decomposed signal proposed in this paper can effectively modulate the amplitude of signal, increase the ratio of the fault impulse component to the noise component, and improve the accuracy of classification.(2)By optimizing the input of the classifier, the classification accuracy of the eigenclass has been improved with effectiveness. Compared with traditional methods, this novel method ensures both the highest accuracy and stability, as well as the verified effectiveness.(3)In this paper, the eigenvalue attributes between indexes are used to construct the eigenvalue matrix. Concerning its originality, this method is innovative and open to further exploration and improvement in the follow-up research.

Data Availability

The data of this paper can only be obtained with the consent of the corresponding author.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

This work was supported by the National Key Research and Development Program of China (no. 2019YFB1705402) and National Natural Science Foundation of The People’s Republic of China (Grant 61973041).