Literature DB >> 35042894

Cross-condition and cross-platform remaining useful life estimation via adversarial-based domain adaptation.

Dongdong Zhao1, Feng Liu2.   

Abstract

Supervised machine learning is a traditionally remaining useful life (RUL) estimation tool, which requires a lot of prior knowledge. For the situation lacking labeled data, supervised methods are invalid for the issue of domain shift in data distribution. In this paper, a adversarial-based domain adaptation (ADA) architecture with convolution neural networks (CNN) for RUL estimation of bearings under different conditions and platforms, referred to as ADACNN, is proposed. Specifically, ADACNN is trained in source labeled data and fine-tunes to similar target unlabeled data via an adversarial training and parameters shared mechanism. Besides a feature extractor and source domain regressive predictor, ADACNN also includes a domain classifier that tries to guide feature extractor find some domain-invariant features, which differents with traditional methods and belongs to a unsupervised learning in target domain, which has potential application value and far-reaching significance in academia. In addition, according to different first predictive time (FPT) detection mechanisms, we also explores the impact of different FPT detection mechanisms on RUL estimation performance. Finally, according to extensive experiments, the results of RUL estimation of bearing in cross-condition and cross-platform prove that ADACNN architecture has satisfactory generalization performance and great practical value in industry.
© 2022. The Author(s).

Entities:  

Year:  2022        PMID: 35042894      PMCID: PMC8766616          DOI: 10.1038/s41598-021-03835-2

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Remaining useful life (RUL) estimation, which is one facet of prognostics and health management (PHM) aiming to provide users with an integrated view about the health state of a machine or an whole system[1], defined as “the length from the current time to the end of useful life”[2]. With the development of sensing technology, deep learning (DL) technology with nonlinear deep representation capabilities has gradually matured. A large amount of studies[3-9] on RUL estimation of bearing in same operating condition have obtained good results. Although every healthy mechanical component will has a rated working time after it is created, varying operating conditions and platforms will affect the service life of mechanical component to varying degrees. In the case where the training data and the test data come from different operating conditions and platforms, the generalization ability on the test data is usually poor. In the case of varying operating conditions and platforms, there are many factors influencing on performance of RUL estimation, such as feature extractor and regressive predictor, but in the final analysis, it is caused by the number of observation samples, first predictive time (FPT) detection method, etc. For example, when migrating from a source bearing entity with observation samples N1 to a target bearing entity with observation samples N2, if N1 is much smaller than N2, the RUL predicted by the target domain usually fluctuates more strongly. We assume that the bearing entity degrades linearly after the failure prediction time point. The source bearing entity 1 has N1 samples, target bearing entity 1 has N2 samples, and their FPT are respectively FPT1 and FPT2. It is a challenge for feature extractors and regressive predictors to accurately predict RUL. Particularly, the feature extractor needs to obtain discriminative features, and the regressive predictor needs to map the discriminative features to actual RUL value as accurately as possible. Though there are many effective domain adaptation (DA) methods for classification, but we practically found that few regressive DA methods. Jiang et al.[10] suggested that the regression space is usually continuous on the contrary, for example, there is no clear decision boundary. In fact, the observation samples have limited, the RUL estimation is still in a large discrete space. Jiang et al.[10] firstly proposed regressive DA for keypoint detection, which proved the effectiveness of DA in regressive prediction tasks. For RUL estimation of bearing under varying conditions or platforms, the key factor is how to assign cross-invariant features built by neural network corresponding same health status a equal estimated RUL, which is vital for unsupervised RUL estimation in cross-domain. Therefore, we found that in fact, the idea of regressive DA is still the same as the classification algorithm, such as SVM trying to find margins of boundaries between different classes. RUL estimation under varying operating conditions or varying platforms aims to find a boundary to distinguish the different operating states of bearing entity. In whole, there are three key points worth paying attention to: Motivated by domain adaptation neural networks (DANN)[11], in this paper, we introduce the original intention proposed by the DANN architecture into machinery RUL estimation. Through combining with the superiority of CNN in vibration signals process, adversarial-based domain adaptation (ADA), refers to ADACNN, consists of three parts: feature extractor, regressive predictor and domain classifier. Feature extractor firstly acquires domain-invariant and discriminative representation from raw vibration signals. The domain classifier, as an important auxiliary tool, forces feature extractor finding common space where samples have corresponding domain-invariant representation. The regressive predictor input a domain-invariant feature outputs a corresponding estimated RUL close to actual RUL value of the domain-invariant feature. Domain-invariant features. , F(x) is the feature representation of sample x, denotes the samples data, where RUL equals to 0.5, of the i-th source bearing entity. denotes the samples, where RUL equals 0.5, of the j-th target bearing entity. Discriminative features. is not equal to or far from , and is not equal to or far from . If the input of regressive predictor are some similar domain-invariant features respectively come from source domain and target domain, and the predicted results of regressive predictor should be approximately equal with each other. We use the formula to express as: , where Y() is the function of regressive predictor. Overall, in this paper, we proposes a novel neural networks framework with ADA for RUL estimation of bearings in different condition and platform. The main contributions of this paper are as follows: For scenarios lacking labels or no label, ADACNN can be simply switched for maximizing known labels’ value in real application scenarios. At the same time, it ensures excellent estimation accuracy in source domain and generalization ability in target domain. To our best knowledge, this framework with ADA is firstly introduced for RUL estimation of bearing in varying conditions and platform. In addition, in this paper, we just pay attention to a harder unsupervised case, that is, there are all unlabeled data in target domain. The ADACNN was verified on two public datasets: cross-condition experimental scenarios and cross-platform experimental scenarios on FEMTO and XJTU-SY dataset introduced afterthis. The proposed methodology compares with two non-adapted models respectively training with only source data and only target data to verify the generalization ability of proposed ADACNN.

Preliminaries

Transfer learning (TL) was grouped into three classifications: inductive TL, transductive TL and unsupervised TL[12]. Given a source domain and a corresponding learning task , a target domain and a corresponding learning task . In[12], DA belongs to transductive TL, that is, and . DA is divided into three categories in[13]: discrepancy-based methods, adversarial-based methods and reconstruction-based methods. Among them, DA methods are the most popular. Discrepancy-based methods use source and target domain data to fine-tune the model to reduce domain shift. Discrepancy-based DA can be divided into the following according to the criteria used: class criterion, statistic criterion, architecture criterion and geometric criterion. Adversarial-based methods use domain classifier to encourage domain confusion through an adversarial objective. Specifically, a domain classifier tries to guide feature extractor to find a common space with domain-variant from source or target domain, until the classifier cannot distinguish whether the feature comes from source or target, the prediction (classification or regression) function of the source domain can be shared with target domain data. Reconstruction-based methods utilize data reconstruction as an auxiliary task to ensure feature invariance between domains (source and target)[11,12]. Obviously, DANN[11] belongs to adversarial-based methods. As shown in the Fig. 1, the red part is a supervised source training process, the green part is unsupervised learning phase in target domain, and black part represents feature extractor which is always based on CNN or RNN or their variants. The feature extractor is parametrised by , prediction of main task (classification or regression that depends on main tasks) is parametrised by , the domain classifier is parametrised by . Of course, if the source domain is a supervised classification task in the adversarial model, then represents a classifier, and if it is a regression task, represents a regressive predictor. Source data with fully labeled and target data with no label are input into DANN. For the target data, it is an unsupervised learning process. The target data participate in the training of the to enhance the generalization ability of . The target data labels are only used when participating in the evaluation during the test phase. We record source data as and target data as , and which means the distribution of source domain is different from the distribution of target domain. The is used to try to find a common space where no discriminative information between source domain data and target domain data, instead of maintaining its own characteristics in different domains. The source domain features extracted by are fed to , then and are optimised by backpropragation. The target domain data features and source domain data features extracted by are fed to the , then and are also optimised by backpropragation. The original aspiration is to minimize the loss of predictor (classifier and regressor) and maximize the loss of the domain classifier. After some iterative training, cannot distinguish whether the feature comes from the source domain or the target domain, we think that DANN has found a common feature space between source domain data and target domain data. By parameters sharing mechanism, the trained DANN can fine-tune to target domain data.
Figure 1

The architecture of DANN.

The architecture of DANN. Therefore, from the perspective of theoretical realization, the composition of the loss function of DANN is divided into two parts: the prediction loss of the main task and the domain classification loss. The prediction loss of i-th batch examples is defined as the Eq. (1):where is the regularisation factor, and it’s weight is . Inspired by proxy distance, the optimisation problem of is denoted as the Eq. (2): Equations (1) and (2) are made up of a min–max adversarial optimisation procedure. DANN includes a deep feature extractor (black box in Fig. 1) and a deep label predictor (red box in Fig. 1), which together construct a traditional standard feedforward architecture. The domain classifier (green box in Fig. 1) is connected to the feature extractor. Last but not least, gradient reversal layer (GRL) plays an indispensable role in DANN, which builds a bridge between feature extractor and classifier for guiding feature extractor to acquire domain-invariant features. In the updation processes of model parameters by back-propagation, the gradient is multiplied by a certain negative constant through GRL. In short, the training processes of DANN are constraint through the min–max formula (Eqs. 1 and 2), and stopped training when the optimal balance train-off is found.

Proposed method

Problem formulation

In this paper, we proposed a framework with ADA for RUL estimation, which constitutes feature extractor, domain classifier, and a regressive predictor. It should be pointed out that we assume the source domain training are run-to-failure vibration data. Let , , represents the whole source sample data, is the number of the source bearing entity. denotes the i-th sample of the j-th bearing entity in source domain, where is the number of samples for the j-th bearing entity in source domain. By analogy, , , represents the whole target sample data, is the number of the target bearing entity. denotes the i-th sample of the j-th bearing entity in target domain, where is the number of samples for the j-th bearing entity in target domain. The methodology proposed in this paper mainly includes two steps as follows: Data preparation. Calculate RUL percentage label corresponding to the i-th sample of the j-th bearing entity in source domain. Building and training the ADACNN model. In addition to and in source domain as input of ADACNN, unlabeled data in target domain are also used as input when training ADACNN, which includes three parts shown in Fig. 2:
Figure 2

The architecture of ADACNN.

A regressive predictor parameterized by is introduced to accomplish main regression task through a supervised learning in source domain. A feature extractor parameterized by finds a common space with domain-variant feature from source and target domain data. A domain classifier parameterized by combined with GRL can not make the data output by the feature extractor be distinguished. The architecture of ADACNN.

Experimental design

Data preparation

The source domain data are divided into training data (, ) and test data (, ). The target domain data are also divided into training data and test data (, ). are used when training the ADACNN. are used as input for observing the maturity of model training. participates in unsupervised training ADACNN to improve its generalization ability, involves in the evaluation of ADACNN. It should be note that the training data are the run-to-failure data, and the test data are truncated data. The training label generation: taking FPT as the boundary, the actual RUL values of data samples before FPT point are equal to 1, and the actual RUL value corresponding to the j-th sample after FPT point drops from 1 to 0 successively, which denoted as , where represents the j-th sample after FPT point of i-th bearing entity, is equal to the number of whole run-to-failure samples of i-th bearing entity, and represents the number of samples before FPT point of i-th bearing entity. The test label generation: . represents the FPT of i-th bearing entity. We assume that known truncation points are always after FPT point, and K denotes the number of samples between truncation point and FPT point is K. Usually, before FPT point, it is not necessary or difficult to predict RUL because there is no obvious sign of degradation, and after this point, signs of degradation began to appear. Therefore, a FPT detection mechanism is very important to capture real-time changes. Kurtosis[3,14,15] was used to detect FPT. Many feature-fused methods combine time domain with frequency domain vibration characteristics. References[16,17] fused some features into one index descripting degradation process, and it is worth mentioning that these features are obtained by calculating the mahalanobis distance (MD) from original healthy state, which is a relative feature and suitable for some scenarios of vibration signal processing. In this paper, we will explore the impact of different FPT detection methods on RUL estimation.

Data normalization

In order to speed up the training speed and align the test data with the training data during the test, we have some pre-processing for the vibration data. Data normalization mainly includes four parts: and where norm(a, b) represents the normalization function and consists of two steps: firstly to calculate the mean m and variance d of b, and secondly to normalize a by m and d.

Building the ADACNN

Initialize feature extractor: The input parameters of feature extractor include input data, number of CNN layers, number of filters per layer, and dropout rate. The kernel size of first layer equals to 25 by default (Has been proven its effectiveness in[18]), and the remaining layers are initialized according to input parameters. The feature extractor mainly includes one dimensional convolution layer (Conv1D), activation layer, Dropout, and MaxPooling1D. The output of feature extractor are latent features, its dimension depends on the initialization parameters f. Initialize regressive predictor: The input parameters of regressive predictor include input data, number of convolution network layers, and number of nodes per layer. The architecture of regressive predictor mainly consist of FCN, activation layer, and dropout layer. Finally, the output predicted value is between 0 and 1, indicating RUL percentage. The closer the predicted value is to 1, the healthier it is, and the closer it is to 0, the closer it is to a fault. Initialize classifier: The classifier includes FCN, Activation layer, Dropout layer. Construct regression model: Source regression model consists of feature extractor and regressive predictor. The output latent features of the feature extractor is fed to the regressive predictor after passing through a flatten layer. Construct domain classification model: Through the parameter sharing mechanism of feature extractor, domain classification model mainly consists of feature extractor, GRL, and domain classifier.

Training the ADACNN

Initialization: Start iterative training with iteration . Patience value M = 0. In order to reduce the memory pressure, data are read in batches in each iteration. Training source regressive predictor and domain classifier: As shown in Fig. 2, through the forward propagation mechanism, the source regression model takes and as input and RUL prediction value as output, according to the known to calculate the prediction loss. The domain classifier model takes and as input, outputs binary classification and calculate the loss of domain classification, and then update the parameters of the regressive predictor and the feature extractor and classifier through the backward propagation method of gradient learning as shown in Fig. 2. The i-th updation formula are defined as Eq. (3): . Evaluate the ADACNN model by calculating the loss of RUL estimation: Calculate the accuracy of current model using and by root mean square error (RMSE) evaluation metrics. If is less best accuracy , then will be assigned to , otherwise, the patience value M is increased by 1. Judgement 1: If M is greater than the preset value, stop iterative training and save the model parameters of the i-th iteration. Judgement 2: If iteration i increases to threshold, stop iterative training and save the result of the i-th iteration. Start a new iteration: The entire experimental flow chart is shown in Fig. 3, the value of i plus one is re-assigned to i, and continue going to the step 2.
Figure 3

The schematic diagram of the process of the proposed model.

Testing the ADACNN: Use target test data to evaluate the accuracy of the trained model. The schematic diagram of the process of the proposed model.

Experimental setup

Dataset description

FEMTO dataset. The FEMTO dataset comes from an experimental platform called PRONOSTIA, where bearings’ degradation experiments are allowed to conduct in only few hours, this platform can obtain true bearing degradation data by accelerating bearing degradation under different operating conditions so that some data-driven techniques are studied further. PRONOSTIA includes three main parts: a rotating part, a degradation generation part, and a measurement part. For more rotating part and degradation generation, please refer to[19]. For the measurement part, there are two types of signals: temperature and vibration with horizontal and vertical respectively from their own acceleration sensor and temperature sensor. The algorithm proposed in this paper only uses vibration signals, and the sampling frequency of the acceleration sensor is 25.6 kHz. As tabulated in Table 1. FEMTO data set includes three different operating conditions, we use A to represent the FEMTO dataset, and Ai-j to represent the j-th bearing of i-th conditions in FEMTO. Six bearings data are run-to-failure data, which we use as training data. The 11 bearing data are truncated to predict the remaining life, which we use as test data.
Table 1

The information of FEMTO dataset.

Operating conditionsTraining dataTest data
A1 (1800 rpm and 4000 N)A1-1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} A1-2A1-3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} A1-7
A2 (1650 rpm and 4200 N)A2-1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} A2-2A2-3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} A2-7
A3 (1500 rpm and 5000 N)A3-1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} A3-2A3-3
The information of FEMTO dataset. XJTU-SY dataset. The XJTU-SY dataset was collected by the Xi’an Jiaotong university and the Changxing Sumyoung Techonology Company[20]. 32768 data points are collected on 1.28 s of every minute with sampling rate of 25.6 kHz. The tested processes of bearing are stopped when the amplitude of the vibration signal is higher than 20 g for protecting the test bed. There are two PCB 352C33 accelerometers are placed on the housing of the tested bearings, which are respectively on the vertical and horizontal axis. The information of XJTU-SY dataset are shown in Table 2.
Table 2

The information of XJTU-SY dataset.

Operating conditionsTraining dataTest data
B1 (2100 rpm and 12 kN)B1-1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} B1-2B1-3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} B1-5
B2 (2250 rpm and 11 kN)B2-1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} B2-2B2-3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sim $$\end{document} B2-5
The information of XJTU-SY dataset.

Data preprocessing

For FEMTO, 2560 data points are collected on 0.1 s of every 10 s with sampling frequency of 25.6 kHz. For XJTU-SY, 32768 data points are collected on 1.28 s of every minute with sampling rate of 25.6 kHz. In this study, we reduced the dimensions of the sample data of the two datasets to 1280 by way of dimensionality reduction. For FEMTO, we extract 0.05 s of data from every 0.1 s sample data. In other words, we reduce the dimensionality by half. For XJTU-SY, we extract 0.05 s of data from every 1.28 s of sample data. In other words, we reduce the dimensionality of sample data by 25.6 times.

Comparative methods

Comparison against methods with different FPT mechanisms. For FPT detection mechanism influences the performance on RUL estimation, we choose three FPT detection methods: MD, kurtosis and no FPT. MD and Kurtosis sensitive to early failure has been widely used in FPT detection. No FPT detection mechanism means that the bearing will degenerates from the initial state. We denotes these three methods as MD-ADACNN, Kur-ADACNN and NoFPT-ADACNN respectively. Comparison against non-adapted methods. In order to verify whether the proposed ADA method works, we use the following two baselines for comparison. The Source-Only method is trained on the source domain data and tested in the target domain . The Target-Only method is trained on the target domain data and tested on the target domain data (There is no intersection between training data and test data). To be fair, the parameters of feature extractor and regressive predictor of the Source-Only and Target-Only methods are consistent with the parameters of ADACNN.

Implementation details

Evaluation metrics

Root mean square error (RMSE), and Score are used as performance metrics to evaluate the error between the predicted RUL and the true RUL. RMSE has been used in many publications[21-23], and its definition formula is the Eq. (4)where and is the predicted RUL and actual RUL of i-th test sample, K denotes the total number of test samples. A larger RMSE value means a larger prediction error. Score, defined as Eq. (5), used in this study was first proposed in[19] and has been used in many studies[3,22,23]. The predicted RUL is greater than or less than the actual RUL should be treated differently. In other words, in the case of the same absolute value, the penalty for a positive value is less than the penalty for a negative value where K is the total number of test samples. , , and respectively represent actual RUL, predicted RUL and the difference between actual RUL and predicted RUL for the i-th testing data sample.

Hyper-parameter selection

In adaptation training processes, the learning rate of source domain regression and domain classification and the parameter of CNN largely determine the experimental performance. Therefore, we use the grid-search method to find the optimal learning rate (, , [Layer, units, Dropout]), and then manually fine-tune other parameters presented in Table 3. Overall, we did 6 cross-condition experiments (E1–E6) on the FEMTO dataset, 2 cross-condition experiments (E7–E8) on the XJTU-SY dataset, and 12 cross-platform experiments (E9–E20) on the FEMTO and XJTU-SY. Their parameter pairs are tabulated in Table 4.
Table 3

Hyperparameter evaluated in the proposed method.

Hyper-parameterRange
Layers: (\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{f}$$\end{document}θf, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{y}$$\end{document}θy, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{d}$$\end{document}θd)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{1, 2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
Units: (\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{f}$$\end{document}θf, f, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{y}$$\end{document}θy, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{d}$$\end{document}θd)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{16, 32, 64, 128, 256, 512\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
Learning rate: (\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{y}$$\end{document}θy (\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda _{y}$$\end{document}λy), \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\theta _{d}$$\end{document}θd (\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda _{d}$$\end{document}λd))\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{0.0001, 0.001, 0.01\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
Dropout rate\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{0.1, 0.3, 0.5, 0.7, 0.9\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha $$\end{document}α\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{0.8, 1.0, 2.0\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
Batch size\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{64, 128, 256\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
Threshold of patience M\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{20\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
Max iteration N\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{$$\end{document}{200\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\}$$\end{document}}
Table 4

Selected hyperparemeter for each source-target experiment pair.

No.From toCNN: Layer, (units), [Dropout]fSource regression: Layers, (units), [Dropout]Domain classification: Layers, (units), [Dropout]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha $$\end{document}αBatch size(\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda _{y}$$\end{document}λy, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda _{d}$$\end{document}λd)
E1A1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A22, (128, 64), 0.95121, (64), 0.12, (256, 128), 0.922560.01, 0.01
E2A1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A32, (128, 32), 0.5642, (32, 16), 0.32, (32, 16), 0.30.82560.0001, 0.0001
E3A2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A12, (64, 32), 0.1642, (32, 32), 0.12, (16, 16), 0.112560.001, 0.001
E4A2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A32, (128, 64), 0.15122, (64, 32), 0.12, (64,32), 0.122560.001,0.001
E5A3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A12, (64,32), 0.31282, (64, 64), 0.12, (64, 64), 0.122560.001, 0.001
E6A3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A22, (128, 32), 0.95122, (128, 128), 0.12, (128, 64), 0.122560.001, 0.001
E7B1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B22, (128, 32), 0.1642, (256, 128), 0.12, (128, 64), 0.50.81280.001, 0.001
E8B2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B12, (128, 32), 0.9642, (256, 128), 0.12, (128, 64), 0.90.81280.001, 0.001
E9A1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B12, (32, 32), 0.1641, (64), 0.11, (64), 0.50.8640.001, 0.001
E10A1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B22, (128, 32), 0.1641, (64), 0.12, (64, 64), 0.50.8640.001, 0.001
E11A2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B12, (32, 32), 0.1641, (64), 0.11, (64), 0.50.8640.001, 0.001
E12A2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B22, (32, 32), 0.1641, (64), 0.11, (64), 0.50.8640.001, 0.001
E13A3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B12, (32, 32), 0.1641, (64), 0.11, (64), 0.50.8640.001, 0.001
E14A3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B22, (32, 32), 0.1641, (64), 0.11, (64), 0.50.8640.001, 0.001
E15B1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A12, (32,32), 0.9641, (64), 0.92, (64,64), 0.50.8640.001,0.001
E16B1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A22, (128, 32), 0.9641, (64), 0.92, (64, 64), 0.50.8640.001, 0.001
E17B1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A32, (128, 32), 0.9641, (64), 0.92, (64, 64), 0.50.8640.001, 0.001
E18B2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A12, (64,64), 0.1641, (32), 0.92, (64,64), 0.51640.001,0.001
E19B2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A22, (128, 32), 0.9641, (64), 0.92, (64, 64), 0.52640.001, 0.001
E20B2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A32, (128, 32), 0.9641, (64), 0.12, (64, 64), 0.50.8640.001, 0.001
Hyperparameter evaluated in the proposed method. Selected hyperparemeter for each source-target experiment pair.

Discussion

It should be pointed out that all the pictures in the following content are generated by MATLAB software based on experimental data.

Cross-condition

In Fig. 4, the horizontal axis represents every test units of same operating condition in target domain (5, 5 and 1 on FEMTO, 3 and 3 on XJTU-SY), and the vertical axis represents RUL percentage. The thick histogram and the thin histogram respectively represent the predicted value and label of the same method. The closer the highest points of the thick histogram and the thin histogram are, the higher the accuracy. As described in Fig. 4, it can be seen that no matter which data set it is verified on, the predicted value of MD-ADACNN method is closer to its actual label. In addition, the MD-ADACNN method usually gives a predicted value slightly smaller than the true value, which will provide constructive warnings for engineering operation and maintenance engineers. However, for the other two FPT detection mechanisms, the prediction accuracy of the Kur-ADACNN method is obviously the lowest. The kurtosis-based FPT detection mechanism is an indicator in the vibration data, and the MD-based FPT detection mechanism is a relatively joint indicator of multiple indicators in the vibration data after dimensionality reduction in a relatively healthy state. Therefore, it can be seen from the results that the MD-based method is a more suitable FPT detection mechanism which is closer to the bearing degradation trend. It is worth emphasizing that the impact of FPT is only effective for experimental data sets, and does not mean that a specific FPT mechanism on all data sets can always maintain the best performance. Of course, the experimental work done can provide a certain degree of reference for the RUL prediction research of bearings from specific operating conditions or platforms to similarly configured operating conditions or platforms (from E1 to E20).
Figure 4

RUL estimation comparisons with different FPT detection mechanisms on FEMTO and XJTU-SY datasets: (a) A3  A1, (b) A3  A2, (c) B1  B2, (d) B2  B1.

RUL estimation comparisons with different FPT detection mechanisms on FEMTO and XJTU-SY datasets: (a) A3  A1, (b) A3  A2, (c) B1  B2, (d) B2  B1. From Fig. 5, we find that the RUL estimation results of MD-ADACNN method are between the results of Source-Only and Target-Only methods, and even more closer to the actual RUL value than Target-Only method (Fig. 5b), which proves effectiveness of ADA.
Figure 5

RUL estimation comparisons with source-only and target-only methods on FEMTO and XJTU-SY datasets. (a) E5: A3 A1, (b) E6: A3  A2, (c) E7: B1 B2, (d) E8: B2  B1.

RUL estimation comparisons with source-only and target-only methods on FEMTO and XJTU-SY datasets. (a) E5: A3 A1, (b) E6: A3  A2, (c) E7: B1 B2, (d) E8: B2  B1. From the perspective of the FPT detection mechanisms, as shown in the cross-condition results of the three methods (MD-ADACNN, NoFPT-ADACNN and Kur-ADACNN) on the FEMTO dataset listed from E1 to E6 in Table 5, MD-ADACNN has obtained the four best RMSE and Score accuracy (E1: A1 A2, E3: A2  A1, E5: A3  A1 and E6: A3  A2). In the other two cross-condition (A1  A3 and A2  A3), although MD-ADACNN has a larger error than the other two methods, from the 6 cross-condition experiments as a whole, MD-ADACNN predicts RUL more stable, and the two methods NoFPT-ADACNN and Kur-ADACNN, especially Kur-ADACNN, are greatly affected by the total number of cycle of units under varying conditions. In the FEMTO data set, the cycle of units under condition A3 is shorter, and the cycle of units under condition A1 and A2 is longer. Therefore, these two methods have better results in RUL estimation of cross-condition bearings from longer to shorter cycles, but very poor in the opposite case.
Table 5

RMSE/score ± standard deviation comparison between source-only, target-only, NoFPT-ADACNN, Kur-ADACNN and MD-ADACNN on FEMTO dataset and XJTU-SY dataset.

No.Source-onlyTarget-onlyMD-ADACNNNoFPT-ADACNNKur-ADACNN
RMSEScoreRMSEScoreRMSEScoreRMSEScoreRMSEScore
E153.2101.332.664.636.2 ± 1.674.0 ± 46.441.5 ± 2.791.9 ± 66.736.5 ± 2.588.3 ± 88.2
E226.012.58.71.08.5 ± 14.45.1 ± 14.03.4 ± 4.00.3 ± 0.42.9 ± 0.50.3 ± 0.1
E339.4239.822.515.326.8 ± 5.914.4 ± 10.944.2 ± 7.4328.0 ± 700.064.3 ± 8.98542.5 ± 9330.2
E430.520.224.610.719.8 ± 18.810.3 ± 4.59.9 ± 9.11.9 ± 2.83.8 ± 1.30.4 ± 0.1
E567.02084.733.5930.431.4 ± 3.27.1 ± 4.740.0 ± 7.4136.6 ± 215.952.0 ± 4.0544.4 ± 322.3
E648.81752.630.735.935.4 ± 0.837.2 ± 9.143.6 ± 14.61088.6 ± 2835.637.3 ±  2.089.4 ± 43.4
E718.310.113.01.013.3 ± 0.31.2 ± 0.614.0 ± 2.92.1 ± 1.117.6 ± 3.49.6 ± 1.0
E85.60.73.10.33.6 ± 0.70.4 ± 0.45.3 ± 2.20.6 ± 0.34.1 ± 0.80.7 ± 0.1
RMSE/score ± standard deviation comparison between source-only, target-only, NoFPT-ADACNN, Kur-ADACNN and MD-ADACNN on FEMTO dataset and XJTU-SY dataset. From the perspective of whether to use domain adaption technology in Table 5, in most experiment with cross-condition (E1: A1  A2, E2: A1  A3, E3: A2  A1, E5: A3  A1 and E6: A3  A2), RMSE(Source-Only) > RMSE(MD-ADACNN) > RMSE(Target-Only). In these cross-condition (E2: A1  A3 and E4: A2  A3), RMSE(MD-ADACNN) RMSE(Target-Only), which shows that the generalization ability of ADA technology from condition A2 with a long cycle time to condition A3 with a short cycle time exceeds supervised algorithms that only use target data, that is, to some extent, the data in target domain are benefit to guide the whole ADA model fine-tune to target data. Judging from the results of the five methods (Source-Only, Target-Only, MD-ADACNN, NoFPT-ADACNN and Kur-ADACNN in Table 5) verified on the FEMTO data set, basically, we found the effects of these three methods (MD-ADACNN, NoFPT-ADACNN and Kur-ADACNN) are superior to that of Source-Only, so these results prove that the validity of domain adaption technology in the research field of cross-condition bearing RUL estimation. Observing the results listed from E7 to E8 in Table 5. on the XJTU-SY data set from the same perspective, the same conclusion was confirmed again.

Cross-platform

In order to further verify the performance of the proposed ADACNN method in the inter-platform domain for RUL estimation, we choose the three conditions in the FEMTO dataset as the source domain or target domain data, and at the same time, 2 conditions in the XJTU-SY dataset are used as target domain or source domain data, so there are a total of 12 experiments (E9–E20 tabulated in Table 6 between two platforms. Different from the cross-condition, the cross-platform experiment part only explores the research on the remaining life of different FPT establishment mechanisms under the same DA technology, because the superiority of the DA technology has been clearly proved in the previous experiments.
Table 6

RMSE/score ± standard deviation comparison under cross-platform on FEMTO and XJTU-SY.

No.From toMD-ADACNNNoFPT-ADACNNsKur-ADACNN
RMSEScoreRMSEScoreRMSEScore
E9A1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B16.1 ± 2.60.9 ± 0.56.6 ± 2.10.8 ± 0.42.3 ± 2.00.2 ± 0.1
E10A1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B260.9 ± 4.64027.6 ± 1904.854.1 ± 3.42531.6 ± 1341.165.2 ± 3.922933.6 ± 10923.1
E11A2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B11.7 ± 1.60.1 ± 0.24.5 ± 3.70.4 ± 0.54.6 ± 3.50.7 ± 0.9
E12A2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B28.5 ± 12.710.4 ± 30.118.7 ± 15.418.3 ± 29.363.1 ± 15.597874.2 ± 125079.4
E13A3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B14.4 ± 2.10.5 ± 0.36.9 ± 2.60.9 ± 0.64.4 ± 5.20.5 ± 0.7
E14A3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} B213.7 ± 9.07.8 ± 11.623.9 ± 10.371.1 ± 141.444.2 ± 6.4117.3 ± 85.6
E15B1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A132.7 ± 5.159.3 ± 86.941.5 ± 8.481.6 ± 70.745.8 ± 13.3101.3 ± 86.4
E16B1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A232.6 ± 1.726.5 ± 7.836.3 ± 3.344.8 ± 23.036.2 ± 2.064.1 ± 21.5
E17B1 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A321.6 ± 6.810.2 ± 8.67.7 ± 14.03.0 ± 6.25.2 ± 1.80.7 ± 0.2
E18B2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A130.2 ± 5.825.1 ± 16.031.9 ± 7.114.9 ± 6.034.8 ± 10.039.6 ± 18.9
E19B2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A235.5 ± 1.943.4 ± 13.435.8 ± 3.2122.1 ± 179.747.0 ± 9.01510.6 ± 2003.2
E20B2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\xrightarrow []{}$$\end{document} A31.5 ± 2.60.1 ± 0.310.7 ± 14.94.4 ± 8.06.1 ± 0.50.8 ± 0.1
RMSE/score ± standard deviation comparison under cross-platform on FEMTO and XJTU-SY. It can be seen from Table 6 that compared with NoFPT-ADACNN and Kur-ADACNN, when the evaluation metric is RMSE, MD-ADACNN obtained the optimal value in 9 out of 12 cross-platform experiments (E11: A2  B1, E12: A2 B2, E13: A3  B1, E14: A3 B2, E15: B1  A1, E16: B1 A2, E18: B2  A1, E19: B2 A2, and E20: B2  A3). When the evaluation metric is Score, MD-ADACNN obtained the optimal value in 8 of the 12 cross-platform experiments (E3: A2B1, E12: A2  B2, E13: A3  B1, E14: A3 B2, E15: B1  A1, E16: B1 A2, E19: B2  A2, and E20: B2 A3). We still find that the Kurtosis-based FPT detection method is unstable. It can be seen from Fig. 6 that in Fig. 6b,d, the Kur-based FPT detection mechanism considers the RUL label at the prediction point to be 1, while RUL label of the prediction point determined by the No FPT and MD-based FPT detection mechanism are different not big. From the principle that the majority obeys the minority, for test unit = 1 in Fig.6b and test unit = 4 in Fig.6d, we can think that Kur-based does not perform well. On the whole, whether it is A B or BA, the MD-ADACNN predicted value is not far from the corresponding label, and even more often slightly smaller than the corresponding label value.
Figure 6

RUL estimation comparisons with different FPT detection mechanisms between two platforms: (a) E11: A2  B1, (b) E12: A2  B2, (c) E16: B1  A2, (d) E19: B2  A2.

RUL estimation comparisons with different FPT detection mechanisms between two platforms: (a) E11: A2  B1, (b) E12: A2  B2, (c) E16: B1  A2, (d) E19: B2  A2.

Feature visualization

In order to demonstrate the effectiveness of the proposed model, we use the t-SNE method to visualize high-level features. Fig. 7 is the feature visualization in the case of knowledge transfer between different conditions under the same platform, that is, in Fig. 7a, the source domain test entity is based on condition 1 in FEMTO dataset, and the target test entity is based on condition 2 in FEMTO dataset, and in Fig. 7b, the source domain test entity is based on condition 2 in FEMTO dataset, and the target test entity is based on condition 1 in FEMTO dataset. From Fig. 7 we can see that the high-level features on the subspace corresponding to the source domain test entity data and the target domain test entity data are fully fused, which proves that the proposed domain adaptation method has played a role.
Figure 7

The feature visualization for cross-condition.

The feature visualization for cross-condition. Figure 8 describes feature visualization for cross-platform. In Fig. 8a, the source data is condition 3 based on platform a, and the target data is condition 2 based on platform b. In Fig. 8b, the source data is condition 2 based on platform a, and the target data is condition 3 based on platform b. In the above two cases, it has big bridge between source test data and target test data. Because this research background is cross-platform and cross-condition, and the source domain condition is single, the training data information is not rich, so for the remaining life prediction task, the source domain knowledge is transferred to the target domain is quite difficult. Despite this, we can see from Fig. 8 that both the high-level features corresponding to source domain data and target domain data have a large overlap area to some extent.
Figure 8

The feature visualization for cross-platform.

The feature visualization for cross-platform. Due to the difficulty in predicting the remaining life of cross-platform migration, such as different operating conditions, different sampling frequencies, and large differences in life span. In short, compared with Fig. 7, the monotonicity of Fig. 8 is not so obvious, but this does not mean that it is a complete failure. We can understand that such feature representation is a greater challenge for the subsequent regression predictor.
  1 in total

1.  Bearing Fault Diagnosis Based on the Switchable Normalization SSGAN with 1-D Representation of Vibration Signals as Input.

Authors:  Dongdong Zhao; Feng Liu; He Meng
Journal:  Sensors (Basel)       Date:  2019-04-29       Impact factor: 3.576

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.