Next Article in Journal
Erratum: Toro, C. et al. On the Detection of Spectral Emissions of Iron Oxides in Combustion Experiments of Pyrite Concentrates. Sensors 2020, 20, 1284
Previous Article in Journal
Intelligent UAV Deployment for a Disaster-Resilient Wireless Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Ensemble Adaptive Sparse Bayesian Transfer Learning Machine for Nonlinear Large-Scale Process Monitoring

1
School of Automation Science and Engineering, South China University of Technology, Guangzhou 510640, China
2
Centre for Technology in Water and Wastewater, School of Civil and Environmental Engineering, University of Technology Sydney, Ultimo, NSW 2007, Australia
3
School of Data Science and Information Engineering, Guizhou Minzu University, Guiyang 550025, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(21), 6139; https://doi.org/10.3390/s20216139
Submission received: 7 October 2020 / Revised: 26 October 2020 / Accepted: 26 October 2020 / Published: 28 October 2020
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Process monitoring plays an important role in ensuring the safety and stable operation of equipment in a large-scale process. This paper proposes a novel data-driven process monitoring framework, termed the ensemble adaptive sparse Bayesian transfer learning machine (EAdspB-TLM), for nonlinear fault diagnosis. The proposed framework has the following advantages: Firstly, the probabilistic relevance vector machine (PrRVM) under Bayesian framework is re-derived so that it can be used to forecast the plant operating conditions. Secondly, we extend the PrRVM method and assimilate transfer learning into the sparse Bayesian learning framework to provide it with the transferring ability. Thirdly, the source domain (SD) data are re-enabled to alleviate the issue of insufficient training data. Finally, the proposed EAdspB-TLM framework was effectively applied to monitor a real wastewater treatment process (WWTP) and a Tennessee Eastman chemical process (TECP). The results further demonstrate that the proposed method is feasible.

1. Introduction

Due to the increasing diversification of industrial demand, the combination of process and equipment results in system structures become increasingly complex. Therefore, if the operation status of a plant cannot be monitored comprehensively and efficiently, it will not only cause serious economic losses [1], but also may cause irreversible damage to social communities. Timely detection and prediction of faults has become a focus of attention in academia and industry [2,3,4]. Recently, data-driven process monitoring has developed as the best form of “whistleblower” for extreme or abnormal events in a plant. This is because the data-driven process monitoring method does not need to establish an accurate mechanism model; rather, it uses a data-driven model to establish a global monitoring method for complex large-scale industrial processes [5]. Moreover, data-driven monitoring methods have been successfully applied in many different scenarios [1,6,7,8,9,10,11]. Liu et al., proposed a variational Bayesian principal component analysis (PCA) model to effectively monitor a wastewater treatment process (WWTP) [6]. Ge et al., proposed a two-step information extraction strategy to monitor a Tennessee Eastman chemical process (TECP) [7]. Zhu et al., proposed a novel two-step strategy probabilistic independent component analysis- probabilistic PCA (PICA-PPCA) to improve the robustness of the traditional method [8]. However, the above-mentioned data-driven methods ignore some general characteristics. Firstly, due to the complexity of external disturbances, the collected datasets contain nonlinear information [12]. Therefore, the performance of the above-mentioned linear methods usually degrades. Secondly, due to the increasing complexity of industrial equipment, monitoring models require an increasing amount of real-time training data. When sufficient training data are not available to train a reliable model, the monitoring model produces false and missed alarms, resulting in unpredictable losses to the factory.
To address the nonlinear characteristics of industrial data, the academia and industrial communities have undertaken a significant amount of research [1,13,14]. Ma et al., proposed using a deep convolutional neural network to diagnose the faults of rotating machinery [1]. However, the neural network-based method has some disadvantages, such as high computational cost and poor interpretability, especially when the network layer increases, which leads to an increase in the required tuning parameters. Based on the above considerations, the kernel function method, as a powerful technology, has been effectively used to expand the traditional statistical monitoring model. Lee et al., used a kernel PCA (KPCA)-based method to monitor a nonlinear wastewater treatment system, and experiment results show that KPCA performed better than PCA [13]. Wang et al., proposed using kernel independent component analysis (KICA) to diagnose nonlinear process faults. The experimental results show that KICA is superior to ICA [14]. In addition, the kernel-based support vector machine (SVM) has become one of the most popular method [15]. Liu et al., presented an in-depth discussion on the application of kernel-based methods in industry [16].
However, the number of support vectors will increase rapidly as the size of training sets increases [17], thus unnecessarily increasing the computational burden. To overcome this problem, the computational burden can be reduced by increasing the sparsity of the model. Sparsity is an important and desirable property for algorithm design and model construction. Firstly, sparsity is able to control the complexity of the model and avoids over-fitting. Secondly, prediction using a sparse model is highly effective in computation. Inspired by [17], the RVM under the sparse Bayesian framework is selected accordingly. In recent years, RVM has gained more attention. Liu et al., used an RVM to predict the difficult-to-measure variables of a WWTP [18]. Wu et al., proposed a multi-kernel RVM to predict the quality-related faults of a WWTP [19]. Hu et al., developed an RVM to predict the remaining useful life (RUL) of field pump impellers [9]. Because the posterior distributions of many of the weights are sharply peaked around zero, the sparsity of the RVM can easily use automatic relevance determination (ARD) to remove the zero-weight “relevance” vectors [20]. Furthermore, the sparsity of the RVM can meet the needs of real-time monitoring of factories. Therefore, based on the sigmoid function, Bernoulli distribution, Bayesian derivation, and Markov chain rule, we re-derived the probability RVM, allowing the probabilistic relevance vector machine (PrRVM) to monitor WWTPs and the TE chemical process.
The re-derived PrRVM is still limited by the following factors. When the amount of training data is insufficient, the performance of the PrRVM is weakened. Based on previous research [21,22], one solution lies in augmentations of the available data. The training dataset can be artificially expanded through transformations of samples, such as adding additional noise to the raw data. Another approach is to create synthetic data to assist in model training [21]. However, these methods rely too heavily on the original data, and expansion of the dataset is subject to significant uncertainty, which can be counterproductive. To address the above problems, transfer learning is considered to be embedded in PrRVM. Transfer learning aims to transfer the learned knowledge from one domain (source domain) to another domain (target domain). Therefore, the insufficient training data problem can be potentially addressed by knowledge transformation from additional datasets with sufficient supervised information. Transfer learning can be divided into three categories [23]: instance-based transfer learning, feature-based transfer learning, and model-based transfer learning. Based on [24], we used adaptive boosting technology and instance-based transfer learning to update the weight vector of source domain (SD) data and labeled target domain (LTD) data. If SD data can improve the method performance, its weight is increased. For LTD data, when it is misclassified, its weight will be increased to ensure more attention is paid to its optimization in the next iteration. In each iteration period, the updated data will be used to train a novel PrRVM detection model.
Note that the data collected by the process industries (WWTP and TECP) are not designed for transfer learning. Therefore, the dataset must be split before executing the corresponding strategy. Firstly, the real-time collected TD dataset is split into two components: the first component is the labeled target domain (LTD) dataset, which is defined as the training dataset. The second component is the unlabeled target domain (ULTD) dataset, which is defined as the real-time testing dataset. The SD dataset is the auxiliary training dataset, which is the out-of-date dataset. Then, the SD dataset and LTD dataset are updated by adaptive boosting technology and transfer learning. To summarize, we propose a modified version of a PrRVM for fault diagnosis that can enable a high quantitative fault diagnosis performance in the design process. Additionally, transfer learning is embedded in the PrRVM to solve the problem of insufficient training data. The ensemble monitoring model constructed using two-layer iteration (weight iteration and hyperparameter iteration) with the ensemble rule is termed the ensemble adaptive sparse Bayesian transfer learning machine (EAdspB-TLM). Finally, key performance indicators (KPIs) are used to evaluate the performance of different methods.
The paper is organized as follows: Section 2 presents the basic theory of the approach. Section 3 provides a detailed formula derivation of the EAdspB-TLM. In Section 4, the EAdspB-TLM is used to monitor different types of faults, and the experiment results are discussed and analyzed. Finally, the paper ends with conclusions in Section 5.

2. Theoretical Foundation

2.1. Transfer Learning

The purpose of transfer learning is to gain knowledge from an environment (source domain) to help the learning task in a new environment (target domain) [23]. To facilitate the subsequent use of transfer learning algorithms, the general symbols related to transfer learning are defined as follows:
(1)
Detection model Φ : X Y , where X represents the training data or testing data. Y represents the corresponding sample label. In this study, the premise is to assume that the training data are not sufficient to train a reliable detection model Φ .
(2)
Domain ( D ): The symbol of domain is represented by D = { X , P ( X ) } , where X = { x 1 x n } X , X is a feature space. D s = { X s , P ( X s ) } is the source domain (SD). D t = { X t , P ( X t ) } = D t 1 D t 2 is the target domain (TD). D t 1 and D t 2 are the LTD and ULTD, respectively. In this paper, LTD data are used as the training data; ULTD data are used as the testing data.
(3)
Task ( T ) : T = { Y , f ( ) } , Y { 0 , 1 } is the sample label. f ( ) is the corresponding prediction function, f ( X ) = P ( Y | X ) . Its task is to minimize the deviation between the predicted label and the real label Y .

2.2. Sparse Bayesian for Fault Diagnosis

The essence of data-driven fault diagnosis is to identify the running state of the equipment. The corresponding labels can be set for different running states; for example, the data label of the normal working condition is set to 0, and the data label of the fault state is set to 1. Then the fault diagnosis model in the framework of a sparse Bayesian (PrRVM) is equivalent to a supervised classifier. In this study, the PrRVM is a sparse model with probabilistic output. Suppose the training dataset is { x i , y i } i = 1 n , where x i . is the put data, then y i { 0 , 1 } . is the corresponding label. The prediction formula of PrRVM can be expressed as follows:
y j = Φ ( x j ; w ) = i = 0 n w i f i ( x j ) + ε
where w = [ w 0 , w 1 , , w n ] is the weight vector. ε represents the additive noise, let ε ~ N ( 0 , σ 2 ) . f i ( x j ) = k ( x j , x i ) is kernel function, which aims to map low dimensional non-separable data to high dimensional space. When the weight vector w and variance σ 2 are known, the label vector y = [ y 1 , w 2 , , w n ] T can be derived using the following probability expression:
p ( y | w , σ 2 ) = i = 1 n N ( y i | u , σ 2 ) = ( 2 π σ 2 ) n / 2 exp ( y Ψ w 2 2 σ 2 )
according to [17], Ψ is the n × ( n + 1 ) “design” matrix, where
Ψ = [ 1 k ( x 1 , x 1 ) 1 k ( x 2 , x 1 ) k ( x 1 , x n ) k ( x 2 , x n ) 1 k ( x n , x 1 ) k ( x n , x n ) ]
w and σ 2 can be estimated by expectation maximization, but it is subject to over-fitting [17]. To avoid over-fitting, a common approach is to impose some additional constraints on the parameters. We use Bayesian strategy and define an explicit prior probability distribution on the parameters to “constrain” the parameters. Assuming that the zero-mean Gaussian prior distribution on the weight vector w can be expressed as follows:
p ( w | a ) = i = 0 n N ( w i | 0 , a i 1 )
a = [ a 0 , a 1 , , a n ] is the hyperparameter vector. w and σ 2 can be further solved by Bayesian inference and rules. Here, we first assume that w and σ 2 are known, and then derive the solution formula of the classification problem. Firstly, the logistic sigmoid function σ ( z ) = ( 1 + e z ) 1 is introduced. Assuming that the data obey the Bernoulli distribution, the corresponding likelihood function can be expressed as follows:
p ( y | w , σ 2 ) = i = 1 n σ ( Φ ( x i ; w ) ) y i [ 1 σ ( Φ ( x i ; w ) ) ] 1 y i
When the predicted value y = 1 , it indicates that the system is out of control.

3. Ensemble Adaptive Sparse Bayesian Transfer Learning Machine for Process Monitoring

3.1. Adaptive Boosting Technology and Transfer Learning

The proposed process monitoring framework is shown in Figure 1. The adaptive sparse Bayesian transfer learning machine is mainly composed of two components. The first component is the adaptive boosting technology in the transfer learning framework, and the second component is the PrRVM fault diagnosis model in the Bayesian framework. The first part was proposed by Dai et al. [24], and named the TrAdaBoost algorithm. In this paper, TrAdaBoost is used to assign the data weights. Before the algorithm is implemented, suppose that the following symbols represent the divided SD data and label: Data: X s R p s × n s , label Y s 1 R 1 × n s . LTD data: X t 1 R p t × n t 1 , and the corresponding label Y t 1 R 1 × n t 1 . ULTD data: X t 2 R p t × n t 2 . n s and p s represent the source domain sample number and the corresponding monitored variable number, respectively. n t 1 and n t 1 represent the number of samples of LTD and ULTD, respectively. p t is the number of monitored variables in the target domain, and p t = p s . The procedure of the TrAdaBoost algorithm can be derived as follows:
Firstly, initialize the weight vector τ 1 = ( τ 1 1 τ n s + n t 1 1 ) , where:
τ i 1 = { 1 n s i = 1 , , n s 1 n t 1 i = n s + 1 , , n s + n t 1
Secondly, call the detection model (PrRVM); according to the detection results, the corresponding data weights are updated process as follows:
Based on [24], set β = 1 / ( 1 + 2 l n n s L ) . L is the number of iterations. Re-define the weight of SD data and LTD data, where:
τ i j = τ i j i = 1 n s + n t 1 τ i j
The sub-PrRVM ( φ j ) is trained using the data with the weight distribution of Equation (7). Then, return the detection model φ j : X Y . Calculate the error of φ j on X t 1 :
ϵ j = i = n s + 1 n s + n t 1 ω i j | φ j ( x i ) Y i | i = n s + 1 n s + n t 1 τ i j
The change parameter β j of X t 1 is obtained as follows:
β j = ϵ j 1 ϵ j
Then, updating the weight vector:
τ i j + 1 = { τ i j β | ϕ j ( x i ) Y i | i = 1 , , n s τ i j β j | ϕ j ( x i ) Y i | i = n s + 1 , , n s + n t 1
Finally, L sub-detection models ( φ 1 , φ 2 , , φ L ) are obtained through L iterations of the whole process.
In this paper, the common formulas are presented. The corresponding rigorous theoretical proof can is provided in previous research papers. For example, the weight distribution formula refers to the Hedge (β) theorem [25]. The proof of 1 / ( 1 + 2 l n n s L can be found in [26].

3.2. Adaptive Probabilistic Relevance Vector Machine

In this section, the evolution steps of the adaptive PrRVM within the joint framework of transfer learning and sparse Bayesian are further deduced. According to Section 2.2, we can obtain the probability derivation process of p ( y | w , σ 2 ) . In the derivation process, w and σ 2 need to be updated in each training process. Therefore, the iterative process of w and σ 2 in the transfer learning framework is re-defined. Assume that the posterior probability of w , σ 2 , and a can be expressed as p ( w , σ 2 , a | y ) . According to Bayesian inference, p ( w , σ 2 , a | y ) can be further decomposed as follows:
p ( w , σ 2 , a | y ) = p ( w , y , σ 2 , a ) p ( y , σ 2 , a ) p ( y , σ 2 , a ) p ( y ) = p ( w | y , σ 2 , a ) p ( σ 2 , a | y )
The solution of unknown parameters w , σ 2 , and a depends on p ( w | y , σ 2 , a ) and p ( σ 2 , a | y ) . For the classification problem, the posterior probability of weight w cannot be calculated directly. Here, we assume that the hyperparameter vector a is known, and p ( w | y , σ 2 , a ) can be further derived as follows:
p ( w | y , σ 2 , a ) = p ( w , y , σ 2 , a ) p ( y , σ 2 , a ) = p ( y | w , σ 2 , a ) p ( w | σ 2 , a ) p ( σ 2 , a ) p ( y | σ 2 , a ) p ( σ 2 , a )
To facilitate the subsequent derivation, we omit the indirect relationship between variables. According to the Bayes rule and Markov property, Equation (12) can be further transformed as follows:
p ( w | y , σ 2 , a ) = p ( y | w , σ 2 ) p ( w | a ) p ( y | σ 2 , a ) = p ( y | w , σ 2 ) p ( w | a ) p ( y | w , σ 2 ) p ( w | a ) d w
Therefore, p ( w | y , σ 2 , a ) p ( y | w , σ 2 ) p ( w | a ) . In addition, we can further deduce p ( σ 2 , a | y ) p ( y | σ 2 , a ) p ( σ 2 ) p ( a ) . Because we cannot directly solve p ( w | y , σ 2 , a ) and p ( σ 2 , a | y ) , we can solve p ( y | σ 2 , a ) and p ( y | w , σ 2 ) to derive the desired result. Here, the formulas of p ( y | w , σ 2 ) and p ( w | a ) can be expressed as follows:
p ( y | w , σ 2 ) = i = 1 n σ ( Φ i ) y i [ 1 σ ( Φ i ) ] 1 y i
p ( w | a ) = i = 0 n N ( w i | 0 , a i 1 ) = i = 0 n ( 2 π a i 1 ) 1 2 exp ( 1 2 a i w 2 ) = ( 2 π ) n + 1 2 | Λ | 1 2 exp ( 1 2 w T Λ w )
where Φ i = Φ ( x i ; w ) , Λ is the diagonal matrix:
Λ = ( a 0 0 0 0 a 1 0 0 0 0 a n )
When the hyperparameter vector a is fixed, Newton’s method can be used to solve p ( y | w , σ 2 ) p ( w | a ) :
log ( p ( y | w , σ 2 ) p ( w | a ) ) = log p ( y | w , σ 2 ) + log p ( w | a ) = i = 1 n [ y i log σ ( Φ i ) + ( 1 y i ) log ( 1 σ ( Φ i ) ] + Θ 1 2 w T Λ w
where Θ = log [ ( 2 π ) n + 1 2 i = 0 n a i 1 2 ] , because Equation (17) is a penalized logistic log-likelihood function, and necessitates iterative maximization [17]. The second-order Newton method is used to derive the target function. In addition, it can be further deduced that log p ( w | y , σ 2 , a ) l o g ( p ( y | w , σ 2 ) p ( w | a ) ) . Based on [27], we should take the derivative of w . Assuming that the solved extreme point is w M P , the second derivative result of w can be expressed as follows:
G = 2 w ( log p ( w | y , σ 2 , a ) ) | w M P = 2 w ( i = 1 n [ y i log σ ( Φ i ) + ( 1 y i ) log ( 1 σ ( Φ i ) ] + Θ 1 2 w T Λ w ) = Ψ T Η Ψ Λ
where Ψ is shown in Equation (3). Η is the diagonal matrix, Η = diag ( h 1 , h 2 , h n ) , and h i = σ ( Φ i ) ( 1 σ ( Φ i ) . Η can be written as
Η = ( σ ( Φ 1 ) ( 1 σ ( Φ 1 ) 0 0 σ ( Φ n ) ( 1 σ ( Φ n ) )
Based on [28], the covariance matrix and ( Ψ T Η Ψ Λ ) can be linked as follows:
Σ = ( G ) 1 = ( Ψ T Η Ψ + Λ ) 1
It can be seen that the Laplace approximation effectively maps the classification problem to a regression problem with data-dependent noise [29], with the inverse noise variance for ε given by σ ( Φ i ) ( 1 σ ( Φ i ) . In addition, according to w ( log p ( w | y , σ 2 , a ) ) | w M P = 0 and Σ , w M P can be further derived as follows:
w M P = Σ Ψ T Η y
Next, we can iteratively update hyperparameter vector a by fixing Σ and w M P . According to the relation p ( σ 2 , a | y ) p ( y | σ 2 , a ) p ( σ 2 ) p ( a ) , we only need to further simplify log ( p ( y | σ 2 , a ) ), and then repeat the previous derivation steps. The following relation can be obtained:
a i ( log p ( y | σ 2 , a ) ) = 1 2 a i 1 2 i i 1 2 u i 2 = 0
where u i = Σ Ψ T Η y = w M P . Equation (22) can be further converted as follows:
α i n e w = 1 α i Σ i i ( u i ) 2

3.3. Updating the Weight Vector and Sparse Analysis

In this section, the weight w is defined as the “hidden” variable. Then a general algorithm of expectation maximization (EM) is selected accordingly. EM mainly includes an expectation (E) step and a maximization (M) step. The adaptive PrRVM derived in this paper is used for classification. Assuming that ε is random additive noise, when the output is i = 0 n w i f i ( x j ) + ε 0 , the corresponding prediction label is p ( y = 1 | w , σ 2 ) = i = 1 n σ ( Φ i ) y i [ 1 σ ( Φ i ) ] 1 y i . The probit mode can be presented as follows:
p ( y = 1 | w , σ 2 ) = p ( i = 0 n w i f i ( x j ) + ε ) 0 )
The probability derivation of the weight vector w can be expressed as p ( w | y , σ 2 , a ) p ( y | w , σ 2 ) p ( w | a ) . The corresponding log-posterior is log ( p ( w | y , σ 2 , a ) ) = log p ( y | w , σ 2 ) + log p ( w | a ) . Suppose that the hyperparameter at time t is denoted as α i ( t ) . According to [30], define a new Q function and let Q ( w t | w t + 1 ) = log ( p ( w t | y ( t + 1 ) , ( σ 2 ) ( t ) , a ( t ) ) ) . The following expectation step can be derived:
Q ( w t | w t 1 ) = log p ( y ( t ) | w t , ( σ 2 ) ( t ) ) + log p ( w t | a ( t ) )
In the maximization step stage, we can update a ( t + 1 ) at the time of t + 1 through w t . Calculating the partial derivative of Q ( ) , we can then obtain the following α i ( t + 1 ) :
α i ( t + 1 ) = 1 Σ i i t + ( u i t ) 2
This scenario illustrates that the hyperparameters can be updated adaptively with available new inputs. In addition, during the update process, it is found that some α i n e w will approach infinity. At this time, the automatic correlation decision (ARD) can be used to update the corresponding u and Σ . When α i n e w approaches infinity, ARD will make the corresponding u i and Σ i i equal to zero [20]. w i is updated to zero. In this way, the matrix becomes sparse. Finally, it is assumed that the parameter probability estimation of the adaptive PrRVM is expressed by the following symbols: the weight parameter w = [ w 0 , w 1 , , w n ] and Λ = d i a g { a 0 , a 1 , , a n } . Additionally, rank ( Λ ) < n + 1 . The iteratively updated sparse matrix Σ = [ 0 0 0 ( Ψ T Η Ψ + Λ ) 1 ] , and the final prediction label y can be obtained.

3.4. Ensemble Detection Model and Key Performance Indicator

The finite number of adaptive sparse Bayesian transfer learning machines { φ 1 , φ 2 , , φ L } can be derived by Section 3.1 and Section 3.3. Moreover, the effective system decision making needs to consider the detection results of multiple adaptive sparse Bayesian transfer learning machines simultaneously. Based on [24], the following ensemble detection model can be constructed:
φ ( x ) = { 1 , j = [ L 2 ] L l n ( 1 β j ) φ j ( x ) 1 2 j = [ L 2 ] L l n ( 1 β j ) 0 , O t h e r w i s e
where:
β j = ϵ j / ( 1 ϵ j ) , i . e . , β j = i = n s + 1 n s + n t 1 ω i j | ϕ j ( x i ) Y i | i = n s + 1 n s + n t 1 ω i j / ( 1 i = n s + 1 n s + n t 1 ω i j | ϕ j ( x i ) Y i | i = n s + 1 n s + n t 1 ω i j )
When the above ensemble detection model is obtained, it is necessary to verify the performance of the model. KPIs are the critical decision tools for evaluating the method performance. They are the quantifiable and results-based statements. In this study, missed alarm rate (MAR), false alarm rate (FAR), accuracy, and pre-alarm rate (PAR) were carefully selected as KPIs. The corresponding formulas are as follows:
MAR = Fr ( Normal | Fault ) = F N F N + T P
FAR = Fr ( Fault | Normal ) = F P F P + T N
PAR = ϖ MAR + ( 1 ϖ ) FAR
accuracy = T P + T N T P + F P + T N + F N
Note that “normal” is the fault-free condition. Fr ( | ) represents the conditional frequency [12]. TP is true positive; TN represents true negative; FP is false positive. The PAR is constructed by combining the false alarm and missed alarm indicators [12]. ϖ is the weight parameter ( 0 ϖ 1 ).

4. Case Studies

4.1. Experimental Design and Compared Approaches

In this section, the dataset splitting steps are introduced in detail. Firstly, the SD data are filtered according to LTD data, in such a way that SD data and LTD data have the same types of labels. The data splitting is shown in Figure 2. Firstly, LTD data have the same distribution as ULTD data. IN contrast to the previous transfer learning, we combine the SD data and LTD data to form a new training set, and use the ULTD data as a testing set. The proportion of LTD data is 1%–10%. The “proportion” formula is defined as PR = LTD SD , where LTD is the number of labeled samples in the target domain, and SD is the number of samples in the source domain. Moreover, the main aim of the experiment is to monitor the single fault of the system. Multiple fault cases can be expanded accordingly. To verify the proposed method, traditional statistical methods and transfer learning methods were used to monitor a chemical plant and WWTP simultaneously. The traditional statistical methods PCA-T2 [31], SVM [32], and RVM [17] were trained using LTD data. RVMt and the proposed method were trained by the SD data and LTD data simultaneously.
In this study, the proposed EAdspB-TLM framework was used to monitor the TE chemical plant and a full-scale wastewater treatment plant (WWTP). The main tools used in the study were a personal computer (PC), MATLAB R2016a, SigmaPlot 12 and office software. The parameters of the PC are CPU Intel Core i7-6700HQ, 8 GB RAM, and 1 TB SSD. The data are from a TE simulation platform and a real WWTP.

4.2. Case Study on the Tennessee Eastman Chemical Process

4.2.1. Background

The Tennessee Eastman chemical process (TECP) was designed by a chemical company as a testing process control and diagnosis method. As shown in Figure 3, the TECP consists of five core units: reactor, compressor, stripper, condenser, and separator. The process includes measured variables and manipulated variables. There are four gaseous reactants (A, B, C, D, E) and two liquid products (G and H). The reaction equation is as follows:
{ A ( g ) + C ( g ) + D ( g ) G ( liq ) A ( g ) + C ( g ) + E ( g ) H ( liq ) A ( g ) + E ( g ) F ( byproduct ) 3 D ( g ) 2 F ( byproduct )
where F is the byproduct in the reactor, and the process is irreversible. More detailed reaction information of the TEP can be found in [33]. Moreover, the simulation program and operation introduction can be downloaded from http://depts.washington.edu/control/LARRY/TE/download.html#Basic_TE_Code. According to [33], 52 observation variables were selected for process monitoring depending on the process importance. Firstly, the platform started with a 25 h steady state. Then, the simulation ran for 97 h in each case. The sampling time was set up as 3 min. The source domain dataset resulted from the initial 59 h simulation. The corresponding dataset started with a normal working condition, but with faults being imposed after 39 simulation hours. Target domain (TD) data were collected from the simulation period of 59–97 h. TD data mainly includes two parts: the LTD and ULTD datasets. In this study, the ULTD dataset is defined as the testing dataset.

4.2.2. Analysis and Discussion of Experimental Results

EAdspB-TLM differs from previous modeling methods. It has the abilities of adaptive adjustment and transfer learning. To verify the performance of the proposed EAdspB-TLM, five fault cases were used. The fault type description is shown in Table 1. Simultaneously, according to the engineering experience and cross-validation, the proposed method basic parameters were set by trial and error: the kernel function is “Gauss” ϖ = 0.6, and the maximum number of iterations and period are 1000 and 100, respectively.
In this study, step and sticking faults are the most noteworthy among the above five type faults. When the external disturbance is strong, it is easy to cause step faults of the sensor or other equipment. Table 2 presents monitoring results for Fault 1. In addition to the transfer learning methods of EAdspB-TLM, unsupervised statistical (PCA) and supervised statistical (SVM, RVM) methods are also presented. It is worth noting that the step fault of D feed temperature is a kind of fault that is difficult to monitor. Because the abnormality is not obvious, most monitoring models cannot effectively monitor this fault [3,34]. According to the experimental results, when PR = 8%, the detection accuracy of EAdspB-TLM is the highest among the five methods, with an accuracy of 87.2%. When PR = 3%, the accuracy of EAdspB-TLM is only 85.48%. In addition, the PAR of EAdspB-TLM is the lowest among the five methods. For example, when PR = 8%, the PAR of EAdspB-TLM is 13.52%. Moreover, the PAR of PCA-T2 and RVM are 49.06% and 16.83%, respectively. Moreover, the missed alarm rate of EAdspB-TLM is the lowest among the five methods. This shows that the proposed EAdspB-TLM is effective.
Fault 5 relates to a control problem with the reactor cooling water valve, which is a common sticking fault in the Tennessee Eastman chemical process (TECP). The reactor is an important component in the normal operation of the chemical plant. Once the fault occurs, other components (reactor, compressor, etc.) will not function normally. Therefore, it is imperative to monitor Fault 5 in real-time. The monitoring results for Fault 5 are tabulated in Table 3. When the PR value increased from 3% to 8%, the detection accuracy of EAdspB-TLM improved from 93.22% to 96.88%. Moreover, the detection accuracy of EAdspB-TLM is much higher than that of the other four methods. It is worth noting that when the PR value reaches 8%, the accuracy of EAdspB-TLM is 96.88%. The detection accuracy of EAdspB-TLM to Fault 5 is much higher than that of Fault 1. This indicates that the complexity of Fault 1 is higher than that of Fault 5.
Overall, the performance of the monitoring method improves with the increase of PR value. Table 4 shows the average values of false and missed alarms of the five methods, that is, the average value of all faults in monitoring the TE chemical process. In terms of false alarm, when PR increases from 3% to 8%, false alarms of EAdspB-TLM decrease from 2.2% to 1.62%, false alarms of RVM decrease from 21.6% to 5.12%, and false alarms of SVM decrease from 32.02% to 19.06%. This shows that with the increase of PR, the false alarms of the methods will gradually decrease. In addition, the average value of the two comprehensive KPIs is shown in Figure 4: the PR alarm rate is shown in Figure 4a, and the fault diagnosis accuracy is shown in Figure 4b. It can be seen that with the increase of PR value, the fault diagnosis accuracy of the five monitoring methods gradually increases. In addition, the EAdspB-TLM method has the highest fault diagnosis accuracy; when the PR value increases to 8%, the accuracy of EAdspB-TLM is 94.61%.

4.3. Case Study of Full-Scale WWTP

4.3.1. Background

In this case study, the proposed method was used to monitor a real full-scale WWTP. The plant serves a population of 480,000, with a daily treatment flow of 170,000 m3 and a hydraulic retention time of 16.5 h. A long solid residence time (SRT) is used to achieve good nitrogen removal performance, and is typically maintained at 15–22 days. The schematic of the WWTP is shown in Figure 5. It is mainly composed of three components: selector, oxidation ditch, and secondary settler. Due to external disturbances, such as weather, temperature, and sludge activity, the filamentous sludge bulking occurs frequently and is difficult to monitor online in real time. The data were collected from 1 September to the following 31 March. Fifteen observation variables were selected as modeling variables. The sampling interval was one day and filamentous sludge bulking occurred during this period. The source domain dataset is based on the first samples of 110 days. This dataset starts with normal working conditions, but with faults occurring after 70 sample days.

4.3.2. Analysis and Discussion of Experimental Results

Filamentous sludge bulking is a type of drift fault [35]. In contrast to the abrupt fault, sludge bulking may return to normal after the self-regulation of microorganisms in the early stage [36]. During this period, the abnormality is less obvious (Figure 6). Figure 6a shows the dynamic trend of BOD5 (the five-day biochemical oxygen demand), and Figure 6b shows the curve of the sludge volume index (SVI). These can be used to determine if sludge bulking occurs in the WWTP. Although these indicators can be used to identify whether there is sludge bulking in the WWTP, the experiment is time-consuming. Therefore, real-time monitoring of the WWTP cannot be effectively implemented. In addition, the consecutive filamentous sludge bulking will cause the secondary pollution to the environment. Therefore, it is desirable to design an effective method for real-time monitoring of sludge bulking of WWTPs.
When the data are obtained by the TECP, the technology of Section 4.1 was used to split the data accordingly. Then the proposed method and the other four methods were simultaneously used to monitor the wastewater treatment process; the false alarm rate, missed alarm rate, accuracy, and pre-alarm-rate of the five methods are tabulated in Table 5. Because sludge bulking is a slow drift fault, false and missed alarms become more obvious. When PR = 8%, the false alarm rate of PCA-T2 is 11.11%. The FAR of the SVM and RVM are both 1.85%. In comparison, the false alarm rate of EAdspB-TLM is zero. Furthermore, EAdspB-TLM is not the best in terms of missed alarms. The missed alarm rate of RVMt is higher than that of EAdspB-TLM. This unconventional result implies that EAdspB-TLM may not always be optimal. Thus, we need to further explore the effectiveness of EAdspB-TLM using the comprehensive KPIs (PAR and accuracy). According to Table 5 and Figure 7, the EAdspB-TLM-based pre-alarm rate is the lowest among the five methods (Figure 7a). For example, when PR = 8%, the PAR of EAdspB-TLM is 4.62%. In addition, the PARs of the comparison methods RVMt and SVM are 5.19% and 13.05%, respectively. In addition, when the PR value increased from 3% to 8%, the pre-alarm rate of EAdspB-TLM decreased from 5.09% to 4.62%. Based on the above analysis, we can conclude that the performance of EAdspB-TLM is the best among the six monitoring methods. At the same time, with the increase of PR value, the performance of the six methods is improved. Additionally, the fault diagnosis accuracy further verifies this conclusion, which is shown in Figure 7b. When PR = 8%, the average detection accuracy of EAdspB-TLM reaches 96.77%.

5. Conclusions

In this paper, a process monitoring framework, termed EAdspB-TLM, is proposed for monitoring nonlinear large-scale processes. When training data are insufficient to train a reliable model, traditional process monitoring methods cannot work well. As a result, faults of wastewater treatment and chemical processes cannot be identified and pre-alarmed in time, thus increasing the cost of system maintenance. Therefore, the proposed EAdspB-TLM was equipped with the ability of transfer learning, which allows useful information of unused data to be transferred to assist in training the model. EAdspB-TLM effectively alleviates the problem of insufficient label data in factories. Furthermore, the corresponding results also further verify the feasibility of the proposed EAdspB-TLM. According to the experimental results, with the increase of labeled target domain data, the diagnostic accuracy of EAdspB-TLM is improved. In addition, the pre-alarm rate (PAR) of EAdspB-TLM is also reduced. Overall, EAdspB-TLM achieved the best performance in monitoring the wastewater treatment and TE chemical processes. Using the WWTP as an example, when PR = 8%, the accuracy of the five methods can be ranked as follows: EAdspB-TLM (96.77%) > RVMt (92.47%) > SVM (90.32%) > PCA-T2 (84.95%) > RVM (68.82%).
The batch dataset needed for wastewater treatment process monitoring is drawn mostly from a collection of sensors. However, the data collected by some sensors has little value in training the monitoring model. Therefore, future research work will aim to optimize the number of selected sensors for monitoring and improve the monitoring efficiency of EAdspB-TLM.

Author Contributions

Material preparation, data collection, and analysis were performed by H.C., C.X., J.W., H.C. performed the experiments and wrote the paper. Y.L. reviewed and revised the paper. The funding was provided by J.W., D.H. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (61873096, 62073145), Guangdong Basic and Applied Basic Research Foundation (2020A1515011057), Guangdong Technology International Cooperation Project Application (2020A0505100024). Fundamental Research Funds for the central Universities, SCUT (D2201200). Science and Technology Planned Project of Guizhou Province ([2020]1Y276). The authors thank the anonymous referees and other people for their help to improve the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ma, S.; Cai, W.; Liu, W.; Shang, Z.; Liu, G. A Lighted Deep Convolutional Neural Network Based Fault Diagnosis of Rotating Machinery. Sensors 2019, 19, 2381. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Newhart, K.B.; Holloway, R.W.; Hering, A.S.; Cath, T.Y. Data-driven performance analyses of wastewater treatment plants: A review. Water Res. 2019, 157, 498–513. [Google Scholar] [CrossRef] [PubMed]
  3. Shang, J.; Chen, M.; Ji, H.; Zhou, D. Recursive transformed component statistical analysis for incipient fault detection. Automatica 2017, 80, 313–327. [Google Scholar] [CrossRef]
  4. Liu, K.; Li, Y.; Yang, J.; Liu, Y.; Yao, Y. Generative Principal Component Thermography for Enhanced Defect Detection and Analysis. IEEE Trans. Instrum. Meas. 2020, 69, 1. [Google Scholar] [CrossRef]
  5. Jiang, Q.; Yan, X. Neighborhood Stable Correlation Analysis for Robust Monitoring of Multiunit Chemical Processes. Ind. Eng. Chem. Res. 2020, 59, 16695–16707. [Google Scholar] [CrossRef]
  6. Liu, Y.; Pan, Y.; Sun, Z.; Huang, D. Statistical Monitoring of Wastewater Treatment Plants Using Variational Bayesian PCA. Ind. Eng. Chem. Res. 2014, 53, 3272–3282. [Google Scholar] [CrossRef]
  7. Ge, Z.; Song, Z. Process Monitoring Based on Independent Component Analysis−Principal Component Analysis (ICA−PCA) and Similarity Factors. Ind. Eng. Chem. Res. 2007, 46, 2054–2063. [Google Scholar] [CrossRef]
  8. Zhu, J.; Ge, Z.; Song, Z. Non-Gaussian Industrial Process Monitoring With Probabilistic Independent Component Analysis. IEEE Trans. Autom. Sci. Eng. 2016, 14, 1309–1319. [Google Scholar] [CrossRef]
  9. Hu, J.; Tse, P.W. A Relevance Vector Machine-Based Approach with Application to Oil Sand Pump Prognostics. Sensors 2013, 13, 12663–12686. [Google Scholar] [CrossRef] [Green Version]
  10. Feng, Q.; Zhao, X.; Fan, D.; Cai, B.; Liu, Y.; Ren, Y. Resilience design method based on meta-structure: A case study of offshore wind farm. Reliab. Eng. Syst. Saf. 2019, 186, 232–244. [Google Scholar] [CrossRef]
  11. Sun, B.; Li, Y.; Wang, Z.; Ren, Y.; Feng, Q.; Yang, D. An Improved Inverse Gaussian Process with Random Effects and Measurement Errors for RUL Prediction of Hydraulic Piston Pump. Measurement 2020, 108604. [Google Scholar] [CrossRef]
  12. Cheng, H.; Liu, Y.; Huang, D.; Liu, B. Optimized Forecast Components-SVM-Based Fault Diagnosis with Applications for Wastewater Treatment. IEEE Access 2019, 7, 128534–128543. [Google Scholar] [CrossRef]
  13. Lee, J.-M.; Yoo, C.; Choi, S.W.; Vanrolleghem, P.A.; Lee, I.-B. Nonlinear process monitoring using kernel principal component analysis. Chem. Eng. Sci. 2004, 59, 223–234. [Google Scholar] [CrossRef]
  14. Wang, L.; Shi, H. Multivariate statistical process monitoring using an improved independent component analysis. Chem. Eng. Res. Des. 2010, 88, 403–414. [Google Scholar] [CrossRef]
  15. Yélamos, I.; Escudero, G.; Graells, M.; Puigjaner, L. Simultaneous fault diagnosis in chemical plants using support Vector Machines. Comput. Aided Chem. Eng. 2007, 24, 1253–1258. [Google Scholar]
  16. Liu, Y.; Xie, M. Rebooting data-driven soft-sensors in process industries: A review of kernel methods. J. Process. Control. 2020, 89, 58–73. [Google Scholar] [CrossRef]
  17. Tipping, M.E. Sparse Bayesian learning and the relevance vector machine. J. Mach. Learn. Res. 2001, 1, 211–244. [Google Scholar]
  18. Liu, Y.; Liu, B.; Zhao, X.; Xie, M. Development of RVM-Based Multiple-Output Soft Sensors With Serial and Parallel Stacking Strategies. IEEE Trans. Control. Syst. Technol. 2019, 27, 2727–2734. [Google Scholar] [CrossRef]
  19. Wu, J.; Cheng, H.; Liu, Y.; Huang, D.; Yuan, L.; Yao, L. Learning soft sensors using time difference–based multi-kernel relevance vector machine with applications for quality-relevant monitoring in wastewater treatment. Environ. Sci. Pollut. Res. 2020, 27, 28986–28999. [Google Scholar] [CrossRef]
  20. Mackay, D.J.C. Bayesian Methods for Backpropagation Networks. In Models of Neural Networks III; Springer Science and Business Media LLC: Berlin, Germany, 1996; pp. 211–254. [Google Scholar]
  21. Li, X.; Zhang, W.; Ding, Q.; Li, X. Diagnosing Rotating Machines with Weakly Supervised Data Using Deep Transfer Learning. IEEE Trans. Ind. Inform. 2019, 16, 1688–1697. [Google Scholar] [CrossRef]
  22. Dong, A.; Chung, F.; Deng, Z.; Wang, S. Semi-Supervised SVM With Extended Hidden Features. IEEE Trans. Cybern. 2015, 46, 2924–2937. [Google Scholar] [CrossRef] [PubMed]
  23. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar] [CrossRef]
  24. Dai, W.; Yang, Q.; Xue, G.-R.; Yu, Y. In Boosting for Transfer Learning. In Proceedings of the 24th International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007; pp. 193–200. [Google Scholar]
  25. Freund, Y.; Schapire, R.E. A Desicion-Theoretic Generalization of on-line Learning and an Application to Boosting. In European Conference on Computational Learning Theory; Springer: Berlin, Germany, 1995; pp. 23–37. [Google Scholar]
  26. Rojas, R. AdaBoost and the Super Bowl of Classifiers a Tutorial Introduction to Adaptive Boosting; Freie University: Berlin, Germany, 2009; Volume 1, pp. 1–6. [Google Scholar]
  27. Mackay, D.J.C. The Evidence Framework Applied to Classification Networks. Neural Comput. 1992, 4, 720–736. [Google Scholar] [CrossRef]
  28. Tzikas, D.; Likas, A.; Galatsanos, N. Sparse Bayesian Modeling with Adaptive Kernel Learning. IEEE Trans. Neural Netw. 2009, 20, 926–937. [Google Scholar] [CrossRef]
  29. Bishop, C.M.; Tipping, M.E. Bayesian regression and classification. In NATO Science Series, III: Computer and Systems Sciences; IOS Press: Amsterdam, The Netherlands, 2003; Volume 190, pp. 267–288. [Google Scholar]
  30. Chen, H.; Tino, P.; Yao, X. Probabilistic Classification Vector Machines. IEEE Trans. Neural Netw. 2009, 20, 901–914. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. GarcÇa-Alvarez, D. Fault detection using principal component analysis (PCA) in a wastewater treatment plant (WWTP). In Proceedings of the International Student’s Scientific Conference, Wismar, Germany, 15 January 2009; pp. 55–60. [Google Scholar]
  32. Hsu, C.-W.; Lin, C.-J. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13, 415–425. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Downs, J.; Vogel, E. A plant-wide industrial process control problem. Comput. Chem. Eng. 1993, 17, 245–255. [Google Scholar] [CrossRef]
  34. Yin, S.; Ding, S.X.; Haghani, A.; Hao, H.; Zhang, P. A comparison study of basic data-driven fault diagnosis and process monitoring methods on the benchmark Tennessee Eastman process. J. Process. Control. 2012, 22, 1567–1581. [Google Scholar] [CrossRef]
  35. Liu, Y.; Guo, J.-H.; Wang, Q.; Huang, D. Prediction of Filamentous Sludge Bulking using a State-based Gaussian Processes Regression Model. Sci. Rep. 2016, 6, 31303. [Google Scholar] [CrossRef] [Green Version]
  36. Seka, A.M.; Van De Wiele, T.; Verstraete, W. Feasibility of a multi-component additive for efficient control of activated sludge filamentous bulking. Water Res. 2001, 35, 2995–3003. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the proposed process monitoring framework.
Figure 1. Schematic diagram of the proposed process monitoring framework.
Sensors 20 06139 g001
Figure 2. The flowchart of data splitting.
Figure 2. The flowchart of data splitting.
Sensors 20 06139 g002
Figure 3. Schematic layout of the Tennessee Eastman chemical process (TECP) [33].
Figure 3. Schematic layout of the Tennessee Eastman chemical process (TECP) [33].
Sensors 20 06139 g003
Figure 4. Average of comprehensive key performance indicators (KPIs) based on the proposed method and four other methods: (a) pre-alarm rate (PAR); (b) accuracy.
Figure 4. Average of comprehensive key performance indicators (KPIs) based on the proposed method and four other methods: (a) pre-alarm rate (PAR); (b) accuracy.
Sensors 20 06139 g004
Figure 5. Schematic layout of a real full-scale wastewater treatment process (WWTP).
Figure 5. Schematic layout of a real full-scale wastewater treatment process (WWTP).
Sensors 20 06139 g005
Figure 6. Isolated variables change curves of sludge bulking: (a) Five-day biochemical oxygen demand (BOD5); (b) the sludge volume index (SVI).
Figure 6. Isolated variables change curves of sludge bulking: (a) Five-day biochemical oxygen demand (BOD5); (b) the sludge volume index (SVI).
Sensors 20 06139 g006
Figure 7. Average of comprehensive KPIs based on the proposed method and four other methods: (a) PAR; (b) accuracy.
Figure 7. Average of comprehensive KPIs based on the proposed method and four other methods: (a) PAR; (b) accuracy.
Sensors 20 06139 g007
Table 1. Faults description of TECP.
Table 1. Faults description of TECP.
No.DescriptionType
Fault 1D feed temperature (stream 2)Step
Fault 2C header pressure loss-reduced availability (stream4)Step
Fault 3Condenser cooling water inlet temperatureRandom
Fault 4Reaction kineticsSlow drift
Fault 5Reactor cooling water valveSticking
Table 2. Monitoring results for Fault 1.
Table 2. Monitoring results for Fault 1.
KPI PCA-T2SVMRVMRVMtEAdspB-TLM
3%0.51790.06550.01790.10710.0565
FAR5%0.38940.02490.05610.10900.0654
8%0.37710.03370.03700.06060.0438
3%0.28420.74420.90440.27130.2222
MAR5%0.32720.83110.73610.23750.2190
8%0.36240.78470.79290.24520.1962
3%0.37770.47270.54980.20560.1560
PAR5%0.35210.50860.46410.18610.1576
8%0.36830.48430.49060.17140.1352
3%0.60720.57120.50760.80500.8548
Acc5%0.64430.53860.57570.82140.8514
8%0.63100.55120.54520.83730.8720
Table 3. Monitoring results for Fault 5.
Table 3. Monitoring results for Fault 5.
KPI PCA-T2SVMRVMRVMtEAdspB-TLM
3%0.52680.09520.12800.01490.0060
FAR5%0.27100.12460.23360.01870.0093
8%0.13800.07410.06060.00340.0067
3%0.37980.35140.50650.16800.1214
MAR5%0.52510.42740.37730.14510.1135
8%0.53950.41960.40330.07900.0654
3%0.43860.24890.35510.10670.0752
PAR5%0.42350.30630.31980.09450.0718
8%0.37890.28140.26620.04880.0419
3%0.55190.76760.66940.90320.9322
Acc5%0.59140.71140.68860.91290.9343
8%0.64010.73490.75000.95480.9608
Table 4. Average of false alarm rate (FAR) and missed alarm rate (MAR).
Table 4. Average of false alarm rate (FAR) and missed alarm rate (MAR).
KPI
(Average)
PCA-T2SVMRVMRVMtEAdspB-TLM
3%0.39290.32020.21610.03810.0220
FAR5%0.33270.14890.26170.03490.0218
8%0.43970.19060.05120.01890.0162
3%0.27390.32920.54470.15350.1323
MAR5%0.32030.57360.41210.15300.1277
8%0.30740.35910.42620.13410.0845
Table 5. Monitoring results for filamentous sludge bulking.
Table 5. Monitoring results for filamentous sludge bulking.
KPI PCA-T2SVMRVMRVMtEAdspB-TLM
3%0.08770.07020.00000.15790.0175
FAR5%0.03570.03570.00000.16070.0179
8%0.11110.01850.01850.12960.0000
3%0.53660.43901.00000.00000.0732
MAR5%0.56100.43900.95120.00000.0732
8%0.20510.20510.71790.00000.0769
3%0.35700.29150.60000.06320.0509
PAR5%0.35090.27770.57070.06430.0510
8%0.16750.13050.43820.05190.0462
3%0.72450.77550.58160.90820.9592
Acc5%0.74230.79380.59790.90720.9588
8%0.84950.90320.68820.92470.9677
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cheng, H.; Liu, Y.; Huang, D.; Xu, C.; Wu, J. A Novel Ensemble Adaptive Sparse Bayesian Transfer Learning Machine for Nonlinear Large-Scale Process Monitoring. Sensors 2020, 20, 6139. https://doi.org/10.3390/s20216139

AMA Style

Cheng H, Liu Y, Huang D, Xu C, Wu J. A Novel Ensemble Adaptive Sparse Bayesian Transfer Learning Machine for Nonlinear Large-Scale Process Monitoring. Sensors. 2020; 20(21):6139. https://doi.org/10.3390/s20216139

Chicago/Turabian Style

Cheng, Hongchao, Yiqi Liu, Daoping Huang, Chong Xu, and Jing Wu. 2020. "A Novel Ensemble Adaptive Sparse Bayesian Transfer Learning Machine for Nonlinear Large-Scale Process Monitoring" Sensors 20, no. 21: 6139. https://doi.org/10.3390/s20216139

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop