Next Article in Journal
Complexity as Causal Information Integration
Previous Article in Journal
Fractional Norms and Quasinorms Do Not Help to Overcome the Curse of Dimensionality
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evidential Estimation of an Uncertain Mixed Exponential Distribution under Progressive Censoring

School of Mathematics and Statistics, Northwestern Polytechnical University, Xi’an 710072, Shaanxi, China
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(10), 1106; https://doi.org/10.3390/e22101106
Submission received: 13 August 2020 / Revised: 27 September 2020 / Accepted: 29 September 2020 / Published: 30 September 2020
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
In this paper, the evidential estimation method for the parameters of the mixed exponential distribution is considered when a sample is obtained from Type-II progressively censored data. Different from the traditional statistical inference methods for censored data from mixture models, here we consider a very general form where there is some uncertain information about the sub-class labels of units. The partially specified label information, as well as the censored data are represented in a united frame by mass functions within the theory of belief functions. Following that, the evidential likelihood function is derived based on the completely observed failures and the uncertain information included in the data. Then, the optimization method using the evidential expectation maximization algorithm (E2M) is introduced. A general form of the maximal likelihood estimates (MLEs) in the sense of the evidential likelihood, named maximal evidential likelihood estimates (MELEs), can be obtained. Finally, some Monte Carlo simulations are conducted. The results show that the proposed estimation method can incorporate more information than traditional EM algorithms, and this confirms the interest in using uncertain labels for the censored data from finite mixture models.

1. Background

Mixture models are of great importance in many applied sciences such as survival analysis, pattern recognition, image analysis, economics, and so on [1,2]. In reliability analysis, there are only one population and one type of failure in the simple case of life distributions. However, in real applications, there may be more than one failure cause [3,4]. In this case, the response of the modeling process can be seen as from several distinct sub-populations, and finite mixture models can be used to represent the time of failures.
Suppose the life lengths of these n units are independent and identically distributed (i.i.d.) random variables with probability density function (pdf) f ( x ; θ ) and cumulative distribution function (cdf) F ( x ; θ ) . Consider the mixture model of the form:
f ( x ; θ ) = k = 1 m π k f k ( x ; λ k ) ,
where θ = { λ 1 , , λ m ; π 1 , , π m } , with  π j 0 , j = 1 , , m and j π j = 1 . In this paper, we focus on the case where f k ( x ; λ k ) is the exponential density, although the model we develop can be applied more generally. The mixed exponential distribution (MED) model with m components has its pdf and cdf respectively as:
f ( x ; θ ) = k = 1 m π k f k ( x ; λ k ) = k = 1 m π k λ k e λ k x
and:
F ( x ; θ ) = 1 k = 1 m π k s k ( x ; λ k ) = 1 k = 1 m π k e λ k x ,
where the parameters are π = ( π 1 , , π m ) and λ = ( λ 1 , , λ m ) with:
k = 1 m π k = 1 , 0 < π k < 1 ,
and:
λ k > 0 , k = 1 , 2 , , m .
As we can see, the samples from mixed distributions are not precisely labeled by their origin sub-classes; thus, the heterogeneous dataset cannot be explicitly decomposed to homogeneous subgroups. Consequently, it brings about a barrier for estimating finite mixture models [5].
Consider a life-testing experiment where n units are placed under observation. For some reason, such as to save time and cost, we have to terminate the experiment before all items have failed. In real applications, the removal of units prior to failure is often pre-planned. Data obtained from such a type of experiments are called censored data. The most common censoring schemes are termed Type-I and Type-II censoring. In Type-I censoring, the experiment continues up to a prescribed time T. Any failures that occur after T cannot been observed. The endpoint of the experiment T is assumed to be independent of the failure times. In Type-II censoring, the experiment is terminated upon the N t h failure, where N < n is prefixed. The experiments under these two kinds of test schemes have the drawback that they do not allow removal of samples at time points other than the termination of the experiment. However, the Type-II progressive censoring, which can be seen as a generalization of Type-II censoring, allows removing some fixed number of units at the time of the first observed m 1 failures and removing all the remaining at the time of the m t h failure, at which time, the experiment terminates [6]. As a result, it is efficient in time and money [7] and has become very popular in the last few years [8,9,10,11,12,13,14].
The parameter estimation problems for different lifetime distributions, including the mixed exponential distribution and some other mixed distribution models, have been widely studied under different censoring schemes [15,16,17]. The most commonly adopted methods to obtain the estimations for mixture models are simply using the expectation maximization (EM) algorithm or Bayesian method. Lee and Scott [18] presented the EM algorithm for fitting multivariate Gaussian mixture models to data that are truncated, censored, or truncated and censored. In [19,20], the authors discussed the parameter estimation methods of the MED under Type-II progressively censored data and progressively hybrid Type-II censored data, respectively. Tahir et al. [21] studied the problem of estimating the parameters of a three-component mixture of exponential, Rayleigh and Burr Type-XII distributions using the Type-I right censoring scheme in the Bayesian framework. The maximum likelihood and Bayesian estimators of the parameters of a heterogeneous population represented by a finite mixture of two Pareto distributions were discussed in [22]. Feroze and Aslam [23] introduced the Bayesian approach for estimating the parameters of the two-component mixture of Weibull distribution under doubly censored samples. Feroze and Aslam [24] considered the Bayesian analysis of the three-component mixture of the Rayleigh distribution under doubly censored samples. All these works assume that we do not have any information about the label of the sub-class at all. However, in many applications, often, it is easy to get some imprecise and uncertain knowledge about the sub-class labels. For instance, in medical surveillance databases, we can find partially labeled data provided by experts or from experience, that is, while not completely unlabeled, there is only uncertain information about class values [5]. The above-mentioned methods cannot deal with the possible available partial label information in the dataset from mixture models.
Zio [25] pointed out that uncertain information is a big challenge in reliability engineering. Specifically, there indeed exits many kinds of uncertain information in reliability analysis for censored data. First, we only know that the failure time of the censored unit belongs to the interval [ t * , + ) . Second, as mentioned, in the finite mixture models, there may be some uncertain information about the sub-class labels of the data. The theory of belief functions (also known as Dempster–Shafer theory (DST)) is appealing to represent data uncertainty. As an extension of probability theory, it has many advantages in dealing with uncertain information. Similar to the probability distribution over the discernment frame, in DST, the basic belief assignment (BBA) defined on the power set of the frame is used to represent the available information. Many scholars have studied how to measure the uncertainty of BBA [26,27,28,29]. Due to the effectiveness of DST in dealing with uncertain knowledge, it has already been widely used in many fields such as data classification/clustering [30,31], target recognition [32], decision making [33], fault diagnosis [34,35], complex networks [36,37], and so on.
In this paper, we try to handle the uncertain information in mixed distributions under Type-II progressive censoring using the theory of belief functions. Note that for the progressively censored data considered in this work, the incomplete observations are not influenced by the specific values taken by the random variables. Thus, the mechanism that causes complete failure data cases to become incompletely reported can be ignored. Under such a kind of coarsening at random (CAR) assumption, the statistical inference can proceed based on the so-called “face-value likelihood”, which measures the probability of incomplete observations by its marginal probability according to the underlying complete data distribution [38,39]. The EM algorithm is quite effective at maximizing the face-value likelihood [38], and it has been widely used for progressively censored data [7,40]. Different from the traditional estimation method using the EM algorithm, the proposed evidential parameter estimation model can take not only the uncertain censored observations, but also the prior partial information about class labels into consideration. The two kinds of uncertain information are modeled in a united frame by mass functions in belief function theory, and then, the evidential likelihood function is derived. The optimization method to get the optimal estimators, called maximal evidential likelihood estimates (MELEs), is derived based on the evidential-EM (E2M) algorithm [41]. Experimental results show that the proposed model can take advantage of the the partial information about the sub-class labels effectively and consequently improve the performance of the estimation model.
The rest of this paper is organized as follows. In Section 2, some basic concepts and the rationale of our method are briefly introduced. In Section 3, the evidential estimation model to get MELEs, which can maximize the evidential likelihood function, is presented in detail. In order to show the effectiveness of our approach, a real data application is discussed in Section 4, while some Monte Carlo simulations are conducted in Section 5. Finally, we conclude the paper in Section 6.

2. Preliminary Knowledge

Some necessary background knowledge related to this paper will be recalled in this section. The Type-II progressive censoring scheme is first introduced in Section 2.1. The basic knowledge about the theory of belief functions is then presented in Section 2.2. Finally, the evidential likelihood is recalled in Section 2.3.

2.1. Type-II Progressively Censoring Scheme

The Type-II progressive censoring scheme can be described as follows. Suppose n independent identical items are placed on a life-test. The integer N ( < n ) , which denotes the number of failures we would like to observe in the experiment, is fixed before the experiment. Let R = ( R 1 , R 2 , , R N ) , where R 1 , R 2 , , R N are also fixed integers describing the progressive censoring scheme. They should satisfy:
R 1 + R 2 + + R N + N = n .
At the time of the first failure, say T 1 , R 1 of the remaining units are randomly removed. Similarly, at the time of the i t h failure, say T i , R i ( i = 1 , 2 , , N ) of the remaining units are removed. At the time of the N t h failure, say T N , the remaining R N = n N i = 1 N 1 R i items are removed, and the experiment terminates. Therefore, in the presence of the Type-II progressive censoring scheme, the observed failures are { T 1 , , T N } . For further details on this censoring scheme, the readers may refer to the excellent monograph of Balakrishnan and Aggarwala [6].

2.2. Theory of Belief Functions

The theory of belief functions is a mathematical theory that generalizes the theory of probabilities by giving up the additivity constraint. In this theory, justified degrees of support are assessed according to an evidential corpus, which is the set of all evidential pieces of evidence held by a source that justifies the degrees of support assigned to some subsets.
Let Ω = { ω 1 , ω 2 , , ω c } be the finite domain of reference, called the discernment frame. The c elements in Ω are nonempty and mutually exclusive hypotheses related to a given problem. The belief functions are defined on the power set 2 Ω = { A : A Ω } . The function m : 2 Ω [ 0 , 1 ] is said to be the basic belief assignment on 2 Ω , if it satisfies:
A Ω m ( A ) = 1 .
Every A 2 Ω such that m ( A ) > 0 is called a focal element. The difference from probability models is that masses can be given to any subset of Ω instead of only to the atomic element of Ω . The credibility and plausibility functions are defined as in Equations (6) and (7), respectively:
B e l ( A ) = B A , B m ( B ) , A Ω ,
P l ( A ) = B A m ( B ) , A Ω .
Each quantity B e l ( A ) measures the total support given to A, while P l ( A ) represents the potential amount of support to A. The function p l : Ω [ 0 , 1 ] such that p l ( ω ) = P l ( { ω } ) is called the contour function associated with m.
According to the type of the focal elements, we can define some particular mass functions. A categorical mass function is a normalized mass function that has a unique focal element A * . This kind of mass function can be defined as:
m ( A ) = 1 if A = A * Ω 0 otherwise .
A vacuous mass function is a particular categorical mass function focused on Ω . It is a special kind of categorical mass function with a unique focal element Ω . This type of mass function is defined as follows:
m ( A ) = 1 if A = Ω 0 otherwise .
The vacuous mass function emphasizes the case of total ignorance.
A Bayesian mass function is a mass function for which all focal elements are elementary hypotheses, i.e., the focal elements are all singletons. It is defined as follows:   
m ( A ) = a [ 0 , 1 ] if | A | = 1 0 otherwise .
As all focal elements are single points, this mass function is a probability distribution over frame Ω . Specifically, if a Bayesian mass function is categorical, it describes that there is no uncertainty at all, and we are completely sure about the state of the variable concerned.
If m 1 and m 2 are two independent mass functions defined on Ω , a new mass function can be formed by combining m 1 and m 2 using Dempster’s rule directly:
( m 1 m 2 ) ( A ) = 1 1 k B C = A m 1 ( B ) m 2 ( C ) ,
where k is defined as
k = B C = m 1 ( B ) m 2 ( C ) .
This describes the conflict between m 1 and m 2 . When k = 1 , the two masses are completely in conflict, and they cannot be combined using Dempster’s rule.
Suppose m 1 is a Bayesian mass function, and its corresponding contour function is a probability distribution function defined by p 1 ( ω ) = m ( { ω } ) . Assume m 2 is an arbitrary mass function with contour function p l 2 ( ω ) . The fused mass function of m 1 and m 2 by Dempster’s rule yields a Bayesian mass function. Its corresponding contour function can be defined by [41]:
( p 1 p l 2 ) ( ω ) = 1 1 k p 1 ( ω ) p l 2 ( ω ) ,
where k is the conflict between m 1 and m 2 . It can be written as:
k = 1 w Ω p 1 ( ω ) p l 1 ( ω ) .
The item w Ω p 1 ( ω ) p l 1 ( ω ) can be regarded as the mathematical expectation of p l 2 with respect to p 1 . If  m 2 is categorical and such that m 2 ( A ) = 1 , then p 1 p l 2 is the probability distribution by conditioning p 1 with respect to A.
Let m X and m Y be two mass functions defined on Ω X and Ω Y , respectively, and P l X and P l Y are the associated plausibility functions. P l X Y is the plausibility function defined on the product frame Ω X × Ω Y . Variables X and Y are called cognitively independent if:
P l X Y ( A × B ) = P l X ( A ) P l Y ( B ) , A Ω X , B Ω Y .
If two variables are cognitively independent, the evidence on one variable does not affect the beliefs on other variables. It is clear that cognitive independence reduces to stochastic independence when m X and m Y are Bayesian.

2.3. Evidential Likelihood

Let X be a random vector with probability density function p X ( · ; θ ) , where θ is an unknown parameter. Assume that X is perfectly observed, and let x 0 be a realization of X. The complete likelihood function given x 0 would be defined as:
L ( θ ; x 0 ) = p ( x 0 ; θ ) , θ Θ .
If x is not precisely observed, but we only know that x A for A Ω X , where Ω X is the set from which X can take values. Then, the likelihood function given such imprecise data can be defined as [41]:   
L ( θ ; A ) = x A p ( x ; θ ) , θ Θ .
More generally, if the observation x is not only imprecise, but also uncertain, it can be described by a mass function m on Ω X with focal elements A 1 , , A r . The corresponding masses assigned to each focal element are m ( A 1 ) , , m ( A r ) . The likelihood function given such uncertain data can be extended as [41]:
L ( θ ; m ) = i = 1 r m ( A i ) L ( θ ; A i ) .
It can be rewritten as:
(19a) L ( θ ; m ) = i = 1 r m ( A i ) x A i p ( x ; θ ) (19b) = x Ω X p ( x ; θ ) A i x , A i Ω X m ( A i ) (19c) = x Ω X p ( x ; θ ) p l ( x ) (19d) = E θ [ p l ( X ) ] (19e) L ( θ ; p l ) .
From Equation (19c), we can see that L ( θ ; m ) depends on m through its associated contour function p l . As a result, L ( θ ; m ) can be denoted by L ( θ ; p l ) indifferently. We can see that L ( θ ; p l ) is the expectation of p l ( x ) with respect to p ( x ; θ ) . It is often called the evidential likelihood function [42].
The evidential likelihood function L ( θ ; p l ) can be seen as a special case of L ( θ ; A ) and L ( θ ; x ) . If  p l is associated with a categorical mass function m A , then:
p l ( x ) = 1 , x A Ω X , 0 , otherwise .
Equation (19c) equals the likelihood function given the imprecise data in Equation (17). If  p l is associated with a certain mass function:
p l ( x ) = 1 , X = x 0 , 0 , otherwise .
Equation (19c) degrades to the complete likelihood defined in Equation (16).

3. Maximum Evidential Likelihood Estimates for Uncertain Progressively Censored Data

The EM algorithm is usually adopted to obtain the maximal likelihood estimates (MLEs) for the datasets with missing information. However, it cannot deal with the partially labeled information included in the finite mixture model. In this section, we discuss the evidential likelihood for the uncertain progressively censored data from mixture distributions. The optimal statistical estimates that can be obtained by maximizing the evidential likelihood function (named MELEs) are derived, and the relation with the traditional EM model is also discussed.

3.1. Evidential Likelihood for Uncertain Progressively Censored Data

Assume that X 1 , X 2 , , X n are n independent variables that follow the mixture distribution in the form of Equation (1). Denote their probability density function by f ( x ; π , λ ) , the probability distribution function by F ( x ; π , λ ) and the survival function by s ( x ; π , λ ) . Suppose n units are placed on a life-test under the progressive censoring scheme.
In the Type-II progressive censoring scheme, we can denote the observed failure times by T i , i = 1 , 2 , , N , and use T = ( t 1 , t 2 , , t N ) to describe the completely observed data. At the time of the i t h observed failure, we remove R i ( i = 1 , 2 , , N ) items. Denote the failure times of these units by:
Z = { z i j , i = 1 , 2 , , N ; j = 1 , 2 , , R i } .
It is easy to know that Z cannot be observed in the experiment.
Let B i ( 1 ) = ( b i 1 ( 1 ) , b i 2 ( 1 ) , , b i m ( 1 ) ) . Each b i j ( 1 ) in B i ( 1 ) denotes whether the completely observed failure t i comes from the j t h component of the mixture model. If  T i f j ( x ; θ ) , b i j ( 1 ) = 1 . Otherwise, b i j ( 1 ) = 0 . Similarly, we can define the indicator for the unobserved data z i j . Let B i j ( 2 ) = ( b i j 1 ( 2 ) , b i j 2 ( 2 ) , , b i j m ( 2 ) ) , where  b i j k ( 2 ) represents whether the j t h removed unit at the time t i comes from the k t h component in the mixture model. Let B ( 1 ) = { B i ( 1 ) , i = 1 , 2 , , N } and B ( 2 ) = { B i j ( 2 ) , i = 1 , 2 , , N , j = 1 , 2 , , R i } .
The complete variables in this life-test can be denoted by W = ( X , B ) , where X = ( T , Z ) and B = ( B ( 1 ) , B ( 2 ) ) . It is clear that the observed data in this model are T, and the hidden variables are Z and B. Let θ = ( π , λ ) , then the probability function of W under the Type-II progressive censoring scheme is:
p ( w ; θ ) = i = 1 N k = 1 m π k f k ( t i ; λ k ) b i k ( 1 ) · j = 1 R i k = 1 m π k f k ( z i j ; λ k ) b i j k ( 2 ) .
The likelihood function based on the complete data is:
L ( θ ; w ) = p ( w ; θ ) ,
As mentioned, in this life-test, the observed data are T = { t i , i = 1 , 2 , , N } . The likelihood function based on the observed data is:
L ( θ ) = C i = 1 N [ f ( t i ; θ ) ( s ( t i ; θ ) ) R i ] ,
where C is a constant number. The maximum likelihood estimates can be obtained by maximizing Equation (22). The EM algorithm is usually adopted to get the MLEs for the progressively censored data [7]. As we can see, this likelihood function is based on the observed data, but it ignores some possible uncertain information of the partial labels. In fact, there are two kinds of uncertain information in this case. One is the failure time of the removed units during the experiment, for which we only know their censored time. The other is the sub-class label of each unit. Here, we use the mass functions in belief function theory to model these two types of uncertain information in a united frame. In order to establish a united frame for the data, we can describe all the data obtained in the life-test using different kinds of mass functions.
If X is a completely observed data and we assume that X = t * , we can model this information with a certain mass function. Its contour function can be defined as:
p l 1 ( X ) ( x ) = 1 , x = t * , 0 , otherwise .
If X is the censored data at time t * , we only know that x [ t * , + ) . We can model this uncertain information using the following contour function:
p l 2 ( X ) ( x ) = 1 , x [ t * , + ) , 0 , otherwise .
As mentioned before, in real applications, some uncertain information could be obtained from the experts or from historical data. The plausibility is adopted to express the uncertain information about the partial class labels of units. For the completely observed data t i , the plausibility about the proposition that the unit with failure t i comes from the k t h component is p l t i , k ( B ) . For the censored data z i j , we can use p l z i j , k ( B ) .
Suppose that the failure time X and the sub-class label B are cognitively independent. Therefore, the contour function of the complete data w i = ( x i , b i ) can be defined by:
p l ( w i ) = k = 1 m p l x i , k ( B ) b i k · p l ( X ) ( x i ) .
where p l x i , k ( B ) is the plausibility that x i comes from the k t h sub-group and  p l ( X ) ( x i ) describes the information about the failures. For the completely observed data and censored data, p l ( X ) ( x i ) can be defined as Equations (23) and (24), respectively.
Considering the progressive censoring scheme, given the data w = ( x , b ) , the contour function of x can be defined as:
(26a) p l ( w ) = i = 1 n p l ( X ) ( x i ) p l ( B ) ( b i ) (26b) = i = 1 n k = 1 m p l x i , k ( B ) b i k p ( X ) ( x i ) (26c) = i = 1 N k = 1 m p l t i , k ( B ) b i k ( 1 ) p l 1 ( X ) ( t i ) · j = 1 R i k = 1 m p l z i j , k ( B ) b i j k ( 2 ) p l 2 ( X ) ( z i j ) .
The evidential likelihood for the progressively censored data can be derived as:
(27a) L ( θ ; p l ) = i = 1 n E α [ p l ( W i ) ] (27b) = i = 1 n E α [ p l ( X ) ( X i ) p l ( B ) ( B i ) ] (27c) = i = 1 n k = 1 m π k E α [ p l ( X ) ( X i ) p l ( B ) ( B i ) | b i k = 1 ] (27d) = i = 1 n k = 1 m π k p l x i , k ( B ) E α [ p l ( X ) ( X i ) | b i k = 1 ] = i = 1 N k = 1 m π k p l t i , k ( B ) E α [ p l 1 ( X ) ( t i ) | b i k ( 1 ) = 1 ] (27e) · i = 1 N j = 1 R i k = 1 m π k p l z i j , k ( B ) E α [ p l 2 ( X ) ( z i j ) | b i j k ( 2 ) = 1 ] (27f) = i = 1 N k = 1 m π k p l t i , k ( B ) f k ( t i ; λ k ) · j = 1 R i k = 1 m π k p l z i j , k ( B ) s k ( t i ; λ k ) ,
where θ = ( π , λ ) denotes the unknown parameters.
Corollary 1.
If there is no information about the sub-class label of each unit, we can use the vacuous mass function on Ω B , i.e., p l x i , k ( B ) = 1 . Therefore, for the observed data and the censored data, we can get:
p l t i , k ( B ) = 1 , i = 1 , 2 , , N ; k = 1 , , m , p l z i j , k ( B ) = 1 , i = 1 , 2 , , N ; j = 1 , 2 , , R i ; k = 1 , , m .
In this case, the evidential likelihood function defined in Equation (27f) can be rewritten as:
L ( θ ; p l ) = i = 1 N k = 1 m π k f k ( t i ; λ k ) · k = 1 m π k s k ( t i ; λ k ) R i .
This is the same as Equation (22).
From Corollary 1, we can see that the traditional likelihood can be regarded as a special case of the evidential likelihood for the progressively censored data when there is no prior information about the sub-class label of each unit.

3.2. The Optimal Estimates under the Evidential Likelihood

The E2M algorithm can be evoked to derive the MELEs for the uncertain progressively censored data from mixture models. Given the initial parameter value of θ = θ ( 0 ) , the E-step and M-step are repeated alternatively.
  • E-step: Derive the Q function, which can be seen as the expectation of l o g L ( θ ; w ) with respect to p ( w | p l ; θ ( q ) ) .
    From Equation (13), the probability mass function p ( · | p l ; θ ( q ) ) can be expressed as:
    p ( w | p l ; θ ( q ) ) = p 1 ( w ; θ ( q ) ) p l 2 ( w ) w Ω p 1 ( w ; θ ( q ) ) p l 2 ( w ) = p ( w ; θ ( q ) ) p l ( w ) L ( θ ( q ) ; p l ) .
    The probability function p ( · | p l ; θ ( q ) ) can be seen as the Dempster combination of p ( w ; θ ( q ) ) and p l ( w ) . The former describes the random uncertainty due to the underlying population, while the latter reflects the epistemic uncertainty due to the partial knowledge.
    According to Equation (30), the Q function at the q t h iteration is:
    (31a) Q ( θ , θ ( q ) ) = E [ l o g L ( θ ; w ) | p l ; θ ( q ) ] (31a) = w Ω W l o g ( L ( θ ; w ) ) p ( w ; θ ( q ) ) p l ( w ) L ( θ ( q ) ; p l ) ,
    where L ( θ ; w ) is the likelihood function based on the complete data (see Equation (21)).
  • M-step: This is the same as the traditional EM algorithm. In this step, the value of θ can be obtained by maximizing Q ( θ , θ ( q ) ) .
Theorem 1.
L ( θ ( q ) ; p l ) , q = 0 , 1 , 2 , of the evidential likelihood function obtained by the above-mentioned E2M algorithm is not decreasing, i.e., it verifies:
L ( θ ( q + 1 ) ; p l ) L ( θ ( q ) ; p l ) .
Proof. 
From Equation (30), we know:   
p ( w | p l ; θ ) = p ( w ; θ ) p l ( w ) L ( θ ; p l ) = L ( θ ; w ) p l ( w ) L ( θ ; p l ) .
From the above equation, we can get:
L ( θ ; w ) L ( θ ; p l ) = p ( w | p l ; θ ) p l ( w ) g ( w | p ; θ ) .
Then, we can have:
log L ( θ ; p l ) = log L ( θ ; w ) log g ( w | p l ; θ ) .
Taking the expectation of both sides of the above equation with respect to p ( w | p l ; θ ( q ) ) , then it is easy to get:
log L ( θ ; p l ) = E log L ( θ ; w ) | p l ; θ ( q ) E log g ( x | p l ; θ ) | p l ; θ ( q ) .
Let
H ( θ , θ ( q ) ) = E log g ( x | p l ; θ ) | p l ; θ ( q ) ,
and then, we have:
log L ( θ ( q + 1 ) ; p l ) log L ( θ ( q ) ; p l ) = Q ( θ ( q + 1 ) , θ ( q ) ) Q ( θ ( q ) , θ ( q ) ) H ( θ ( q + 1 ) , θ ( q ) ) H ( θ ( q ) , θ ( q ) ) .
It is easy to see that Q ( θ ( q + 1 ) , θ ( q ) ) Q ( θ ( q ) , θ ( q ) ) since in the M-step, the value of θ is updated by maximizing Q ( θ , θ ( q ) ) with respect to θ. For any θ,
H ( θ , θ ( q ) ) H ( θ ( q ) , θ ( q ) ) (39a) = E log g ( w | p l ; θ ) g ( w | p l ; θ ( q ) ) | p l ; θ ( q ) (39b) log E g ( w | p l ; θ ) g ( w | p l ; θ ( q ) ) | p l ; θ ( q ) (39c) log g ( w | p l ; θ ) g ( w | p l ; θ ( q ) ) p l ( w ) g ( w | p l ; θ ( q ) ) d w (39d) log g ( w | p l ; θ ) p l ( w ) d w (39e) log p ( w | p l ; θ ) d w = 0 ,
where the inequality (39b) is obtained from Jensen’s inequality.
As we can see, the first difference in Equation (38) is nonnegative, while the second one is no larger than zero. Thus, it is easy to get log L ( θ ( q + 1 ) ; p l ) log L ( θ ( q ) ; p l ) , and consequently, we have the inequality (32). □
Corollary 2.
When the contour function p l ( w ) degrades to the categorical mass function, the probability function p ( w | p l ; θ ( q ) ) degrades to the conditional probability function P ( · | y , θ ( q ) ) in EM.

3.3. The Estimation Methodology for a Mixed Exponential Distribution

The likelihood function based on the complete data w for Type-II progressively censored data from MED is:
L ( π , λ ; w ) = i = 1 N k = 1 m π k λ k e λ k t i b i k ( 1 ) · j = 1 R i k = 1 m π k λ k e λ k z i j b i j k ( 2 ) .
The log-likelihood is:
l o g L ( π , λ ; w ) = i = 1 N k = 1 m b i k ( 1 ) l o g π k + l o g λ k λ k t i + j = 1 R i k = 1 m b i j k ( 2 ) l o g π k + l o g λ k λ k z i j .
As before, let θ = ( π , λ ) denote the unknown parameters of the MED model. Given the initial parameter θ ( 0 ) = ( π ( 0 ) , λ ( 0 ) ) , the alternative iterations of the E2M algorithm are described in the following.
  • E-step: Let the parameters at the q t h iteration be θ ( q ) = ( π ( q ) , λ ( q ) ) . The expectation of l o g L ( π , λ ; w ) with respect to p ( w | p l ; θ ( q ) ) (Q function) can be defined as:
    (42a) Q ( θ , θ ( q ) ) = E [ l o g L ( π , λ ; w ) | p l ; θ ( q ) ] = i = 1 N { k = 1 m ( l o g π k + l o g λ k λ k t i ) E [ b i k ( 1 ) | p l ; θ ( q ) ] + (42b) j = 1 R i k = 1 m ( l o g π k + l o g λ k ) E [ b i j k ( 2 ) | p l ; θ ( q ) ] λ k E [ b i j k ( 2 ) z i j | p l ; θ ( q ) ] } .
    From Equation (13), we known:
    (43a) p ( x i , b i k ( 1 ) = 1 | p l ; θ ( q ) ) = π k ( q ) f k t i ; λ k ( q ) p l t i , k ( B ) p l ( X ) t i l = 1 m π l ( q ) x i f l x i ; λ k ( q ) p l x i , l ( B ) p l ( X ) x i d x i (43b) = π k ( q ) f k t i ; λ k ( q ) p l t i , k ( B ) p l ( X ) t i l = 1 m π l ( q ) f l t i ; λ k ( q ) p l t i , l ( B ) .
    As b i k ( 1 ) is a Boolean variable, it is easy to get:
    E [ b i k ( 1 ) | p l ; θ ( q ) ] = P ( b i k ( 1 ) = 1 | p l ; θ ( q ) ) = π k ( q ) f k t i ; λ k ( q ) p l t i , k ( B ) l = 1 m π l ( q ) f l t i ; λ k ( q ) p l t i , l ( B ) a i k ( q ) .
    Similarity, we can get:
    p ( z i j , b i j k ( 2 ) = 1 | p l ; θ ( q ) ) = π k ( q ) f k z i j ; λ k ( q ) p l z i j , k ( B ) p l ( X ) z i j l = 1 m π l ( q ) s l z i j ; λ k ( q ) p l z i j , l ( B ) .
    E [ b i j k ( 2 ) | p l ; θ ( q ) ] = P ( b i j k ( 2 ) = 1 | p l ; θ ( q ) ) = π k ( q ) s k t i ; λ k ( q ) p l z i j , k ( B ) l = 1 m π l ( q ) s l z i j ; λ l ( q ) p l z i j , l ( B ) e i j k ( q ) .
    For the expectation of b i j k ( 2 ) z i j , we can get:
    E [ b i j k ( 2 ) z i j | p l ; θ ( q ) ] = E [ z i j | b i j k ( 2 ) = 1 , p l ; θ ( q ) ] P ( b i j k ( 2 ) = 1 | p l ; θ ( q ) ) ,
    where P ( b i j k ( 2 ) = 1 | p l ; θ ( q ) ) can be obtained by Equation (46). From Equations (45) and (46), we can get:   
    p ( z i j | b i j k ( 2 ) = 1 , p l ; θ ( q ) ) = p ( z i j , b i j k ( 2 ) = 1 | p l ; θ ( q ) ) P ( b i j k ( 2 ) = 1 | p l ; θ ( q ) ) = f k ( z i j ; λ k ( q ) ) p l ( X ) ( z i j ) s k ( t i ; λ k ( q ) ) .
    As z i j [ t i , + ) , we can get:
    E [ z i j | b i j k ( 2 ) = 1 , p l ; θ ( q ) ] = t i z i j f k ( z i j ; λ k ( q ) ) s k ( t i ; λ k ( q ) ) d z i j ξ i k ( q ) .
    Therefore,
    E [ b i j k ( 2 ) z i j | p l ; θ ( q ) ] = ξ i k ( q ) e i j k ( q ) .
    Substituting Equations (44), (46), and (50) into Equation (42b), we can get the Q function in the q t h iteration:
    Q ( θ , θ ( q ) ) = i = 1 N k = 1 m ( l o g π k + l o g λ k λ k t i ) a i k ( q ) + j = 1 R i k = 1 m ( l o g π k + l o g λ k ) e i j k ( q ) λ k ξ i k ( q ) e i j k ( q ) ,
    where θ = ( π , λ ) , π = ( π 1 , , π m ) , λ = ( λ 1 , , λ m ) , and  k = 1 m π k = 1 .
Corollary 3.
If there is no uncertain information in the experiment, i.e., p l x i , k = 1 , the conditional expectations defined in Equations (44), (46), and (50) are the same as those in the conventional EM algorithm:
E [ b i k ( 1 ) | p l ; θ ( q ) ] = π k ( q ) f k t i ; λ k ( q ) l = 1 m π l ( q ) f l t i ; λ l ( q ) , E [ b i j k ( 2 ) | p l ; θ ( q ) ] = π k ( q ) s k t i ; λ k ( q ) l = 1 m π l ( q ) s l t i ; λ l ( q ) , E [ b i j k ( 2 ) z i j | p l ; θ ( q ) ] = π k ( q ) t i z i j f k ( z i j ; λ k ( q ) ) d z i j l = 1 m π l ( q ) s l t i ; λ l ( q ) .
  • M-step: In order to get the updated estimates, we have to maximize the Q ( θ , θ ( q ) ) function with respect to parameters π and λ . The Lagrangian method can be adopted to consider the constraint k = 1 m π k = 1 . Let:
    L ( θ , η ) = Q ( θ , θ ( q ) ) + η ( k = 1 m π k 1 ) .
    We can get:
    L ( θ , η ) π k = A k ( q ) π k + η = 0 ,
    L ( θ , η ) λ k = A k ( q ) λ k C k ( q ) i = 1 N t i a i k ( q ) = 0 ,
    L ( θ , η ) η = k = 1 m π k 1 = 0 ,
    where
    A k ( q ) = i = 1 N a i k ( q ) + i = 1 N j = 1 R i e i j k ( q ) ,
    C k ( q ) = i = 1 N j = 1 R i ξ i k ( q ) e i j k ( q ) .
    From Equations (52) and (54), we know:
    η = k = 1 m A k ( q ) .
    Thus, it is easy to get:
    π k ( q + 1 ) = A k ( q ) k = 1 m A k ( q ) , k = 1 , , m ,
    λ k ( q + 1 ) = A k ( q ) C k ( q ) + i = 1 N t i a i k ( q ) , k = 1 , , m .
    where A k ( q ) and C k ( q ) are defined in Equations (55) and (56), respectively.
The E-step and M-step will be repeated iteratively until:
i = 1 m π i ( q + 1 ) π i ( q ) 2 ϵ
and:
i = 1 m λ i ( q + 1 ) λ i ( q ) 2 ϵ ,
where ϵ is a given small positive constant. Suppose that the algorithm stops at the l t h step, then θ ( l ) = ( π ( l ) , λ ( l ) ) are the MELEs for the parameters. The whole process of the estimation is summarized in Algorithm 1.
Algorithm 1: The evidential estimation approach for uncertain MED under progressive censoring.
  • Input: The observed failures ( t 1 , , t N ) and the censoring scheme ( R 1 , , R N ) .   
  • Initialization:
      (1)
    Let q = 0
      (2)
    Initialize the parameters θ ( q ) = π ( q ) , λ ( q )
  • repeat
      (1)
    E-step: Calculate the Q function based on the current parameters Q ( θ , θ ( q ) ) .
      (2)
    M-step: Maximize Q ( θ , θ ( q ) ) with respect to θ and update the parameters θ ( q + 1 ) .
      (3)
    q = q + 1
  • until Both Equations (59) and (60) hold true.   
  • Output: The MELEs of the parameters.

4. Data Analysis

In this section, a real dataset of the air conditioning system of an airplane is considered to illustrate the practical application of the proposed evidential estimation model. The group of failure times of the system is given as follows:
1 , 3 , 5 , 7 , 11 , 11 , 11 , 12 , 14 , 14 , 14 , 16 , 16 , 20 , 21 , 23 , 42 , 47 , 52 , 62 , 71 , 71 , 87 , 90 , 95 , 120 , 120 , 225 , 246 , 261 .
This dataset was analyzed by different authors using various models [43,44]. In [19], the authors found that the MED model with two components ( m = 2 ) provided a good fit to the given failure data. The fitted probability density function can be obtained by the maximal likelihood estimations:
f ^ ( x ) = π ^ 1 λ ^ 1 e λ ^ 1 x + π ^ 2 λ ^ 2 e λ ^ 2 x , x 0 ,
with π ^ 1 = 0.3459 , π ^ 2 = 0.6541 , λ ^ 1 = 0.0647 , λ ^ 2 = 0.0121 . If we do not have any information about the sub-class labels of each sample, the MLEs of the above MED can be obtained using the EM algorithm [19]:
π ^ 1 MLE = 0.1136 , π ^ 2 MLE = 0.8864 , λ ^ 1 MLE = 0.1091 , λ ^ 2 MLE = 0.0161 .
Suppose we have some uncertain information about the sub-class labels. This type of knowledge can be obtained by experts or from historical data. Here, we can use the evidential c-means (ECM) clustering algorithm [31] to obtain the contour function for each sample. From the given failures, it is easy to see that they can be divided into two groups. The credal partition found by ECM can provide us with the plausibility of the data belonging to the sub-groups. In ECM, let C = 2 , and the other parameters are set as the default. The plausibility matrix describing the partial information about the sub-class label of each sample is illustrated in Table 1.
The maximum likelihood estimations in the sense of evidential likelihood (denoted by MELE) that can integrate the uncertain label information can be obtained:
π ^ 1 MELE = 0.3433 , π ^ 2 MELE = 0.6567 , λ ^ 1 MELE = 0.0651 , λ ^ 2 MELE = 0.0121 .
We can see that the estimation results by the proposed estimation model are closer to the estimations by the complete data.
As we know, in the traditional EM algorithm, the solution depends on the initial values adopted. In order to show the influences of the initial values for the proposed estimation model, we will try with different initialization strategies. For parameter π , we evoke the estimation model with:
π 1 ( 0 ) = | 0.5 + N ( 0 , 0.1 ) | , π 2 ( 0 ) = 1 π 1 ( 0 ) ,
while for λ , we set:
λ 1 ( 0 ) = | 0.2 + N ( 0 , 0.1 ) | , λ 2 ( 0 ) = | 0.8 + N ( 0 , 0.1 ) | ,
where | x | is the absolute value of x and N ( 0 , 0.1 ) denotes a random number from the normal distribution with μ = 0 , σ 2 = 0.1 . The E2M algorithm and EM algorithm are both repeated 100 times. The mean value and the standard variances of the estimations are reported, which is illustrated in Table 2.
It is easy to see that the standard variation values of MELEs are smaller than those of MLEs. We can conclude that the use of the uncertain information can reduce the influence of the initializations in the estimation model. Many initialization methods have been studied for the EM algorithm, which is worth learning and a reference for the proposed estimation model. However, in this paper, we do not focus on the proper initialization method and leave it for our future research.

5. Simulations

Monte Carlo simulations are conducted in this section to show the performance of the proposed method. Suppose that X 1 , X 2 , , X n are n independent variables from the mixed exponential distribution in Equation (2). Here, we consider the model with two mixture components ( m = 2 ) under the Type-II progressive censoring scheme. The parameters to be estimated are π 1 , π 2 , λ 1 , λ 2 with π 1 + π 2 = 1 .
To simulate uncertainty on class labels, the original generated data are processed as follows. For the i t h observed failure, an error probability q i is drawn randomly from a beta distribution with mean ρ and standard deviation σ . In real applications, q i can be obtained by experts or from historical data. With probability q i , the label of z i is replaced by a completely value of z ˜ i (with a uniform distribution over all the class labels). The contour function on class labels can be defined as:
p l x i , k ( B ) = P ( b i k = 1 | b i k * ) = q i m , b i k * = 0 , q i m + 1 q i , b i k * = 1 , i = 1 , , n ; k = 1 , , m .
If there is no uncertainty, i.e., p l x i , k = 1 , the evidential likelihood degrades to the traditional likelihood.
Let π 1 = 0.3 , π 2 = 0.7 ; λ 1 = 0.6 , λ 2 = 0.1 . Set the initial value of the parameters as:
π 1 ( 0 ) = 0.5 , π 2 ( 0 ) = 0.5 , λ 1 ( 0 ) = 1 , λ 2 ( 0 ) = 0.5 .
In this experiment, we set the number of units placed on the life-test to be n = 20 , 100 , 200 , 400 . We consider different censoring proportions for a given sample size. Let N = n × ( 0.1 , 0.2 , 0.3 ) . For the parameters to model the uncertain label information, we set ρ = 0.2 , σ = 0.2 . Three censoring schemes are considered:
Scheme   1 :   R 1 = , R N 1 = 0 , R N = n N . Scheme   2 :   R 1 = n N , R 2 = , R N = 0 . Scheme   3 :   R 1 = R 2 = = R N 1 = 1 , R N = n 2 N + 1 .
Given the sample size n and the censoring scheme, the E2M algorithm will be repeated s = 100 times. The bias and mean squared error (MSE) values of the estimators are reported. The results are illustrated in Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8.
From the tables, we can see that, with the increasing size of n and N, for the estimators obtained both by the proposed evidential estimation model and by the EM algorithm, the values of the bias and MSE decrease in most cases. This is consistent with our common sense about the maximum likelihood estimations. For different censoring schemes, it is obvious that the estimation efficiency is highly related to the proportion of the censored samples. When the number of units placed in the life-test is fixed, with the increasing number of censored data, the values of MSE increase.
Overall, in most cases, the biases and MSEs of the MELEs by the proposed evidential estimation model are smaller than those of MLEs by the EM algorithm, which can be attributed to the use of the uncertain prior information about the sub-class labels. When the proportion of the censored data is large, the MELEs obtained by the proposed estimation model are significantly better.

6. Conclusions

In this paper, a method for estimating the parameters of a mixed distribution model under Type-II progressive censoring was introduced. The main advantage of the proposed formalism is that it can combine uncertainty captured by the imperfect observation process due to censoring and the partial knowledge about the sub-class information in the mixture model. These two types of uncertain information can be represented in a united frame by the use of belief functions, and then, the evidential likelihood is derived, which can be seen as a generalized likelihood criterion. It can be maximized using the evidential EM algorithm. As an illustration, the method was applied to the MED model. A real dataset was considered, and some Monte Carlo simulations were carried out to demonstrate the performance of the resulting estimates. The results show that the estimations by the proposed method gain better estimation efficiency, which can be attributed to the effective integration of the uncertain information in the model.
More generally, the evidential estimation approach introduced in this paper is applicable to any other mixed distributions under different censoring schemes, where data uncertainty in the parametric statistical model arises from the imperfect observation process and reliability engineering applications such as the life data with multiple failure causes.

Author Contributions

The authors contributed equally to the presented mathematical framework, the programming and the writing of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No.61701409), the Natural Science Basic Research Plan in Shaanxi Province of China (No.2018JQ6005), the Aero Science Foundation of China (No.20182053023), and the Fundamental Research Funds for the Central Universities of China (No.310201911cx041).

Acknowledgments

The authors would like to express deep thanks to the editors and referees for their helpful comments and suggestions which significantly improved the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Figueiredo, M.A.T.; Jain, A.K. Unsupervised learning of finite mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 381–396. [Google Scholar] [CrossRef] [Green Version]
  2. Xiang, S.; Yao, W.; Yang, G. An overview of semiparametric extensions of finite mixture models. Stat. Sci. 2019, 34, 391–404. [Google Scholar] [CrossRef] [Green Version]
  3. He, L.; Lu, Z.; Li, X. Failure-mode importance measures in structural system with multiple failure modes and its estimation using copula. Reliab. Eng. Syst. Saf. 2018, 174, 53–59. [Google Scholar] [CrossRef]
  4. Liao, Y.; Xiang, Y.; Keedy, E. Reliability analysis of flow meters with multiple failure modes in the process industry. In Proceedings of the 2020 Annual Reliability and Maintainability Symposium (RAMS), Palm Springs, CA, USA, 27–30 January 2020; pp. 1–6. [Google Scholar]
  5. Lee, Y. Expectation-Maximization Algorithms for Learning a Finite Mixture of Univariate Survival Time Distributions from Partially Specified Class Values. Ph.D. Thesis, Iowa State University, Ames, IA, USA, 2013. [Google Scholar]
  6. Balakrishnan, N.; Aggarwala, R. Progressive Censoring: Theory, Methods, and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  7. Ng, H.K.T.; Chan, P.S.; Balakrishnan, N. Estimation of parameters from progressively censored data using EM algorithm. Comput. Stat. Data Anal. 2002, 39, 371–386. [Google Scholar] [CrossRef]
  8. Balakrishnan, N.; Kannan, N.; Lin, C.T.; Ng, H.T. Point and interval estimation for Gaussian distribution, based on progressively Type-II censored samples. IEEE Trans. Reliab. 2003, 52, 90–95. [Google Scholar] [CrossRef]
  9. Basak, I.; Balakrishnan, N. Robust estimation under progressive censoring. Comput. Stat. Data Anal. 2003, 44, 349–376. [Google Scholar] [CrossRef]
  10. Abo-Eleneen, Z.A. The entropy of progressively censored samples. Entropy 2011, 13, 437–449. [Google Scholar] [CrossRef] [Green Version]
  11. Kotb, M.S.; Raqab, M.Z. Statistical inference for modified Weibull distribution based on progressively type-II censored data. Math. Comput. Simul. 2019, 162, 233–248. [Google Scholar] [CrossRef]
  12. Bai, X.; Shi, Y.; Liu, Y.; Liu, B. Reliability inference of stress–strength model for the truncated proportional hazard rate distribution under progressively Type-II censored samples. Appl. Math. Model. 2019, 65, 377–389. [Google Scholar] [CrossRef]
  13. Zhang, Z.; Gui, W. Statistical inference of reliability of Generalized Rayleigh distribution under progressively type-II censoring. J. Comput. Appl. Math. 2019, 361, 295–312. [Google Scholar] [CrossRef]
  14. Feroze, N.; Aslam, M.; Khan, I.H.; Khan, M.H. Bayesian reliability estimation for the Topp–Leone distribution under progressively type-II censored samples. Soft Comput. 2020, 1–22. [Google Scholar] [CrossRef]
  15. Cho, Y.; Sun, H.; Lee, K. An estimation of the entropy for a Rayleigh distribution based on doubly-generalized Type-II hybrid censored samples. Entropy 2014, 16, 3655–3669. [Google Scholar] [CrossRef] [Green Version]
  16. Algarni, A.; Almarashi, A.M.; Okasha, H.; Ng, H.K.T. E-Bayesian Estimation of Chen Distribution Based on Type-I Censoring Scheme. Entropy 2020, 22, 636. [Google Scholar] [CrossRef]
  17. Cho, Y.; Sun, H.; Lee, K. Estimating the entropy of a Weibull distribution under generalized progressive hybrid censoring. Entropy 2015, 17, 102–122. [Google Scholar] [CrossRef] [Green Version]
  18. Lee, G.; Scott, C. EM algorithms for multivariate Gaussian mixture models with truncated and censored data. Comput. Stat. Data Anal. 2012, 56, 2816–2829. [Google Scholar] [CrossRef]
  19. Tian, Y.; Tian, M.; Zhu, Q. Estimating a finite mixed exponential distribution under progressively type-II censored data. Commun. Stat. Theory Methods 2013, 43, 3762–3776. [Google Scholar] [CrossRef]
  20. Tian, Y.; Zhu, Q.; Tian, M. Estimation for mixed exponential distributions under type-II progressively hybrid censored samples. Comput. Stat. Data Anal. 2015, 89, 85–96. [Google Scholar] [CrossRef]
  21. Tahir, M.; Aslam, M.; Hussain, Z.; Abbas, N. On the finite mixture of exponential, Rayleigh and Burr Type-XII distributions: Estimation of parameters in Bayesian framework. Electron. J. Appl. Stat. Anal. 2017, 10, 271–293. [Google Scholar]
  22. Abushal, T.A.; AL-Zaydi, A.M. Inference on Constant-Partially Accelerated Life Tests for Mixture of Pareto Distributions under Progressive Type-II Censoring. Open J. Stat. 2017, 7, 323. [Google Scholar] [CrossRef] [Green Version]
  23. Feroze, N.; Aslam, M. Approximate Bayesian analysis of doubly censored samples from mixture of two Weibull distributions. Commun. Stat. Theory Methods 2019, 48, 2862–2878. [Google Scholar] [CrossRef]
  24. Feroze, N.; Aslam, M. Three-Component Mixture of Rayleigh Model Under Doubly Censored Samples: A Bayesian Look. Commun. Math. Stat. 2019, 7, 417–443. [Google Scholar] [CrossRef]
  25. Zio, E. Reliability engineering: Old problems and new challenges. Reliab. Eng. Syst. Saf. 2009, 94, 125–141. [Google Scholar] [CrossRef] [Green Version]
  26. Deng, Y. Deng entropy. Chaos Solitons Fractals 2016, 91, 549–553. [Google Scholar] [CrossRef]
  27. Tang, Y.; Zhou, D.; Xu, S.; He, Z. A weighted belief entropy-based uncertainty measure for multi-sensor data fusion. Sensors 2017, 17, 928. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Huang, Z.; Yang, L.; Jiang, W. Uncertainty measurement with belief entropy on the interference effect in the quantum-like Bayesian networks. Appl. Math. Comput. 2019, 347, 417–428. [Google Scholar] [CrossRef] [Green Version]
  29. Wu, D.; Tang, Y. An improved failure mode and effects analysis method based on uncertainty measure in the evidence theory. Qual. Reliab. Eng. Int. 2020. [Google Scholar] [CrossRef]
  30. Liu, Z.g.; Liu, Y.; Dezert, J.; Cuzzolin, F. Evidence combination based on credal belief redistribution for pattern classification. IEEE Trans. Fuzzy Syst. 2019, 28, 618–631. [Google Scholar] [CrossRef] [Green Version]
  31. Masson, M.H.; Denoeux, T. ECM: An evidential version of the fuzzy c-means algorithm. Pattern Recognit. 2008, 41, 1384–1397. [Google Scholar] [CrossRef]
  32. Pan, L.; Deng, Y. An association coefficient of a belief function and its application in a target recognition system. Int. J. Intell. Syst. 2020, 35, 85–104. [Google Scholar] [CrossRef]
  33. Qin, M.; Tang, Y.; Wen, J. An Improved Total Uncertainty Measure in the Evidence Theory and Its Application in Decision Making. Entropy 2020, 22, 487. [Google Scholar] [CrossRef]
  34. Zhang, H.; Deng, Y. Weighted belief function of sensor data fusion in engine fault diagnosis. Soft Comput. 2020, 24, 2329–2339. [Google Scholar] [CrossRef]
  35. Xiao, F.; Cao, Z.; Jolfaei, A. A novel conflict measurement in decision making and its application in fault diagnosis. IEEE Trans. Fuzzy Syst. 2020. [Google Scholar] [CrossRef]
  36. Zhou, K.; Martin, A.; Pan, Q.; Liu, Z. Median evidential c-means algorithm and its application to community detection. Knowl. Based Syst. 2015, 74, 69–88. [Google Scholar] [CrossRef] [Green Version]
  37. Lepskiy, A.; Meshcheryakova, N. Belief functions for the importance assessment in multiplex networks. In Proceedings of the International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems 2020, Lisbon, Portugal, 15–19 June 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 283–296. [Google Scholar]
  38. Jaeger, M. The AI&M procedure for learning from incomplete data. In Proceedings of the Twenty-Second Conference on Uncertainty in Artificial Intelligence, Cambridge, MA, USA, 13–16 July 2006; pp. 225–232. [Google Scholar]
  39. Couso, I.; Dubois, D. A general framework for maximizing likelihood under incomplete data. Int. J. Approx. Reason. 2018, 93, 238–260. [Google Scholar] [CrossRef] [Green Version]
  40. Zhang, M.; Ye, Z.; Xie, M. A stochastic em algorithm for progressively censored data analysis. Qual. Reliab. Eng. Int. 2014, 30, 711–722. [Google Scholar] [CrossRef]
  41. Denoeux, T. Maximum likelihood estimation from uncertain data in the belief function framework. Knowl. Data Eng. IEEE Trans. 2013, 25, 119–130. [Google Scholar] [CrossRef] [Green Version]
  42. Denoeux, T.; Kanjanatarakul, O.; Sriboonchitta, S. A new evidential k-nearest neighbor rule based on contextual discounting with partially supervised learning. Int. J. Approx. Reason. 2019, 113, 287–302. [Google Scholar] [CrossRef] [Green Version]
  43. Gupta, R.D.; Kundu, D. Exponentiated exponential family: An alternative to gamma and Weibull distributions. Biom. J. J. Math. Methods Biosci. 2001, 43, 117–130. [Google Scholar] [CrossRef]
  44. Mokhtari, E.B.; Rad, A.H.; Yousefzadeh, F. Inference for Weibull distribution based on progressively Type-II hybrid censored data. J. Stat. Plan. Inference 2011, 141, 2824–2838. [Google Scholar] [CrossRef]
Table 1. The plausibility matrix for the data.
Table 1. The plausibility matrix for the data.
ID pl x i , 1 ( B ) pl x i , 2 ( B ) ID pl x i , 1 ( B ) pl x i , 2 ( B )
10.99710.0080160.99760.0076
20.99800.0057170.97750.0858
30.99870.0038180.96870.1264
40.99920.0022190.95900.1770
50.99990.0003200.93880.3094
60.99990.0003210.92470.4584
70.99990.0003220.92470.4584
81.00000.0001230.92660.7394
91.00000.0001240.93190.7860
101.00000.0001250.94350.8549
111.00000.0001261.00001.0000
120.99980.0006271.00001.0000
130.99980.0006280.00011.0000
140.99890.0035290.01890.9930
150.99850.0047300.04740.9813
Table 2. The mean and standard variance (sd) values of the estimations with different initializations. MELE, maximal evidential likelihood estimate.
Table 2. The mean and standard variance (sd) values of the estimations with different initializations. MELE, maximal evidential likelihood estimate.
MLEMELE
ParameterMeansdMeansd
λ 1 0.08980.01010.05980.0083
λ 2 0.01770.03050.02770.0045
π 1 0.19310.10940.09310.0182
π 2 0.80690.10940.40690.0154
Table 3. The statistical properties of the estimators of λ 1 and λ 2 with Scheme 1. λ ^ i MELE and λ ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
Table 3. The statistical properties of the estimators of λ 1 and λ 2 with Scheme 1. λ ^ i MELE and λ ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
BiasMSE
n N λ ^ 1 MLE λ ^ 1 MELE λ ^ 2 MLE λ ^ 2 MELE λ ^ 1 MLE λ ^ 1 MELE λ ^ 2 MLE λ ^ 2 MELE
2020.39400.22790.31930.14215.48903.63820.09900.0808
2040.93570.77830.25080.13734.43762.62570.09330.0764
2060.12811.01120.19340.11895.94672.63580.08580.0683
100100.76660.26610.28250.12493.10090.07900.09700.0642
100200.40590.21580.25040.15091.69190.33330.08230.0652
100300.26010.14220.18600.14270.63540.17280.08650.0657
200200.61750.29110.27750.11677.97790.77490.10080.0645
200400.22380.24700.24510.16391.01830.51440.07470.0604
200600.41490.19860.19920.14543.97970.35260.08100.0665
400400.97310.30650.27390.12687.45990.06860.09360.0635
400800.72620.31630.24610.12429.79000.07160.08200.0540
4001200.24210.19780.20370.12190.37900.25340.07680.0455
Table 4. The statistical properties of the estimators of π 1 and π 2 with Scheme 1. π ^ i MELE and π ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
Table 4. The statistical properties of the estimators of π 1 and π 2 with Scheme 1. π ^ i MELE and π ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
BiasMSE
n N π ^ 1 MLE π ^ 1 MELE π ^ 2 MLE π ^ 2 MELE π ^ 1 MLE π ^ 1 MELE π ^ 2 MLE π ^ 2 MELE
2020.05080.08920.05790.11930.39630.29670.13140.0596
2040.12000.03920.06270.05080.16690.11400.12750.0752
2060.11780.07390.08110.00660.16440.13640.12720.0869
100100.13160.05890.07510.08250.15780.11160.13390.0546
100200.14710.08100.04910.05040.16430.12150.11220.0649
100300.16950.11460.05730.01400.17950.14160.11800.0861
200200.14200.07180.08330.07750.15960.11500.14020.0561
200400.15030.09170.03610.04510.16740.12810.10140.0626
200600.17440.11900.05460.00080.18200.14300.11210.0819
400400.15290.08010.07320.07390.16850.11890.13040.0563
400800.16290.09460.05580.04610.17510.12600.11400.0649
4001200.17990.11700.05810.00370.18600.14050.10970.0794
Table 5. The statistical properties of the estimators of λ 1 and λ 2 with Scheme 2. λ ^ i MELE and λ ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
Table 5. The statistical properties of the estimators of λ 1 and λ 2 with Scheme 2. λ ^ i MELE and λ ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
BiasMSE
n N λ ^ 1 MLE λ ^ 1 MELE λ ^ 2 MLE λ ^ 2 MELE λ ^ 1 MLE λ ^ 1 MELE λ ^ 2 MLE λ ^ 2 MELE
2029.164210.17500.19600.13544.85319.11050.09420.1050
20412.253713.25260.20830.16145.65195.05380.09050.0959
2064.00734.34860.25460.16501.31794.25560.13180.0993
1001012.99489.41790.17900.12727.88629.78240.09250.1078
100201.28181.38950.19910.125911.61203.54320.10770.1003
100300.48170.74000.22860.14771.95375.47080.10820.0890
200201.94400.97010.25300.14252.99151.56400.11630.1022
200400.37920.51350.22340.16971.67464.10210.08340.0781
200600.29991.21550.22860.17421.15391.50560.09620.0792
400402.63391.36800.19420.11981.18061.79620.08480.0982
400804.79860.60040.20310.13951.37830.54360.08400.0810
4001200.27020.09760.21990.14100.67020.49620.09240.0876
Table 6. The statistical properties of the estimators of π 1 and π 2 with Scheme 2. π ^ i MELE and π ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
Table 6. The statistical properties of the estimators of π 1 and π 2 with Scheme 2. π ^ i MELE and π ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
BiasMSE
n N π ^ 1 MLE π ^ 1 MELE π ^ 2 MLE π ^ 2 MELE π ^ 1 MLE π ^ 1 MELE π ^ 2 MLE π ^ 2 MELE
2020.03080.02050.06460.00400.24710.25760.13820.1103
2040.07650.06880.03860.00830.22120.18410.12150.1036
2060.17010.16130.03500.05460.17860.18670.12350.1256
100100.18300.17910.07280.02100.18310.18490.14250.1148
100200.19360.18780.07410.00090.19160.19720.13690.1215
100300.19680.18970.05230.02860.19150.19770.11860.1129
200200.19760.18320.05750.05300.18930.19990.13300.1140
200400.19740.19260.03030.02340.19410.19760.10280.0929
200600.19790.19340.02580.02860.19430.19860.10300.1023
400400.20110.19290.08020.00580.19450.20080.13650.1017
400800.19650.19050.06050.00310.19260.19780.11490.0983
4001200.20010.19310.05900.01990.19560.20200.11950.0981
Table 7. The statistical properties of the estimators of λ 1 and λ 2 with Scheme 3. λ ^ i MELE and λ ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
Table 7. The statistical properties of the estimators of λ 1 and λ 2 with Scheme 3. λ ^ i MELE and λ ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
BiasMSE
n N λ ^ 1 MLE λ ^ 1 MELE λ ^ 2 MLE λ ^ 2 MELE λ ^ 1 MLE λ ^ 1 MELE λ ^ 2 MLE λ ^ 2 MELE
2027.90400.17440.12780.19748.98190.21800.10600.0689
2043.01775.20170.13060.15567.63591.31410.10310.0694
2061.42180.57370.16030.19363.18241.14760.08720.0804
100101.73090.03560.12300.17504.45821.59490.09540.0651
100200.99200.10630.14100.13433.29221.30560.08910.0678
100300.51600.47880.12540.14701.65492.87390.08430.0732
200200.85110.29090.12850.17704.96670.07170.09020.0638
200400.43030.31900.13220.13931.51010.07420.08730.0649
200600.64320.41400.12720.15054.10792.75200.08510.0778
400400.43590.30010.11240.17531.22420.06690.09710.0630
400800.05470.31640.13770.04230.98540.16820.08960.0628
4001200.07700.01610.13910.07520.97640.08560.08290.0743
Table 8. The statistical properties of the estimators of π 1 and π 2 with Scheme 3. π ^ i MELE and π ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
Table 8. The statistical properties of the estimators of π 1 and π 2 with Scheme 3. π ^ i MELE and π ^ i MLE are the MELE by the proposed estimation method and the MLE by the traditional EM algorithm, respectively.
BiasMSE
n N π ^ 1 MLE π ^ 1 MELE π ^ 2 MLE π ^ 2 MELE π ^ 1 MLE π ^ 1 MELE π ^ 2 MLE π ^ 2 MELE
2020.04380.01490.07220.09740.26290.20630.14150.0554
2040.07710.00280.06940.05560.16430.12630.13770.0686
2060.14320.06680.03970.00640.17120.13350.11400.0967
100100.14450.06850.07700.07500.16170.11200.13350.0576
100200.15260.09470.05900.03430.16920.12940.12140.0722
100300.18060.14400.07460.05300.18610.16440.12210.1055
200200.14630.07180.07150.07700.16240.11530.12730.0557
200400.16850.10000.06780.03930.17930.12900.12250.0677
200600.17710.14630.07280.04950.18340.16150.12200.1078
400400.15110.07630.08760.07530.16750.11820.13840.0555
400800.16440.09770.06230.04230.17820.12890.12230.0652
4001200.17180.14150.06090.02480.17940.15570.11560.0961

Share and Cite

MDPI and ACS Style

Zhou, K.; Shi, Y. Evidential Estimation of an Uncertain Mixed Exponential Distribution under Progressive Censoring. Entropy 2020, 22, 1106. https://doi.org/10.3390/e22101106

AMA Style

Zhou K, Shi Y. Evidential Estimation of an Uncertain Mixed Exponential Distribution under Progressive Censoring. Entropy. 2020; 22(10):1106. https://doi.org/10.3390/e22101106

Chicago/Turabian Style

Zhou, Kuang, and Yimin Shi. 2020. "Evidential Estimation of an Uncertain Mixed Exponential Distribution under Progressive Censoring" Entropy 22, no. 10: 1106. https://doi.org/10.3390/e22101106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop