Abstract

In this study, we model a heterogeneous population assuming the three-component mixture of the Pareto distributions assuming type I censored data. In particular, we study some statistical properties (such as various entropies, different inequality indices, and order statistics) of the three-component mixture distribution. The ML estimation and the Bayesian estimation of the mixture parameters have been performed in this study. For the ML estimation, we used the Newton Raphson method. To derive the posterior distributions, different noninformative priors are assumed to derive the Bayes estimators. Furthermore, we also discussed the Bayesian predictive intervals. We presented a detailed simulation study to compare the ML estimates and Bayes estimates. Moreover, we evaluated the performance of different estimates assuming various sample sizes, mixing weights and test termination times (a fixed point of time after which all other tests are dismissed). The real-life data application is also a part of this study.

1. Introduction

In the last decade, finite mixture models have emerged as flexible models due to their applications in applied sciences, engineering, and physical sciences. As explained by Mendenhall and Hader [1], for real-life purposes, an engineer split the failures of a structure into more than one kind of causes. For example, to know the proportion of failures due to a definite reason and to recover the engineering system, Acheson and McElwee [2] separated electronic tube flops into three different faults such as mechanical faults, gaseous faults, and normal decline of the cathode.

Moreover, the mixture models can also be used in a situation when the data are presented in the form of the overall mixture models. The overall mixture models are also called the direct application of the mixture models, and their applications can be seen in medicine, botany, zoology, agriculture, economics, life testing, reliability, and survival analysis. The various aspects of mixture models were discussed by Li and Sedransk [3]. The interested readers can refer the work of Harris [4], Kanji [5], and Jones and McLachlan [6] on the application of mixture models for real-life problems.

The mixture models have been extensively used for heterogeneous nature of the process in comparison to the simple models. Most of the researchers have comprehensively applied mixture distributions in various real-life situations and estimated parameters using the Bayesian and classical methods. For a detailed appraisal of classical techniques for estimation and applications of mixture distributions, we refer to studies by Sultan et al. [7], Abu-Zinadah [8], and Kamaruzzaman et al. [9] among others. On the other hand, the estimation of parameters in Bayesian framework for a mixture of two distributions has been considered by many researchers [1021]. Contrary to the two-component mixture modeling, some authors have discussed situations where data are assumed to follow a three-component mixture of suitable probability distributions [2228].

Censoring is a significant characteristic of the real data application. Due to time and cost problem, it is very difficult to continue the lifetime testing experiment till observing the last failure. Although there are many censoring schemes, the type I right censoring is commonly used in life testing experiments. In this scheme, we consider a fixed censoring time, and the values larger than the specific t (life test termination time) are observed as censored observations. Romeu [29] and Kalbfleisch and Prentice [30] explained various censoring schemes.

To motivate the readers about the mixture modeling, consider a sample of sand which is based on the mixture of some minerals. With the application of mixture modeling, estimates of the proportions of various minerals in the sand can be obtained. Similarly, the grain size distributions for the different minerals can also be estimated. It is worth mentioning that mixture models can be classified into type I (if component densities of the various components belong to the same family) and type II (if the component densities belong to different families) mixtures.

It has been noticed that from the recent literature that the Pareto distribution can be applied efficiently in various situations rather than other distributions to model data. The significance of Pareto distribution in forming various real phenomena is patent from the following revealed research and references mentioned therein, Abdel-All et al. [31], Ismail [32], Sankaran and Nair [33], and Nadarajah and Kotz [34]. Inspired by the wide real-life applications of mixture distributions, the main objective of this study is to develop a new three-component mixture of Pareto distributions (TCMPD) for lifetime data modeling under type I mixture. Furthermore, we also compare the maximum likelihood (ML) estimates, ML variances (MLVs), Bayes estimates, and their posterior variances (PVs) assuming type I right censored data.

2. The Population and the Three-Component Mixture Distribution

The finite -components mixture density function can be defined for random variable as , where is the dth component density function, , is the dth mixing proportion, and . A finite TCMPD is defined aswhere , , .

The cdf of a TCMPD iswhere .

3. Properties of the TCMPD

The statistical properties such as moments, mean, variance, and mode of the TCMPD are derived in this section.rth moment about zero: the rth moment about zero of a TCMPD is derived askth order negative moment: the kth order negative moment is derived asFactorial moments: the factorial moments can be determined asWhere is the real number. The can be obtained ashMean: the mean of a TCMPD is evaluated ashVariance: the variance of a TCMPD is derived asMedian: the median (y) is determined by evaluating the following nonlinear equation for y.Mode: the mode (y) is obtained by solving the following nonlinear equation for y.

Using the above expressions, mean, median, variance, and coefficient of skewness (SK) are calculated for different parameters’ values and are given in Table 1.

It is observed from the entries in Table 1 and Figure 1 that TCMPD is positively skewed as we have SK > 0 for all the entries arranged in Table 1. We noticed that the variance of the TCMPD was a declining function of the mixture distribution’s parameters.

4. Entropies

The entropy is a quantity of unspecified extent of evidence in a function. Here, we derive the expressions of the most commonly used entropy measures such as Shannon’s entropy, entropy, and Re´nyi entropy, in this section. As said by Song [35], Shannon’s entropy has a same behavior as a measure of kurtosis in equating the forms of different functions and computing substance of tails.Shannon’s entropy: Shannon’s entropy for a random variable Y, which follows TCMPD, isRe´nyi entropy: the Rényi entropy (Rényi, [36]) is explained asWhere , and . The Rényi entropy of a 3-CMPD isEntropy: the entropy (Ullah, [37]) is written asWhere , and . The entropy of a TCMPD isWhich can be evaluated by the numerical integration.

5. Inequality Measures

The most common income inequality indices are Bonferroni curve, Gini index, and Lorenz curve.The Gini index: this index (Gini, [38]) for a TCMPD isThe Lorenz curve: this curve (Lorenz, [39]) for a TCMPD isBonferroni curve: the Bonferroni [40] curve for a TCMPD is

6. Order Statistics

In this section, we derive which is pdf of kth order statistic assuming a sample of size n from the TCMPD. The rth raw moment along with mean and variance of 1st and nth order statistics are obtained in this section.

Probability density function of kth order statistic: the pdf of kth order statistic iswhere ,

After little simplification, the pdf (19) of kth order statistic can be written aswhere , , , ,

Probability density function of 1st order statistic: substituting in (21) and simplifying it, the pdf of 1st order statistic iswhere , , , ,

Probability density function of nth order statistic: substituting in (21) and after little algebraic simplification, the pdf of nth order statistic iswhere , , , ,

rth moments, mean, and variance of 1st order statistic: the rth moment about the origin, mean, and variance of 1st order statistic are

rth moments, mean, and variance of nth order statistic: the rth moment about the origin, mean, and variance of nth order statistic are

7. Parametric Estimation

Here, we discuss the parameter estimation methods. In particular, we use the ML and Bayesian methods of estimation to estimate unknown parameters under type I censored data. Before discussing the parameter estimation, we construct the likelihood function.

Assume units, with fixed as test termination time, from the TCMPD are used in a real-life experiment. Let be the ordered values that can only be observed. The remaining largest values are censored from the study, i.e., their exact failure time cannot be recorded due to time constraint. So, , , and are failed observations relating to subpopulations I, II, and III, respectively. The observation is assumed to be censored from each component, whereas the numbers , , and of failed values are obtained from the subpopulations. The , , and observations are assumed to be censored values from subpopulations, whereas . So, the likelihood function using type I censored sample, , , and , for a TCMPD is After substitution and simplification, the likelihood function of TCMPD becomes

7.1. ML Estimators and Variances

The ML estimators of TCMPD for parameter are derived from the solution of nonlinear equations (30)–(34). The equations have been derived by partially differentiating the natural logarithm of the likelihood function as

It is very tough to find out an explicit solution from the nonlinear equations (30)–(34); therefore, to obtain ML estimators, any mathematical or statistical software such as Mathematica (Wolfram, [41]) can be used to solve them by an iterative procedure.

Let , and by using multivariate central limit theorem, that is, , one can get asymptotic variances, where the variance is given on diagonal of the inverted Fisher information matrix which is expectation of the negative Hessian aswhere

Next, we discuss the Bayesian estimation for the estimation of unknown parameters.

7.2. The Joint Prior and Posterior Distributions

Now, we discuss the Bayesian estimation of the unknown parameters. This method allows us to obtain an updated form of the knowledge which is calculated by combining the current and the prior knowledge. In particular, we use uniform and Jeffreys as noninformative priors, which are used when little or no formal prior knowledge on the parameters of concern is available. Box and Taio [42] stated a noninformative prior which gives slight information relative to the testing experiment. Bernardo and Smith [43] defined a noninformative prior has the least influence relative to the data. Jeffreys prior suggested by Jeffreys [44] is obtained by evaluating the Fisher information.

7.2.1. The Joint Prior and Posterior Distributions Assuming the UP

We take the improper UP for component parameters, i.e., , . Also, we take the UP for the unknown proportion parameters, i.e., , . Under the assumption of independence of parameters , and , the joint prior distribution is

By combining the likelihood function and the joint UP prior, i.e., , we obtain the joint posterior distribution of parameters , and given data aswhere , , ,

7.2.2. The Joint Prior and Posterior Distributions Assuming the JP

The JP for component parameters , , is , where is the Fisher information and is . The prior distributions of and are assumed as the UP, i.e., , . So, the joint prior distribution is

Using the JP, the joint posterior distribution of , and that give data iswhere , , ,

7.3. Bayes Estimators and Posterior Variances

The Bayes estimators of the component and mixing proportion, i.e., , and using the UP are obtained aswhere , and are defined as (i) (ii) , and (iii) .where , and are defined as (i) and (ii) .

To measure the accuracy and efficiency of the Bayes estimators, we normally calculate the PVs of the parameters. For the PVs of , and , using the UP are derived as

The Bayes estimators of , and using the JP are obtained as

Also, the PVs of the parameters , and using the JP are derived as

7.4. The Bayesian Prediction

One of the main objectives of statistical modeling is the prediction of the future values. The Bayesian methodology allows us to obtain this in a natural way. In particular, the posterior predictive distribution (PPD) comprises the knowledge about future value given data . Al-Hussaini et al. [45], Bolstad [46], and Bansal [47] have discussed the usefulness of the prediction and predictive distribution comprehensively in the Bayesian framework.

7.4.1. The Posterior Predictive Distribution

For the future value , the PPD using the UP and the JP is

In equation, we consider for UP and for JP, respectively.

7.4.2. Bayesian Predictive Intervals

To obtain the Bayesian predictive intervals (BPIs), let L and U are the two lower and upper endpoints of the BPI, which are obtained from (49). A BPI (L, U) using UP and JP can be obtained by simplifying the given expression:

8. Monte Carlo Simulation

Here, we tabulate a comprehensive simulation to check the performance of different estimation methods. As an analytical comparison of the Bayes and ML estimators is not possible, a Monte Carlo simulation is performed to measure the performance of the Bayes and ML estimators under different aspects. Through the following steps, we obtained the maximum likelihood estimates (MLEs), MLVs, Bayes estimates (BEs), and PVs as(1)A sample of size from mixtures may be taken as(i)Generate a sample of values randomly from (ii)Generate a sample of values randomly from (iii)Generate a sample of values randomly from (2)Select the values that are larger than fixed as censored values(3)For a large number of times, say 1000, repeat steps 1 and 2 for the fixed , , and parametric values(4)Evaluate the MLEs and MLVs of , and using nonlinear equations based on the samples obtained in step 3(5)Calculate the BEs and PVs based on 1000 repetitions

The above procedure is adopted for 50, 100, and 200 as sample sizes , parametric values , and The choice of was done to have 8–20% censored rate in the resulting sample. Moreover, to have a perceptible indication about the behavior of the MLEs, BEs, MLVs, and PVs, we depicted some graphs in Figures 14 and Tables 2 and 3.

From Figures 14, it is observed that parameters , and are overestimated, but is underestimated at different values of and in both estimation methods, i.e., ML and Bayesian. Also, the degree of underestimation of , and is higher for a small at various values of , and an opposite behavior was observed for a large at a given . Furthermore, the parameters , and were observed overestimated to a larger extent when the true values of , and were smaller at different values of for a fixed . In addition, the similar pattern has been observed at different values of for a fixed . The difference of the MLEs and BEs of parameters , and from the nominal values becomes the minimum with the increase of and .

It can be seen from Tables 2 and 3 that at different values of , the difference between the MLVs and the PVs (assuming the UP and the JP) diminishes by increasing sample size. The same remark is true for a large at different values of . Also, noticed that the MLVs and PVs of and are larger for smaller values of , and at different values of and . Also, it is pointed out that the performance of the Bayes estimators using JP is best than Bayes estimators using UP and ML estimators based on lesser associated PVs.

To the extent that the selection of an appropriate prior, Tables 2 and 3 revealed that the JP outperforms as compared to the UP because the variance of JP is smaller than the UP.

Table 4 and Figure 5 showcase the 90% BPI using the UP and the JP. It is pointed out that the width of 90% BPI increases with a decrease in . The same conclusion was observed with a smaller for varying values of . The 90% BPIs, for larger values of , and , were observed narrow at various values of and . Moreover, the BPIs using the JP were observed wider than the BPIs obtained by assuming the UP in the simulation study.

9. Real Data Application

To illustrate the proposed methodology, the mixture lifetime data, in thousand hours, was taken from Davis [48] on three factors, i.e., V805 Transmitter, Transmitter, and V600 Indicator Tube used in aircraft sets. For exponential distributed mixture data , the suitable transformation gives the Pareto distributed mixture data . So, the mention transformation permits to utilize the given mixture data for using the suggested ML and Bayesian estimation techniques. Thus, the proposed mixture of the Pareto distributions can be a fair choice to model the abovementioned data. Moreover, it is unidentified that which factor fails until a failure arises at or before 0.6 hours. To calculate the MLEs, MLVs, BEs, and PVs, the data summary is

The MLEs, MLVs, BEs, and PVs assuming the UP and the JP are presented in Table 5.

From Table 5, it is clear that the performance of the BEs using JP is the best as compared to the MLEs, as the variance of BEs is smaller than the counterpart. Moreover, the BEs using the UP have smaller variances for estimating the unknown parameters. Also, the JP was observed superior to the UP due to smaller associated PV for estimating the unknown parameters.

10. Conclusion

In this study, we proposed TCMPD to model lifetime data. Parameter estimation assuming the type I censoring have been considered using the ML and Bayesian estimation methods. For Bayesian estimation, we assumed the noninformative priors and expressions of the Bayes estimators for the mixing proportion ( and ) and component parameters (, and ) and PVs were derived. To examine the relative presentation of the Bayes and ML estimators under different scenario, a Monte Carlo simulation has been done. To illustrate a practical presentation of proposed mixture distribution, an example has also been analyzed.

From simulated results and depicted graphs, it has been noticed that an increase in under a fixed yields very efficient Bayes and ML estimators. It is also pointed that parameters , and are overestimated (underestimated) to a small (larger) extent with relatively larger (smaller) value of (value of ). More specifically, the amount of overestimation (underestimation) of parameters is smaller for a relatively large parameter value. As the value of (value of ) increases (decreases), the PVs decrease (increase) for a fixed (fixed ). To address the problem of selection of a suitable prior, one can observe that the JP has smaller PVs than the UP. The results depend on real-life mixture data that also support the Monte Carlo simulation study. Finally, it is concluded that the Bayes estimators using the JP performed better as compared to the ML estimators because of smaller variances.

Data Availability

The dataset used to support the findings of this study is available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through research groups program under grant number R.G.P.2/82/42.