Abstract

The Topp Leone distribution (TLD) is a lifetime model having finite support and U-shaped hazard rate; these features distinguish it from the famous lifetime models such as gamma, Weibull, or Log-normal distribution. The Bayesian methods are very much linked to the Fuzzy sets. The Fuzzy priors can be used as prior information in the Bayesian models. This paper considers the posterior analysis of TLD, when the samples are doubly censored. The independent informative priors (IPs) which are very close to the Fuzzy priors have been proposed for the analysis. The symmetric and asymmetric loss functions have also been assumed for the analysis. As the marginal PDs are not available in a closed form, therefore, we have used a Quadrature method (QM), Lindley’s approximation (LA), Tierney and Kadane’s approximation (TKA), and Gibbs sampler (GS) for the approximate estimation of the parameters. A simulation study has been conducted to assess and compare the performance of various posterior estimators. In addition, a real dataset has been analyzed for the illustration of the applicability of the results obtained in the study. The study suggests that the TKA performs better than its counterparts.

1. Introduction

Topp and Leone [1] introduced a lifetime distribution and named it TLD. Nadarajah and Kotz [2] have discussed that the hazard rate function of the TLD is U-shaped. The distributions with U-shaped hazard rate are used to model the human populations where the death rate is high at the infant age owing to infant diseases and birth defects; thereafter, the death rate remains constant up to thirties, and then, it increases again. Some of the manufactured items also have similar life patterns. The advantage of TLD is that it has only two parameters, while most of the distributions with U-shaped hazard rate have at least three parameters which increases the computational difficulties. The regularity conditions are satisfied by the TLD. The distribution function of the TLD is in a compact form, due to which it can be very easily applied when the data are censored, unlike the gamma and lognormal distributions. Another feature of the distribution, which distinguishes it from the other lifetime distributions, is that it has a finite support. Due to these distinguishing aspects, the TLD has received much attention from the researchers recently. Al-Zahrani and Alshomrani [3] analyzed the stress strength reliability for the TLD. Genc [4] derived the expressions for single and product moments for the order statistics from the TLD. Genc [5] considered the estimation of from the TLD. Mir Mostafaee et al. [6] discussed the estimation for moments of order statistics from the TLD. Bayoud [7] estimated the shape parameter of the TLD using the Bayesian and classical methods under type-I censored samples. Bayoud [8] used progressive type-II censored samples to obtain the Bayesian and classical estimates for the shape parameter of the TLD. The approximate maximum likelihood method, LA, and importance sampling method have been used for the estimation. Mir Mostafaee et al. [9] studied the lower k-record values from the TLD using the Bayesian approach. Reyad and Othman [10] used the Topp Leone genesis to produce and generalization of the Burr-XII distribution and called it Topp-Leone Burr-XII distribution. Different properties and estimation of the proposed model have been discussed. The effectiveness of the proposed model has been demonstrated using three real-life datasets. Rezaei et al. [11] introduced a new family of distributions from the TLD and discussed the mathematical properties of this family of distributions.

The U-shaped hazard rate with a fewer number of parameters and finite support for the TLD distinguishes it from repeatedly used lifetime models such as Weibull, Gamma, and Log-normal distributions. Motivated by these features of the TLD, we planned to estimate this distribution under a Bayesian framework. The reason for the choice of Bayesian estimation is that the results under Bayesian inference are often better than the classical methods even if we do not have sufficient prior information; for example, please see Kundu and Joarder [12]. Further, the lifetime data are often censored. Therefore, we have considered doubly censored samples from the TLD for the estimation. Doubly censoring is used to analyze the duration times between two events. For example, let the ages of certain equipment are not known when the monitoring of the equipment starts and tracking of the equipment is stopped after the predetermined observational period. Then, the ages of the equipment which are still functioning will be doubly censored. The doubly censored samples are very useful in reliability analysis for the products which are already under use. In medical applications, the doubly censored samples usually deal with time between infection and onset of a disease; for example, the infectious diseases like COVID-19 often produce doubly censored data. Although Feroze and Aslam [13] and Feroze and Aslam [14] have considered the analysis of doubly censored samples from the single and mixture of TLD, respectively, these contributions considered only single parametric TLD which restricted the flexibility of the TLD to a great extent. In addition, these contributions did not include any credible intervals.

In addition, the Bayesian methods are very much linked to the Fuzzy sets. The Fuzzy priors can be used as prior information in the Bayesian models. The fuzzy approximations enable users to incorporate more flexible sets of priors and likelihood functions in the Bayesian inference. This method allows authors to bypass the closed form conjugacy link between likelihood function and prior [15]. Stegmaier and Mikut [16] suggested that the fuzzy priors can improve the performance of the Bayesian inference. This paper considers the posterior analysis of TLD, when the samples are doubly censored. The independent informative priors (IPs) which are very close to the Fuzzy priors have been proposed for the analysis. It has been noticed that the expressions for the posterior estimators are not available in the closed form, so we have used four approximation techniques, namely, QM, LA, TKA, and GS, for the analysis. The assumption of independent gamma priors has been made for both of the parameters; this assumption is not rare; for example, see Kundu and Howlder [17]. Further symmetric and asymmetric loss functions have been considered for the estimation.

The rest of the paper is placed in the following sections. The model and likelihood function have been introduced in Section 2. The Bayesian estimation has been considered in Section 3. The results regarding the simulation study have been reported in Section 4. The real-life example has been presented in Section 5. The concluding remarks have been given in Section 6.

2. The Model

In this section, the TLD and likelihood function under doubly censored samples have been presented.

The probability density function (pdf) of the TLD is where and are model parameters from TLD.

The cumulative distribution function for TLD is

Suppose that a random sample of size “” is selected from the TLD. Further assume the complete information for ordered observations was available, and the information regarding the smallest “” and largest “” items was incomplete. Therefore, observations can only be used for the analysis from the sample of size “.” Using the doubly censored sample , the likelihood function is

Putting results in (3), we have

3. Posterior Estimation

This section reported the posterior estimation for the model parameters assuming doubly censored samples. The gamma priors have been assumed for model parameters. The posterior estimation has been carried out using two loss functions. Various approximation methods have been suggested for numerical solutions of the estimates.

3.1. Prior and Posterior Distribution (PD)

In this subsection, we have assumed IP for the construction of the PD for the parameters and . Consider the informative gamma priors for the parameters and as and , respectively. Now, under the assumption of independence the combined IP for the parameters and is

The prior given in (5) is quite close to the Fuzzy priors [15].

From (4) and (5), the joint PD for model parameters and under IP are

Putting values in (6), we have the PD as

3.2. Loss Functions

The posterior estimation has been carried out using following two loss functions. (1)Squared Error Loss Function (SELF): Legendre [18] and Gauss [19] proposed the SELF that can be defined as , where is the model parameter. Using SELF, the Bayes estimator (BE) for the parameter is .(2)Precautionary Loss Function (PLF): the PLF, introduced by Norstrom [20], has the following expression . Using PLF, the BE is .

3.3. Quadrature Method (QM)

It should be noted that the assumed priors are flexible; however, the BEs under SELF and PLF cannot be obtained in an explicit form. This is due to the fact that the ratio of the integrals in the BEs cannot be solved directly. In such situations, the BEs can be obtained numerically using QM which can be used to evaluate an integral numerically.

The BEs for the parameters and under SELF using the PD based on IP distribution are

In the Bayesian QM, we choose a set of points between the finite integral in an order to ensure the stability of our uncertainty. Consider the posterior density, where and are the parameters. We evaluate this density over a number of the points in the entire range as where are the increments. The Mathematica software has been used to obtain BEs and associated PRs for the parameters and using SELF and PLF under IPs.

3.4. Lindley’s Approximation (LA)

The QM is not suitable in many situations, for example, in the case of functions with singularities; the method is not well suited. In such situations, some other approximation methods, such as LA, can be used. The LA has an important role in Bayesian inference. Using this method, we can compute BEs without performing any complex numerical integration. Hence, if we need to obtain only the BEs, the LA can be efficiently used to serve this purpose. Bayoud [8] used LA to estimate the shape parameter of the TLD considering uncensored data. We have considered the more general and complex case by performing the LA for both parameters of the TLD using censored data.

Assuming a sufficiently large sample, Lindley [21] proposed that any ratio of the integral of the form where is any function of or , is the log-likelihood function, and is the logarithmic of joint prior for the parameters and , can be evaluated as where and are MLEs of the parameters and , respectively,

and is the element of the inverse of the matrix , all evaluated at the MLEs of the parameters.

Now, the log-likelihood function from (4) can be obtained as

The maximum likelihood estimates (MLEs) of the parameters and can be obtained by differentiating (13) with respect to and and equating to zero, respectively, as

Let , , , , , , , , , and ; then,

Based on (14) and (15), the approximate MLEs have been obtained using numerical methods. The second order derivatives of the log-likelihood function are presented in the following:

Now, ((16)–(18)) have been evaluated at the MLEs of and .

As the third order derivatives with respect to and contain long expressions, therefore, they have not been presented in the paper.

Based on the second order derivatives, the matrix is

Now, using LA, the BEs for the parameters and under IP using SELF are, respectively, presented as

Again, using LA, the BEs for the parameters and under IP using PLF are, respectively, presented as

3.5. Tierney and Kadane’s Approximation (TKA)

The LA requires the evaluation of the third derivatives from the log-likelihood function which at times gets tedious, especially in case of problems with several parameters. This issue can be resolved by considering rather easily computable approximation called TKA. The additional advantage of this approximation is that it has less error as compared to LA. Therefore, in this subsection, we have considered TKA for the approximate Bayes estimation of the parameters from the doubly censored TLD. Consider , where is the logarithmic of the joint IP for the parameters given in (5), and is the logarithmic of likelihood function given in (4).

Further consider and , where is the logarithmic of the function of the parameter(s) or . Then, according to Tierney and Kadane [22], the expression using (7) can be reexpressed as

Now using the Laplace’s method, the approximation for can be given as where and maximize and , respectively, and and are the negatives of the inverse Hessians of and evaluated at and , respectively.

Here, we have where is any constant independent of the parameters and .

Now, can be obtained by solving (25) and (26).

The determinant for the negative of the inverse Hessian of evaluated at is where , , and .

The second order derivatives from contains lengthy expressions; therefore, they have not been presented here. Similarly, MLEs and the second order derivatives for the can be obtained. Further, the BEs and the PRs under IP can be obtained by using (23).

3.6. Gibbs Sampler

Unfortunately, by using QM, Lindley’s method, and TKA, it is impossible to obtain the highest posterior density (HPD) credible intervals. The HPD credible intervals can be obtained by using the GS. Hence, in this subsection, we have used the GS to obtain the approximate BEs along with HPD credible intervals.

Consider a PD with two parameters and . Assuming that the full densities and are extractable, we need to obtain and . Using GS, we start with choosing initial guess values for parameters and . The said initial values are denoted by and . Then, we generate random samples from the conditional distributions in the following sequence:

In order to employ the GS for the PD (7), we extract the conditional distribution for given under IP as

The conditional distribution of the parameter given under IP is

Now, using the conditional distributions from (29) and (30), the GS can be implemented following the methodology proposed by Pandey and Bandyopadhyay [23] using Winbugs software. The generated samples for the parameters and can be used for the estimation of the parameters using SELF and PLF. The BE and the posterior risks (PRs) for the parameter using SELF and IP can be obtained by using the formulae and , respectively. Similarly, the BE and PR for the parameter using PLF and IP can be calculated by using the formulae and , respectively. A similar process can be followed to implement GS for the PD under IP and for the PDs for the parameter .

4. Simulation Study

As the performance of the BEs cannot be compared analytically, we have carried out a simulation study to compare different estimators numerically. The comparison among the BEs is based on the level of the convergence of the estimates and the amounts of the PRs. The following parametric space has been used for the estimation: and . The samples of sizes , 30, 50, and 100 have been considered. The values for the hyperparameters have been considered in such a way that, for these values, the mean of the concerned prior is equal to the true parametric value. The samples have been assumed to be 20% censored: 10% from the left and 10% from the right. The test termination point in each sample is assumed to be so that observed the said censoring rate. The results for the simulated samples have been presented in Tables 18. The results using parametric space , using samples of sizes 15, 30, 50, and 100, have been reported in Tables 14, respectively. Similarly, the results for sample sizes 15, 30, 50, and 100, using parametric space , have been given in Tables 58, respectively.

The results using simulated samples have been presented in Tables 110. The said results advocate that the proposed estimates are consistent in nature because the estimated values of the parameters converged to the true parametric values with increase in the sample size. The performance of the BEs has been better as compared to ML counterparts. The convergence under BEs was better as compared to MLEs. In additions, the amounts of PRs associated with each estimate become smaller as sample size increases. The performance of loss functions and different approximation methods has been compared on the basis of amounts of PRs associated with each estimate.

It is interesting to note that for smaller choice of true parametric values, the SELF performs better than PLF. Conversely, the performance of PLF seems better for estimation. There is a sense of competition among the performance of different approximation methods; however, the TKA provides the minimum amounts of the PRs among all the approximation methods with few exceptions. The amounts of the PRs are directly proportional with the magnitude of the true parametric values which is in accordance with the theory. In general, the parameter () has been estimated more efficiently than the parameter ().

5. Real-Life Example

In this section, the data regarding the failure times (in minutes) for a specific type of electrical insulation has been used for the illustration of the results obtained in the previous section. The data has been used by Fernandez [24] and is reported by Lawless [25]. The twelve items have been considered for the experiment out of which the experimenter was unable to observe the smallest two failure times and experiment was terminated after the ninth failure. The data is as follows: -, -, 24.4, 28.6, 43.2, 46.9, 70.7, 75.3, 95.5, -, -, and -. Therefore, we have , , , and . The descriptive statistics for the data are , , , , and . The data has been censored in such a way that same number of observations has been censored from left and right. We have named this dataset as dataset-1. The other dataset used in the study has been named as dataset-2. The dataset-2 represents the failure times of the air condition system of an airplane, reported by Linhart and Zucchini [26]. The dataset-2 consists of observations 1, 3, 5, 7, 11(3), 12, 14(3), 16(2), 20, 21, 23, 42, 47, 52, 62, 71(2), 87, 90, 95, 120(2), 225, 246, and 261. The number in the parenthesis represents the repeated observations. The chi-square and the Kolmogorov–Smirnov tests have been used to confirm that the data can be modeled using TLD. The BEs and corresponding PRs, based on these data, have been presented in Tables 11 and 12, respectively. For convenience, the numerical values for the hyperparameters have been assumed to be for the estimation.

Tables 11 and 12 contain the BEs and associated PRs using different approximation methods. The amounts for the PRs under PLF are on average smaller to those under SELF. As far as the approximation is concerned, the performance of the TKA seems superior to its counterparts, as the least magnitudes of the PRs are associated with these estimates. Hence, the results obtained by analyzing the real data are in agreement to those obtained from simulated samples. The marginal density plots for model parameters under different approximation methods have been presented in Figures 18. For brevity, we have presented the density plots only for dataset-1. The density plots for parameter are positively skewed. However, the skewness is not that high for density curves of parameter . The density plots under PLF are representing slightly more skewness as compared to those for SELF. From these graphs, it can also be seen that the graphs for marginal densities under SELF and PLF are relatively closer under TKA for both parameters. However, the graphs for marginal densities using other approximation techniques, such as QM, LA, and GS, are quite different under SELF and PLF. This simply suggests that the estimates under TKA are relatively robust with respect to choice of loss functions.

6. Conclusion

This paper considers the posterior analysis for the parameters of the TLD assuming doubly censoring. As the expressions for the BEs cannot be obtained explicitly, we considered different approximation methods to obtain the numerical solutions. The performance of the BEs is compared based on amounts of PRs. The analysis of simulated and real datasets suggests that there is a sense of competition among different approximation methods; however, the performance of the TKA is the best among the approximate methods used in the study. The study can be extended for application of TLD in modeling doubly censored heterogeneous data.

Data Availability

The data used in the paper are available in the paper.

Conflicts of Interest

The authors declare that they have no conflicts of interest.