Abstract

In reliability studies, the best fitting of lifetime models leads to accurate estimates and predictions, especially when these models have nonmonotone hazard functions. For this purpose, the new Exponential-X Fréchet (NEXF) distribution that belongs to the new exponential-X (NEX) family of distributions is proposed to be a superior fitting model for some reliability models with nonmonotone hazard functions and beat the competitive distribution such as the exponential distribution and Frechet distribution with two and three parameters. So, we concentrated our effort to introduce a new novel model. Throughout this research, we have studied the properties of its statistical measures of the NEXF distribution. The process of parameter estimation has been studied under a complete sample and Type-I censoring scheme. The numerical simulation is detailed to asses the proposed techniques of estimation. Finally, a Type-I censoring real-life application on leukaemia patient’s survival with a new treatment has been studied to illustrate the estimation methods, which are well fitted by the NEXF distribution among all its competitors. We used for the fitting test the novel modified Kolmogorov–Smirnov (KS) algorithm for fitting Type-I censored data.

1. Introduction

Modeling real-life events and natural processes using probability distributions is one of the most important processes in statistics and probability, where these processes are characterised by complexity and risk. For these reasons, statisticians have worked on the development of probability distributions, as proven probability distributions continue to fall short of accurately describing data obtained from natural events. These help to expand and modify probability distributions in generalized ways. Generalized probability distributions have emerged as a consequence of the widespread availability of additional parameters. Adding a special parameter to existing probability functions increases the precision of adequacy of the data obtained from natural phenomena as well as the accuracy of the distribution tail form’s description.

In recent years, various research works were undertaken to create new distributions through creating new families and classes by modifying the baseline distribution by adding additional shape parameter(s). There are a large number of well-known classes of distribution that exist in the literature. For examples, see references [110].

One of the most well-known lifetime distributions is the exponential distribution, which has accordingly received considerable attention from statisticians. A superior novel family of distribution dubbed as a new exponential-X (NEX) family was introduced by Huo et al. [11]. It can model data with different shapes to their hazard function, such as increasing, decreasing, and bathtub. The distribution function (CDF) and density function (PDF) for the NEX family are defined aswhere the vector of the parameters is denoted by of the family. It consists of , which are the vectors of parameters for the baseline distribution and the additional shape parameter for the family, respectively.

The standard Fréchet distribution is a well-defined limiting distribution. It is commonly used to characterise variables associated with extreme phenomena like floods, rains, and cash flow. It was first introduced by Fréchet [12]. We can say that is a random variable that has the two-parameter Fréchet distribution if its CDF and PDF are written as follows:where the vector of the parameters consists of the shape and scale parameters, respectively, for the Fréchet distribution.

The Fréchet distribution has received attention of a large number of authors such as Nadarajah and Kotz [13], where the standard Fréchet distribution is generalised as an exponentiated Fréchet distribution. Cordeiro et al. [14] discussed a new class of exponentiated generalised distributions, with regard to which functions such as the Exponentiated Generalised Fréchet, Exponentiated Generalised Normal, Exponentiated Generalised Gamma, and Exponentiated Generalised Gumbel have been introduced. A new lifetime distribution called the Weibull Fréchet distribution, which has four parameters, has been defined and studied by Afify et al. [15]. Teamah et al. [16] presented the Fréchet–Weibull distribution with application to earthquake datasets. Teamah et al. [17] introduced the Fréchet–Weibull Mixture Distribution. Almetwally and Muhammed [18] introduced a new novel bivariate distribution depending on copulas. Furthermore, Teamah et al. [19] presented a right truncated model of the Fréchet–Weibull Distribution. Mead et al. [20] introduced the beta exponential Fréchet distribution. The properties of the Gompertz Fréchet distribution have been studied by Oguntunde et al. [21]. An overview of the various estimators and applications for the Fréchet distribution has been discussed by Ramos et al. [22].

The innovations and encouragements to write this article are to introduce a new exponential-X Fréchet distribution as a good fit for the lifetime models that have increasing, decreasing, and bathtub failure rates. We have studied its mathematical properties such as its linear representation, quantile function, moments, generating function, incomplete moments, mean residual life and mean inactivity time, inequality curves, and order statistics. Furthermore, approaches to parameter estimation such as maximum likelihood, maximum product spacing, and Bayesian methods have been discussed using Type-I censored data. The Markov Chain Monte Carlo (MCMC) algorithm has been used as an approximation to Bayesian estimates. An extensive simulation study has been done. A comparison of the suggested estimate techniques has been conducted. Finally, the NEXF distribution has demonstrated its efficiency at fitting real-life data to a greater extent than other competitor distributions with regard to real-life applications.

The following is the structure of the paper: in Section 2, the proposed distribution is presented, and its behaviour is studied using different graphs. Various mathematical properties have been studied in Section 3. Section 4 discusses the estimating techniques used to determine parameters. Section 5 contains the interval estimation for the distribution’s parameters. The simulation experiments have been performed, and its results were tabulated in Section 6. We worked on a real numerical example for the remission times as a data application, and this application is analyzed and fitted using the suitable fitting algorithm according to Type-I censoring in Section 7. After this, Section 8 contains the current study conclusions and other remarks.

1.1. Type-I Censoring Scheme

Assume that identical components are arranged in a life testing experiment. We perform the experiment using a scheme of a Type-I censored sample. The test comes to an end at predetermined point in time , and only the failure times prior to this time are recorded. Then, the number of units that fail by time , say , is a random variable, where . The lifetimes of the initial failures are determined, as well as the lifetimes of the subsequent failures of the remaining units are censored, where all that is known is that they will be greater than . In this case, the model can be known as a Type-I censoring model (see Balakrishnan and Aggarwala [23]). For more examples of censored sampling based on different schemes, see [2433].

Assume that the Type-I censored sample arises from CDF with PDF . The joint density function of Type-I censored data, , is then given bywhere .

2. The NEXF Distribution

Based on the NEX family with the Fréchet baseline distribution, we have obtained the NEXF distribution according to three parameters. We can derive easily the CDF and PDF of the NEXF distribution by the aid of the the Fréchet distribution and both equations (1) and (2), which can be given as follows:

We can say that a random variable is distributed according to the NEXF distribution with PDF as in equation (6) by NEXF. The two-parameter NEX inverse Rayleigh model is a special case of the NEXF model obtained by letting .

The NEXF distribution’s hazard rate function (HR) is provided by the following:

While its revised hazard rate function (RHR) can be written and defined by the following equation:

In order to make a reasonable study on the distribution, we chose different values for its parameters, and the graphs of its PDF and HR function are plotted in Figures 1 and 2 . From these figures, it may be noticed that the behaviour of the NEXF PDF curve may have different shapes. It may be skewed to right or even to left, or posses symmetric shape or declining shape, while the NEXF HR curves may be constant, decreasing, or upside down, which means that the proposed model is an attractive lifetime model. The NEXF distribution shows great flexibility in terms of its ability to model skewed data, as noted in the application section, so it sees widespread use in various fields such as biology, biomedical experiments, reliability, and survival studies.

3. Mathematical Properties

3.1. Linear Representation

The CDF of equation (1) is represented in the following form:

By applying the previous expansion to the Fréchet distribution, we gain the CDF of the NEXF distribution in an expanded form as per the following:

By calculating the derivative of the above equation in terms of , we can determine and easily find the PDF of the NEXF distribution in an expanded form as per the following:

We can say that and is the Fréchet density function having a as a scale parameter and a shape parameter .

3.2. Quantile Function QF

We can express and define the QF of the NEXF distribution as the inverse of the CDF of equation (5), and it is as follows:where is the Lambert function.

The NEXF distribution’s 3 quartiles are determined by specifying special values for p, so when we set in equation (12) , we get the first and second quartile, respectively, while when we set and , we get the third quartile in equation (12).

Let is distributed uniformly from 0 to 1, then the QF of the NEXF distribution can be adopted to generate a randomized sample that can be used in simulation and other applications throughout the following equation:

3.3. The Moments

The most important thing for any distribution is to find its moments in an easy and simple ways. So, we can find the moments of the proposed NEXF distribution as shown in the following steps:

By assigning , and 4 in the preceding equation, we can drive the moments around the origin, from the first moment to the fourth moment, respectively.

We can derive the central moment of the random variable X, say , which is obtained as

The cumulants of X can be found as follows:

3.4. Moment Generating Function (MGF)

We can write the MGF of the Fréchet distribution using the following equation:where by setting and by expanding the first exponential, we have

Assume that the right generalised hypergeometric function is described as follows:

As a result, we may write the MGF as

Then, the MGF of the NEXF distribution can be written and defined as follows:

We may get and determine the formula for the proposed distribution’s characteristic function simply by substituting for in the preceding equation.

3.5. The Incomplete Moments

One of the fundamental formulas is the incomplete moments, so this subsection is devoted to derive the formula of the incomplete moment of the NEXF distribution:such that is the lower incomplete gamma function.

3.6. Mean Residual Life (MRL) and Mean Inactivity Time (MIT)

The MRL of the NEXF distribution is defined as follows:where is the first incomplete moment of the NEXF distribution.

The MIT of the NEXF distribution is defined as follows:

3.7. Inequality Curves

For the NEXF distribution, the following Lorenz curves are defined:where , is the first incomplete moment, and is the quantile function.

Additionally, we can determine the Bonferroni and Zenga inequality curves based on the correlations to the Lorenz curve as shown in the following equation (for further details and further readings, see [34]):

3.8. Order Statistics

For the order statistic of the NEXF distribution, the PDF and CDF are given as follows:where is a hypergeometric function.

4. Point Estimation Methods

This section is devoted to explain the different estimation methods that have been applied to to find the values of the estimates of the NEXF parameters according to two cases: firstly, a complete sample; secondly, Type-I censored samples. The estimation methods used are the maximum likelihood method (MLE) which is the most famous estimation technique to evaluate the parameters. Additionally, we used an important method of classical estimation which is the maximum product of spacing (MPS) method, and also we used the Bayesian estimation method depending on the squared error loss function.

4.1. Maximum Likelihood Method

Suppose that we have an order sample as following such that be a random sample from the NEXF distribution. This sample is based on the Type-I censored sample with vector parameter , and the likelihood function of the NEXF distribution under the Type-I censored sample takes the form as follows:where does not depend on parameters. The log-likelihood function of NEXF based on the Type-I censored sample takes the form as follows:

Now, by differentiating the log-likelihood equation (29) with respect to , and separately, we get the following equations:

Since the equations in (30)–(32) are not solved analytically, numerical approaches will be used to solve these equations such as the Newton–Raphson method.

4.2. The Maximum Product Spacing Method

In this subsection, we devoted our efforts to study the most famous classical method of estimation, namely, the MPS method which is considered as the first competitive for the MLEs. By referring to Ng et al. [35], Almetwally et al. [36], and Alshenawy et al. [37], the log-MPSEs under type-I censored samples for the NEXF distribution take the form as follows:

The MPSEs of distribution parameters for the Type-I censored samples can be derived through the following steps:(1)First derive the log-product equation (33)(2)Find the partial derivative for equation (33), with respect to each existing parameter, respectively(3)We all know that these equations are extremely tough to solve, so we will start here by using nonlinear optimisation algorithms such as the Newton–Raphson algorithm takes a role in solving these kinds of problems

4.3. Bayesian Estimation

This subsection presents Bayesian parameter estimations for the parameters based on the squared error (SE) loss function of the NEXF distribution parameters , and based on the Type-I censored sample. The prior distributions of the parameters are chosen to be gamma distributions. Thus,

By presuming the independency of the proposed model parameters, we can get and formulate the joint PDF of the priors as follows:

Now, the posterior function of the proposed distribution’s parameters may be calculated using equation (28) and also equation (35) as follows:

Bayesian parameter estimation for the NEXF distribution using the SE loss function is given by

It is very clear that the integrals in equations (37), (38) and (39) are complicated. Consequently, the Markov Chain Monte Carlo (MCMC) and Metropolis-Hastings (MH) algorithm are applied to obtain approximations for these integrals.

4.3.1. Markov Chain Monte Carlo Technique

Multiple integrals, as we all know, are incredibly difficult to solve analytically or even mathematically by hand. To find an estimate for these integrals, we must use the MCMC technique. The MH algorithm, also known as the random walk algorithm, is an integral part of the MCMC technique. It is very similar to the process of acceptance and rejection sampling (Algorithm 1).

(1)Initiate using these values , , and where these values are the values evaluate form the MLEs
(2)Start the loop with value
(3)Simulate , and from normal proposal distribution , , and , repressively
(4)Now, we must know the acceptance probability to determine which value will be stored and considered as the estimate of the parameter, so we will use the following ratio to calculate the acceptance probability: , , and
(5)We will randomize and generate a random sample from the uniform distribution having a range from 0 to 1
(6)If the generated value , we will consider , if not we will consider
(7)In order to find the estimates for the other two parameters, we must make a repetition for Step (6) but for, and
(8)Increase the loop counter by one such that
(9)In order to find accurate approximation for the estimates, we must make a repetition for the steps from ((3)–(8)), repetitions for obtaining values for the parameters of the proposed distribution, and this sample can be written as the following:

We can find the BEs of the distribution parameters by the aid of the MH algorithm under the SE loss function, as follows:

5. Interval Estimation for the Distribution’s Parameters

This part of the paper was devoted for interval estimation as we performed interval estimation for the distribution parameters according to two methods: the asymptotic and Credible CIs methods.

5.1. Asymptotic Confidence Intervals

Asymptotic CI is the most popular approach to establishing approximate confidence limits for parameters, in which MLEs are used to get the Fisher information matrix , which consists of the second derivative with negative signs for the log-likelihood function, where we substitute with the estimates of the MLEs , where

In order to find the asymptotic variance-covariance matrix, we get the inverse matrix for the Fisher information matrix. We have the vector parameter as .

So, the asymptotic CI for the parameters , , and can be established as follows:where is , and are the parameters of the distribution and is the which is the the standard normal distribution's percentile.

5.2. Highest Posterior Density (HPD) Interval Algorithm

According to Bayesian estimation, we find point and interval estimations so we must find interval estimation for the distribution’s parameter. This interval is called the HPD interval or sometimes the credible intervals. We explained the technique for discovering the HPD interval for . This method has been suggested by [38] (Algorithm 2).

(1)Do steps in Algorithm 1
(2)After obtaining the estimates for the parameters, arrange each estimate in ascending order as , where and , , and
(3)Then, the credible CIs for are given by

6. Simulation Study

As of now, for any distribution, we must evaluate its performance by using different values for its parameters so we must perform a simulation study under both a complete and a Type-I censored sample. The Monte Carlo simulation is conducted in this section to compare the performance of the methods used in the paper and determine the behaviours of the parameters using the MLEs, MPS, and Bayesian estimates of the NEXF parameters under Type-I censored samples, as depending on the R programming. Ten thousand random samples were generated from the NEXF distribution according to the following combinations of parameters.

We used different values for the parameters with different combination as shown in Tables 14 , with varying sample sizes as shown in the following tables and different time point to end the experiment. In the results of the CI, is chosen to be 0.05. We may define the optimal approach as one that minimises bias, mean squared error (MSE), and length of the confidence interval (L.CI) of the estimator. Comparing the results of the point estimation depends on the bias, MSE, and L.CI values. Tables 14 show the different results of simulating the point estimation methods suggested throughout this paper.

6.1. Concluding Remarks Conducted from the Simulation Study

The following observations can be easily conducted from the results conducted from the simulation. We used Algorithms 1 and 2 to get the simulation results.(1)As the sample size increases, the MSE, Bias, and CI length of each of the parameters decrease, which is the consistence property(2)As the censoring time of Type-I (T) increases, the value of the Bias is decreased(3)As the censoring time of Type-I (T) increases, the value of the MSE is also decreased(4)As the censoring time of Type-I (T) increases, the value of the Length of CI is also decreased(5)In most cases, the MPS performs in a better case than MLE under Type-I censored samples(6)We concluded that the credible interval provides the smallest length among all CIs, when the sample is generated under Type-I censored sample(7)We can analyze that by increasing , the MSE and Bias and CI length for the parameter decrease while for increases, in most cases(8)We can analyze that by increasing , the MSE and Bias and CI length for the parameter decrease while for decreases, in most cases(9)We can analyze that by increasing , the MSE and Bias and CI length for the parameter decrease while for decreases, in most cases

7. Data Analysis Using Type-I Censored Data

In this part of the paper, we perform a real-world data analysis for type progressive censoring data. We used the new novel modified KS algorithm to fit the Type-I censored data. Algorithm 3 discusses the steps for fitting Type-I censored data (see [39]).

(1)Assume that the random sample has size with predetermined time at which to end the experiment
(2)This sample is a right censored sample at time
(3)Assume that U(x) is the PDF of the uniform distribution U(0,1)
(4)Assume that F(x) is the CDF of the distribution, i.e., , where donates the order statistics
(5)The modified KS statistic for the type-I censored data can be defined as follows: where is a nonnegative integer that is less than or equal to and where satisfies , and hence
7.1. Model Selection Criteria

The selection of models for specific data is one of the basic tasks of the scientific study in choosing a predictive model from a group of candidate models. Several statistical methods are available to determine the fitness of competing distributions, where the most widely used are the Akaike information criteria (AIC) and the Bayesian knowledge criteria (BIC). The optimal distribution for the real data set may be the one with the lowest values. These methods are determined according to the following formulas:

The AIC is given by

The CAIC is

The BIC is evaluated as follows:

The HQIC issuch that is the MLE log-likelihood function value, and is parameters count in the distribution in the proposed model, and is considered as the size of the sample used in calculations. We take the AIC and BIC tests to demonstrate that the distribution presented is the most appropriate fit for the data. In order to compare between a large number of distributions, we must base such a comparison on certain criteria: one of these information criteria is called the Akaike information criterion (AIC) (see Akaike [40]) though there are other criteria which are called the Bayesian information criterion (BIC) (see Schwarz [41] for more information), and we can also refer to the Hannan–Quinn criterion (for more information on the criterion (HQIC), see Hannan and Quinn [42]), and for more information and last criteria called the consistent Akaike information criterion (CAIC), refer to the study by Bozdogan [43]; all these criteria were used to determine which model among all competing ones is the best for statistical modeling of the data.

7.2. Real Data Set Application for Remission Times

This subsection includes a data analysis to demonstrate the distribution’s performance. We can find these data in Bain and Engelhardt [44]. This data represent how long leukaemia patients survive with a new treatment for each patient. We recorded the remission times of leukaemia patients after using a new drug for treatment. The new medication that leads to remission in leukaemia was given to a group of 40 patients; after seven months (210 days), the trial was then terminated. It is obvious that we have a Type-I censored sample with and , where donates the number of recorder times before the experiments reach an end and , where is the time at which the experiments ended.

To verify that the likelihood function exists and has unique values for its estimates, the existence and uniqueness property is studied as it is very important property that proves that the likelihood function has unique and global maximum roots, and the likelihood function is plotted in Figures 3 and 4 , respectively. Figure 3 confirms the existence of MLEs as the derivative of the log-likelihood function with respect to a certain parameter that intersects the x-axis at a single point. In addition, Figure 4 provides the truth that estimates for the parameters are global maximum roots. Obviously, the log-likelihood function is a decreasing function that crosses the x-axis only once. As a result, we can assert that the log-likelihood function has only one unique root, which is the global maximum.

The results in the tables show the values of the parameters and the goodness of fit criteria values for the proposed distribution in the competition that were used in the comparison; in this section, we introduce the competitive distributions used in comparison with our proposed model. Table 5 is concerned with the parameter estimation of the distribution’s parameters. Table 6 is concerned with the values of AIC, BIC, HQIC, and and CAIC of the distributions. We compare the proposed distribution with the Exponential distribution (E), , Fréchet two-parameter distribution (F), , and the Fréchet three parameters distribution (FTP), .

8. Conclusions and Remarks

A new novel model has been studied and introduced in this paper, and this model called the NEXF distribution. Its statistical properties have been studied and a linear representation is obtained which helps to find the moments and generating functions. The three unknown parameters of the NEXF distribution have been estimated using the classical and Bayesian approaches of estimation under complete and Type-I censoring schemes. Simulations were performed to test the efficiency of the estimation methods. The MH algorithm has a great role for obtaining an approximation for the Bayesian estimates. We can easily conduct from the results tabulated in the simulation section that the Bayesian method performs better than the other methods according to the values of the MSE. From the real data example, we can see that the NEXF distribution is the best fit of data than the E, F, and FTP distributions. We used the modified KS test to determine the goodness of fit for each distribution to the data. Also, we discussed the existence and uniqueness of the log-likelihood function graphically. We proved that the roots maximize the log-likelihood function and also proved that these roots are unique, and at last we proved the superiority of the proposed distribution in modeling and fitting healthcare and medical data.

Data Availability

All data are included within the paper, all references for all data are included in the paper, and all links and references are included in the paper.

Conflicts of Interest

The authors declare that there are no conflicts of interest.

Acknowledgments

This research was funded by the Deanship of Scientific Research at Princess Nourah bint Abdulrahman University through the Fast-track Research Funding Program.