709
Views
1
CrossRef citations to date
0
Altmetric
Research Article

On using Shannon entropy measure for formulating new weighted exponential distribution

ORCID Icon, &
Pages 1035-1047 | Received 14 Jan 2022, Accepted 07 Oct 2022, Published online: 31 Oct 2022

Abstract

In this article, a new class of weighted exponential distribution called Entropy-Based Exponential weighted distribution (EBEWD) is proposed. The main idea of the new lifetime distribution is in using the Shannon entropy measure as a weighted function. The mathematical properties of the EBEWD, such as reliability measures, moment measures, variability measures, shapes measures, entropy measures and Fisher’s information, were derived. The unknown parameters of the proposed distribution were estimated using the maximum likelihood estimation method. The efficiency and flexibility of the EBEWD were examined numerically and then by a real-life data application. The results revealed that the EBEWD outperforms some existing distributions in terms of their test statistics.

1. Introduction

The method of weighted distributions as proposed by Fisher’s [Citation1] models of ascertainment on how recorded the probabilities of the events. Weighted distributions are used to select the suitable models for observed data with a technique for fitting models and suitable fits for the data. Rao [Citation2] provides a method of ascertainment concept and how the general formula of weighted distribution is useful for standardizing statistical distributions.

Many researchers start using the concept ofweighted distribution and used the idea of weighted distributions on how effect the original distribution with the new weighted function. Cox [Citation3] suggested using w(x)=x and calling the new distribution by length-biased weighted distribution. Rao [Citation2] used the weight function w(x)=xδ and proposed the so-called size-biased weighted distribution of order  δ. Different weight functions are used in the literature; illustrates some of these weight functions with some related references.

Table 1. Improper weight functions.

The main idea of this paper is to introduce a new weight function that is used in quantifying the maximum amount of information in a random variable called entropy [Citation4] to the context of weighted distributions. Accordingly, the article is organized as follows: the next section introduces the new weigh function. The new exponential probability density function is given in Section 3. The statistical measures of the modified exponential distribution are given from Section 3 to Section 10. Parameter estimation is given in Section 11 while in Section 12 the benefit of fitting data to the new distribution model based on a real case study is discussed. The article ends with a concluding remark section.

2. Proposed weighted function

Let X be a random variable with probability density function (pdf) f(x,θ) where θ is an unknown parameter, then the corresponding weighted density function is defined as follows: (1) g(x,θ)=w(x)f(x,θ)E[w(x)](1) where w(x) is a non-negative weighted function; and E[w(x)] is the expected value of the weighted function. It is assumed that the expected value of the weighted function exists, and it is positive. The problem in formulating the weighted distribution occurs in selecting a suitable weighted function such that the new distribution is more suitable than the original distribution f(x,θ), and provides a better fit for the data than some existing distributions.

The suggested weighted function in this article is entropy, which is defined by Shannon [Citation5] as a negative average of the logarithm of a pdf and called the amount of self-information for a given random variable. Statistically, let X be a continuous random variable with pdf f(x,θ), then the entropy is defined as follows: (2) H(x)=ln(f(x,θ))f(x,θ)dx(2) Based on (2), the suggested weight function will be w(x)=ln(f(x,θ)) Then using this function in (1), we will have the new formulation of the weighted distribution as follows: (3) g(x,θ)=ln(f(x,θ))f(x,θ)E[ln(f(x,θ))] (3)

3. Entropy-based weighted exponential distribution

The exponential distribution is the most known and used continuous distribution for lifetime data applied in many sciences, such as engineering, economics, medicine, biomedical and in many industrial applications. However, the exponential distribution does not provide a significant fit for some real life-time applications, where the hazard rate function and the mean residual life function of exponential distribution are always constant. Therefore, a more suitable distribution is needed; one of the creative ideas is to form a weighted form of the exponential distribution with a given weight function.

Definition 3.1:

A non-negative continuous random variable X is said to have an exponential distribution if it has a pdf and CDF: (4) f(x,λ)={λeλx ; x 0;λ>00 ;otherwise  (4) and (5) F(x,λ)={1eλx;x0;λ>0(5) where λ is the distribution parameter and called the rate of the distribution. When λ=1, then the distribution is the standard exponential.

Corollary 3.1:

Let X be a non-negative continuous random variable X with pdf as given in 4, then the entropy is HS(X)= E[ln(f(x,λ))]=1ln(λ)  This result can be used to develop a new pdf. Therefore, using (1), the entropy-based weighted exponential distribution (EB-WED) will be (6) g(x,λ)= λxln(λ)1ln(λ) λ eλx; x>0;ln(λ)1 (6) The cumulative distribution function CDF of the EB-WED is (7) G(x,λ)=0xg(t,λ)dt = 1(1+λxln(λ))1ln(λ) eλx (7) Note that, the pdf given in (6) is a non-negative real-valued function, which means  g(x,λ)>0. This is true iff λxln(λ)1ln(λ)>0 or λeλx>0 Based on this fact, some restriction on the pdf range depends on the parameter value. Three cases were obtained as follows:

  • if 0<λ1; then g(x,λ)>0 ; 0< x<

  • if  1<λe1; then g(x,λ)>0 ;if ln(λ)λ<x<

  • if λ>e1 ; then g(x,λ)>0 ;if 0<x<ln(λ)λ

For better visualization see the pdf of EB-WEB.

Figure 1. The pdf of EB-WED for different values of λ.

Figure 1. The pdf of EB-WED for different values of λ.

shows that the pdf is skewed to the right. However, when λ belongs to the interval 0<λ<1 or 1<λ<e1, then the pdf has a clear shoulder as λ increases. When λ<e1 the pdf has shrinkages and does not have a clear shoulder. Moreover, the CDF is given in noting that G(x,λ)>0 must be held. 1(1+λxln(λ))1ln(λ)eλx>0

Figure 2. The CDF of EB-WED for different values of λ.

Figure 2. The CDF of EB-WED for different values of λ.

4. Moments and related measures of the EB-WED

The rth moment of the EB-WED random variable is defined as μr=E(Xr)=xrg(x,λ)dx Now, based on three cases the pdf given in (7) the rth moment of X is E(Xr)={0xrg(x,λ)dx; if 0<λ1ln(λ)λxrg(x,λ)dx; if 1<λe10ln(λ)λxrg(x,λ)dx; if λ>e1

Theorem 4.1:

Let X be a random variable following the EB-WED. Then, the rth moment of X is E(Xr)={A1; if 0<λ1A1A2; if 1<λe1A2; if λ>e1 where A1=11lnλ[(r+1)!(r)!lnλλr] A2=λ{Γ(r+2,ln(λ))Γ(r+2)}+ln(λ){Γ(r+1)Γ(r+1,ln(λ))}λr(ln(λ)1)

Proof:

The rth moment of the EB-WED can be obtained using the mathematical expectation: E(Xr)=0xrg(x,λ)dx E(Xr)=0xrλxln(λ)1ln(λ)λeλxdx which can be simplified to E(Xr)=11lnλ[λ0xr+1λeλxdxln(λ)×0xrλeλxdx] which can be simplified to E(Xr)=11lnλ[(r+1)!(r)!lnλλr] End of the proof for case 1.

Now case 3 has E(Xr)=11lnλ[λ0ln(λ)λxr+1λeλxdxln(λ)0ln(λ)λxrλeλxdx] The integration within this expected value can be evaluated easily based on the incomplete gamma function and using the following: 0axr  λ eλxdx =1λr{Γ(r+1)Γ(r+1,aλ)} Therefore, E(Xr)=11lnλ[λ1λr{Γ(r+2)Γ(r+2,ln(λ)λλ)}×ln(λ)1λr{Γ(r+1)Γ(r+1,ln(λ)λλ)} which can be simplified to E(Xr)=1λr(1lnλ)[λ{Γ(r+2)Γ(r+2,ln(λ))}ln(λ){Γ(r+1)Γ(r+1,ln(λ))} which ends the proof of case 3. Noting that Case 2 can be obtained directly by finding the differences between case 1 and case 3.

To reduce the number of cases due to the dependence of the variable range on the parameter value, and for the similarities of steps in getting the results, we will consider the results in this article when 0<x<.

Theorem 4.2:

Assume that the random variable X follows the EB-WED, then the moment-generating function, MX(EBWED) (t), of X is defined as MX(t)=λ1lnλ[[λλlnλ+tlnλ](λt)2]

Proof

The moment-generating function is defined as MX(t)=E(etX)=0etxg(x,λ)dx Therefore, MXEB_WED(t)=0etxλxln(λ)1ln(λ)λeλxdx This can be simplified to MXEBWED(t)=λ1lnλ0(λxln(λ))e(λt)xdx After distributing the terms we have MXEBWED(t)=λ1lnλ0λxe(λt)xdx0ln(λ))e(λt)xdx Then MX(t)=λ1lnλ[[λλlnλ+tlnλ](λt)2] This ends the proof.

Theorem 4.3:

The harmonic mean of EB-WED distribution is defined as HxEBWED=λ1ln(λ)[1Γ(0,)lnλ];

Proof:

The following integral is used to compute the Harmonic mean H=01xg(x,λ)dx Therefore H=01xλxln(λ)1ln(λ)λeλxdx After rearranging the terms, we have H=λ1ln(λ)01x(λxln(λ))eλxdx This can be simplified to H=λ1ln(λ)01x(λxeλxdx)01xln(λ)eλxdx Then H=λ1ln(λ)[1lnλ01xeλxdx]; where 01xeλxdx=Γ(0,) This ends the proof.

Theorem 4.4:

The mode of the EB-WED distribution is xM=lnλ+1λ

Proof

The pdf of the EB-WED distribution is g(x,λ)=λxln(λ)1ln(λ)λeλx and its logarithm ln(g(x,λ))=ln[λxln(λ)1ln(λ)λeλx] After rearranging the terms, we have ln(g(x,λ))=ln(λxln(λ))ln(1ln(λ))+ln(λ)λx Finding the first derivative with respect to x, we have ddxln(g(x,λ))=λλxln(λ)0+0λ Setting to zero and solving for x to get x=λlnλ+λλ2 This ends the proof.

5. Measures of variability and distribution shape of the EB-WED

Measures of variability or relative variability can be obtained using the first two moments, based on Theorem 1:

  • The mean E(X) of the EB-WED is μ=E[X]= 0x g(x,λ)dx = 2ln(λ)λ(1lnλ),lnλ1

  • The second moment E(X2) of the EB-WED is E[X2]= 62lnλλ2(1lnλ) ,lnλ1

Then, the population variance is σ2= E[X2](E[X])2= 24lnλ+(lnλ)2λ2(1lnλ)2 ,lnλ1 Moreover, the coefficient of variation for EB-WED is given by CV=σμ which is equivalent to CV= 24lnλ+(lnλ)2λ2(1lnλ)2ln(λ)2λ(1lnλ) =24lnλ+(lnλ)2 ln(λ)2,lnλ2 The measures of shapes can be obtained based on the third and the fourth moments given by Theorem 1:
  • The third moment E(X3) of the EB-WED is E[X3]=0x3g(x,λ)dx= 246lnλλ3(1lnλ),lnλ1

  • The fourth moment E(X4) of the EB-WED is E[X4]=0x4g(x,λ)dx= 12024lnλλ4(1lnλ),lnλ1

Accordingly, Skewness measures symmetry as

SK=E[X3]3μ σ2μ3σ3  is defined as SKEBWED=[(24lnλ+(lnλ)2)(246lnλ)(1lnλ)2(63lnλ)(24lnλ+(lnλ)2)2]λ2(1lnλ)2 Moreover, the coefficient of kurtosis as Ku=E[x4]4μE[x3]+6σ2E[x2] 3E[x4]σ4  for EB-WED is defined as KuEBWED=[(12024lnλ)(1lnλ)3(84lnλ)(246lnλ)(1lnλ)2+(3612lnλ)(24lnλ+(lnλ)2)(1lnλ)+(36072lnλ)(1lnλ)3](24lnλ+(lnλ)2)8λ4(1lnλ)4 gives a numerical illustration of all statistical measures of the EB-WED with different parameter values. The results indicated that the population means values increase as the value of λ increases, noting that when  λ=1, then EB-WED is the standard exponential distribution. While the values of the variance, coefficient of variation, skewness and kurtosis decrease as λ increases.

Table 2. Numerical values of all statistical measures for the EB-WED.

6. Measures of reliability

In this section, the Measure of Reliability of the EB-WED distribution is derived.

Corollary 6.1:

Suppose that X is a non-negative continuous random variable that has the pdf and CDF as given in 7 and 8, respectively, then

  • The hazard function is H(t,λ)=g(t,λ)1G(t,λ) HEB_EWD(t,λ)=(λtln(λ))λ(ln(λ)1λt)

  • The reliability functions R(t,λ)=P(X>t)=1p(Xt)=1G(t,λ) REB_EWD(t,λ)=(ln(λ)1λt)1ln(λ) eλt

  • The reversed hazard rate function RH(x,λ)=g(x,λ)G(x,λ) RHEB_EWD(x,λ)=(λxln(λ))λeλx(ln(λ)1)+1+λxln(λ)

  • The odds function O(x,λ)=G(x,λ)1G(x,λ) OEB_EWD(x,λ)=[(ln(λ)1)eλx+λxln(λ)+1] (1ln(λ)+λx)

Figures show the behaviour of the EB-WED using its reliability measures.

Figure 3. The hazard function of EB-WED for different values of λ.

Figure 3. The hazard function of EB-WED for different values of λ.

Figure 4. The reliability function of EB-WED for different values of λ.

Figure 4. The reliability function of EB-WED for different values of λ.

Figure 5. The reversed hazard function of EB-WED for different values of λ.

Figure 5. The reversed hazard function of EB-WED for different values of λ.

Figure 6. The odd function of EB-WED for different values of λ.

Figure 6. The odd function of EB-WED for different values of λ.

The above figures indicate that as  λ increases the hazard rate function and the odd function increases while it decreases with the other reliability measures.

7. Stochastic orderings

Let X and Y be two continuous random variables with probability density functions fx(x) and fy(x), and the associated cumulative distributions functions are Fx(x),and Fy(x), then

  • Stochastic order Xst Y if  Fx(x)st Fy(x) for all x;

  • Hazard rate order Xhr Y if  hx(x)hr hy(x) for all x;

  • Mean residual life order  Xmr Y if  mx(x)mr my(x) for all x;

  • Likelihood ratio order Xlr Y if  fx(x)fy(x)  decreases in x.

Thus, Xlr Y Xhr  Xmr Y Xst Y

Theorem 7.1:

Let XWBWED(λ1) and YWBWED(λ2) and corresponding distribution functions  Fx(x) and  Fy(x). If λ1>λ2 then Xlr Y and  Xhr Y, Xmr Y and Xst Y

Proof:

  • Stochastic order: needs to show that  Fx(x)st Fy(x) for all x

For  λ1>λ2 then (1+λ1ln(λ1))1ln(λ1) eλ1x (1+λ2ln(λ2))1ln(λ2) eλ2x  Fx(x)st Fy(x) For all x, then X is stochastically smaller than Y based on stochastic order Xst Y

  • Hazard rate order: needs to show that hx(x)hr hy(x) for all x

For  λ1>λ2 then (λ1tln(λ1))λ1(ln(λ1)1λ1x)(λ2tln(λ2))λ2(ln(λ2)1λ2x)  hx(x)hr hy(x) For all x, then X is smaller in the hazard rate ordering than Y-based hazard rate order  Xhr Y

  • Mean residual life order: needs to show that mx(x)mr my(x) for all x

For  λ1>λ2 then mx(x)=x1Fx(t) dt ; my(x)=x1Fy(t) dt mx(x)my(x)=(1+λ1ln(λ1))1ln(λ1)(1+λ2ln(λ2))1ln(λ2) e(λ2λ1)x,x>0;increases in x mx(x)mr my(x) For all x, then X is smaller in the mean residual life order than Y-based mean residual life order  Xmr Y.

  • Likelihood ratio order: needs to show that fx(x)fy(x)  is a decreasing function of x fx(x)fy(x)=[λ1xln(λ1)1ln(λ1)λ1][λ2xln(λ2)1ln(2)λ2]e(λ2λ1)x,x>0

Now for  λ1>λ2 ddxfx(x)fy(x)=[λ1(lnλ21)e(λ2λ1)x[(λ1λ22λ12λ2)x2+((λ12λ1λ2)lnλ2λ22lnλ1+λ1λ2lnλ1)x+(λ2lnλ1λ1lnλ1λ1)lnλ2+λ2lnλ1]](lnλ11)λ2(xλ2lnλ2)2 ,x<0 X is smaller than Y, fx(x)fy(x) decreases in x, then Y is stochastically greater than X based on likelihood ratio order.

8. Information measures

The information measure of the EB-WED random variable is derived as follows:

Theorem 8.1:

Let X be a continuous random variable that has EB-WED distribution, then the Rényi entropy of X is defined as HR(X)=11αln{(λ21ln(λ))α×(1αα+1ααΓ(α)+Γ(α+1,αln(λ)))}

Proof

Using the Rényi entropy as defined by Rényi [Citation22]

HR(X)=RE(α)=11α lng(x,λ)αdx ;α(0,1),then we have HR(X)=11αln0(λxln(λ)1ln(λ)λeλx)αdx;α(0,1) which can be simplified to HR(X)=11αln0(λxln(λ)1ln(λ))αλαeαλxdx HR(X)=11αln0(λxln(λ))α×(λ1ln(λ))αeαλxdx After rearranging the terms, HR(X)=11αln{(λ1ln(λ))α×0(λxln(λ))αeαλxdx} Now if assume that  u=λxln(λ), then du = λ dx

Replace x with u, then the integral will be HR(X)=11αln{(λ1ln(λ))α×ln(λ)(u)αeα(u+ln(λ))duλ} [HR(X)=11αln{(λ1ln(λ))αeαln(λ)λ×ln(λ)(u)αeαudu} which can be simplified to HR(X)=11αln{(λ1ln(λ))αλ(α+1)×ln(λ)(u)αeαudu} Using integral on gamma function, we have HR(X)=11αln{(11ln(λ))α1λ(ααΓ(α)+Γ(α+1,αln(λ))αΓ(α))} After rearranging the terms HR(X)=11αln{(11ln(λ))α×1λ(1αα+1ααΓ(α)+Γ(α+1,αln(λ)))} This ends the proof.

Theorem 8.2:

Let X be a continuous random variable that has EB-WED distribution, then the Shannon entropy of X is defined as follows: HS(X)=λE(X)ln(λ)E(ln(λxln(λ)1ln(λ))) HS(X)=λ2ln(λ)λ(1lnλ)ln(λ){λ1ln(λ){λ((ln(λ)1)ln[ln(λ)]1)+LogIntegral(λ)}}

Proof

Using Shannon entropy HS(X)=g(x,λ)ln(g(x,λ))dx, then the Shannon entropy of EB-WED can be written in terms of integration: HS(X)=λxln(λ)1ln(λ)λeλxln(λxln(λ)1ln(λ)λeλx)dx Then using the properties of the ln function, we have HS(X)=λxln(λ)1ln(λ)λeλx{ln(λxln(λ)1ln(λ))+ln(λ)λx}dx This is can be written in terms of expected value: HS(X)=λE(X)ln(λ)E(ln(λxln(λ)1ln(λ))) where

  • E(X) = 2ln(λ)λ(1lnλ);

  • E(ln(λxln(λ)1ln(λ))=E{ln(λxln(λ))}ln(1ln(λ);

and

  • E{ln(λxln(λ))}=λxln(λ)1ln(λ)λeλxln(λxln(λ))dx

which can be obtained by setting u = λxln(λ) then du =λ dx

Therefore E{ln(λxln(λ))}=uln(u)1ln(λ)e(u+ln(λ))du which is equivalent to λ1ln(λ){λ((ln(λ)1)ln[ln(λ)]1)+LogIntegral(λ)} Noting that  LogIntegral[λ] is the logarithmic integral function that is defined by LogIntegral(λ)= 0λ1ln(t) dt

This ends the proof.

9. Extreme order statistics

In this section, the pdf of the minimum and the maximum order statistic of the EB-WED is given. Let X(1:n), X(2:n), … , X(n:n) denote the order statistics of a random sample of size n; X1, X2 … , Xn; from a continuous population with pdf g(x,λ) and CDF G(x,λ), then the pdf of the ith order statistic X(i:n) is defined as follows: f(i:n)(x,λ)=n!(i1)!(ni)!×G(x,λ)i1[1G(x,λ)]nig(x,λ) Using 7 and 8 will be derived; the pdf of the minimum and the maximum order statistic will be f(1:n)(x)=n[1G(x)]n1g(x) f(1:n)(x)=[nλ(λxlnλ)((1+λxln(λ))1ln(λ))nenλx]1+λxln(λ) f(n:n)(x)=n[G(x)]n1g(x) f(n:n)(x)=n[1+(1+λxln(λ))ln(λ)1eλx]n1×λxln(λ)1ln(λ)λeλx

10. Fisher’s information

Derive Fisher’s information of the EB-WED random variable based on I(λ)=E(d2dλ2ln(g(x,λ)))

Theorem 10.1:

Let X follow the EB-WED distribution, and then its Fisher’s information, FIEB-WED (λ) is I(λ)=1λ2{1+1(1ln(λ))1(1ln(λ))2[ln(λ)1ln(λ)2+2ln(λ)2]+[1+2(ln(λ)3)2λ2(ln(λ)2)6(ln(λ)1)]}

Proof

Fisher’s information is defined as I(λ)=E(d2dλ2ln(g(x,λ))) Now, the log of the pdf of the EB-WED is ln(g(x,λ))=ln[λxln(λ)1ln(λ)λeλx] ln(g(x,λ))=ln[(λx)ln(λ)]ln[(1ln(λ)]+ln(λ)λx By differentiating the last equation with respect to  λ, then dln(g(x,λ))dλ=x1λλxln(λ)+1λ(1ln(λ))+1λx The second differentiation is d2log(g(x,λ))dλ2=1λ2(λxln(λ))(x1λ)2(λxln(λ))21λ2(1ln(λ))+1λ2(1ln(λ))21λ2 Now I(λ)=E(d2dλ2ln(g(x,λ))) I(λ)=E[1λ2(λxln(λ))(x1λ)2(λxln(λ))21λ2(1ln(λ))+1λ2(1ln(λ))21λ2] I(λ)=1λ2(1ln(λ))1λ2(1ln(λ))2+1λ21λ2E[1(λxln(λ))]+E[(x1λ)2(λxln(λ))2] I(λ)=1λ2(1ln(λ))1λ2(1ln(λ))2+1λ21λ2[1ln(λ)+ln(λ)2ln(λ)1]+E[x22xλ+(1λ)2λ2x22xλln(λ)ln(λ)2] I(λ)=1λ2{1+1(1ln(λ))1(1ln(λ))2[ln(λ)1ln(λ)2+2ln(λ)2]+[1+2(ln(λ)3)2λ2(ln(λ)2)6(ln(λ)1)]}

This ends the proof.

provides some value of Fisher’s information for EB-WED when λ = 0.1, 0.2, … , 1.3. The results indicated that as λ increases as I(λ) decreases, hence the data provide a lot of “information” about λ

Table 3. Fisher’s information values of the EB-WED for λ value.

11. Parameter estimation

In this section, we use the maximum likelihood estimation method to find a point estimator of the unknown parameter in the EB-WED. Suppose that X1, X2,, Xn is a random sample of size n (independent and identically distributed) following the EB-WED, as given in 7. Using MLE, then the likelihood function is given as follows: LEBWED(λ)=i=1nλxiln(λ)1ln(λ)λeλxi which is equivalent to LEBEWD(λ)=λ1ln(λ)1n[λxln(λ)]eλx Finding the logarithm of the likelihood function lnLEBEWD(λ)=ln[λ1ln(λ)]+1n[ln[λxiln(λ)]]λ1nxi By differentiating with respect to λ dlnLEBEWD(λ)dλ=ln(λ)2λ(1ln(λ))+1nxi1λλxiln(xi)1nxi The maximum likelihood λˆ of λ is obtained by equating the above equation to zero. Since the equation is complicated, the point estimator can be found by employing a numerical method to find the point estimator numerically under the following condition: d2lnLEBWED(λ)dλ2 < 0

12. Application

In this section, real data analysis will be considered to study the performance of the new distributions. Then a comparison will be made with some fitted lifetime distributions: The Marshall Olkin Esscher transformed Laplace distribution, Lindley distribution and exponential distribution.

  • The one-parameter Marshall – Olkin Esscher transformed Laplace distribution (MOETLD) was provided by George and George [Citation23] with pdf f(x)=φ1+φ2{eφx,x0eφx,x0

  • Lindley distribution was provided by Lindley [Citation24] with pdf f(x)=θ2(1+θ)(1+x)eθx ; x>0,θ>0

12.1. Goodness of fit measures

Maydeu et al [Citation25]. The goodness of fit is some of the statistical model for summarize the fit of observation, and how perform model comparisons, where in there mathematical formulas have γ the value of the log-likelihood function k is the number of estimated parameters and n is the sample size.

(I) Information Criterion Measures:

  • Bayesian Information Criterion (BIC), Schwarz [Citation26]:

It is a statistical measure used to compare between the performances of a set of models, the lower the value the better the model. BIC=2γ+2kln(n)

  • Akaike Information Criterion (AIC), Akaike [Citation27]:

How well the data are explained by the model suitability and complexity of the model, and the performance of a model is predicted by minimizing the function. The lower the value, the better the model. AIC=2γ+2k

  • Consistent Akaike Information Criterion (CAIC), Bozdogan [Citation28]:

It is a model selection criteria provide an asymptotically how will the model fit the data, and give unbaised estimate of the order of the true model. The model with more parameters fits better than the model that has fewer parameters CAIC=2Ln(γ)+k(ln(n)+1)

  • Hannan – Quinn Information Criterion (HQIC) – Hannan and Quinn [Citation29]: the autoregressive of the model, HQIC is strongly consistent; the lower the value the better the model. HQIC=2γ+2k(ln(ln(n))

  • Maximized log-likelihood (MLL), Heggland andLindqvist [Citation30]:

The likelihood ratio function is expressing the relative likelihood of the data given two competing models. The equation of EB-WED: LEBWED (λ)=λ 1ln(λ)1n[λxln(λ)] eλx (II) Test of Hypothesis

  1. Anderson – Darling Criterion (A–D), Stephens [Citation31]:

Test, if the sample of data has specific distributionand use it in calculating critical value. AD=n1n1n((2i1)ln(Fx(xi))+(2n+12i)ln(1Fx(xi)))

  • Camér – von Mises Criterion (C – M), Cramér [Citation32]:

Compare the goodness of fit of a probability distribution to empirical distributions CM=1n(Fx(xi)2i12n)2+112n

  • Chi-Square test, Chakravarty et al. [Citation33]: A measure of goodness of fit to test if the data are distributed according to the distribution used. Pearsonχ2=i=1n(OiEi)2Ei

Oi = an observed frequency for bin i

Ei = an expected frequency for bin i

Dataset: Murthy et al [Citation34], Aircraft windshield components should have high-performance material and must be able to bear high temperatures and pressures, the company needs to test its product and make it high quality and suitable for different aircraft models. It finds failures items that result from the damaged Windshield, and then the time of repairing faulty items is recorded as service time Wallace et al. [Citation35]. In this study, the data are used to measure the service times for Aircraft windshields. The remaining 63 are service times of windshields. The unit for measurement is 1000 h, with mean 2.08, variance 1.55, min 0.046, max 5.14, skewness 0.439 and kurtosis −0.267, reported as

The distribution fitting results are given in Tables and . In addition, visual comparisons between the empirical distribution and the theoretical distribution are given in .

Figure 7. Comparisons between different distributions for fitting aircraft data.

Figure 7. Comparisons between different distributions for fitting aircraft data.

Table 4. Estimation and information criterions for aircraft data.

Table 5. Distribution fitting tests for the aircraft data.

Based on the values of the AIC, BIC, CAIC and HQIC tests the Entropy-Based Exponential Weighted Distribution provides the best fit for the data. Based on the chi-square test for distributions, the proposed weighted distribution shows a significant result to fit the Aircraft data with p-values of 0.45 for EB-WED. This means the proposed distributions are more suitable than the other distributions in fitting the Aircraft data.

13. Concluding remarks

In this article, we proposed a new exponential type of probability density function called the Entropy-Based Weighted Exponential Distribution. Several of its statistical properties, such as the moments’ measures, variability and distribution measures, reliability, hazard, reversed hazard and odds functions, are provided and studied. Moreover, Fisher information and the Maximum-likelihood estimators of the model parameters are obtained. Finally, a real dataset on the Aircraft’s windshield component is analysed and the results are compared to several life-time models. The results showed that the proposed entropy-based exponential model is more flexible and has a better fit for the suggested data.

In this article, we used the Shannon entropy to formulate the weighted distribution. The proposed distribution could be considered a new value distribution that could be used for fitting real-life data. Note that dealing with entropy measures in formulating a new lifetime distribution is not an easy task; however, the window of further research is still open, for example, one can use other entropy measures such as Renyi or Tsallis in formulating the weighted distribution; and the results could be used to compare the new suggested models with the proposed model in this article in terms of model flexibility and bathtub and other reliability measures.

Disclosure statement

No potential conflict of interest was reported by the author(s).

References

  • Fisher RA. The effects of methods of ascertainment upon the estimation of frequencies. Ann Eugen. 1934;6(1):13–25.
  • Rao CR. On discrete distributions arising out of methods of ascertainment. In: Patil GP, editor. Classical and contagious discrete distribution. Calcutta: Pergamon Press and Statistical Publishing Society; 1965, 27(2); p. 311–324.
  • Cox DR. Some sampling problems in technology. In: Johnson NL, Smith H, editor. New developments in survey sampling. New York: John Wiley and Sons; 1968. p. 506–527.
  • Al-Nasser AD, Al-Omari A, Ciavolino E. A size-biased Ishita distribution and application to real data. Qual Quant. 2019;53(1):493–512.
  • Shannon CE. A mathematical theory of communication. Reprinted with Corrections from the Bell Syst Tech J. 1948;27(3):379–423.
  • Rezzoky A, Nadhel Z. Mathematical study of weighted generalized exponential distribution. Journal of the College of Basic Education. 2015;21(89):165–172.
  • Mir K, Ahmed A, Reshi J. Structural properties of length biased Beta distribution of first kind. American Journal of Engineering Research. 2013;2:01–06.
  • Mudasir S, Ahmad S. Structural properties of length-biased Nakagami distribution. International Journal of Modern Mathematical Sciences. 2015;13(3):217–227.
  • Shi X, Broderick O, Pararai M. Theoretical properties of weighted generalized Rayleigh and related distributions. Theoretical Mathematics and Applications. 2012;2:45–62.
  • Ye Y, Oluyede B, Pararai M. Weighted generalized Beta distribution of the second kind and related distributions. Journal of Statistical and Econometric Methods. 2012;1:13–31.
  • Das K, Roy T. Applicability of length biased generalized Rayleigh distribution. Advances in Applied Science Research. 2011;2:320–327.
  • Jabeen S, Jan T. Information measures of size biased generalized Gamma distribution. International Journal of Scientific Engineering and Applied Science. 2015;1:2395–3470.
  • Aleem M, Sufyan M, Khan N. A Class of Modified weighted Weibull distribution and its properties. American Review of Mathematics and Statistics. 2013;1:29–37.
  • Castillo J, Casany M. Weighted Poisson distribution for over dispersion and under dispersion solutions. Annals of Institute of Statistical Mathematics. 1998;50:567–585.
  • Larose D, Dey D. Weighted distributions viewed in the context of model selection: A bayesian perspective. Test. 1996;5(1):227–246.
  • Mahfoud M, Patil G. On weighted distributions. In: Kallianpur G. et al. editors. Statistics and probability essays in honor of C. R. Rao. Amsterdam: North-Holland Publishing Company. 1982; p. 479–492.
  • Al-Kadim A, Hussein A. New proposed length biased weighted exponential and rayleigh distributions with application. Mathematical Theory and Modeling. 2014;4(7):137–52.
  • Al-Kadim K, Hantoosh A. Double weighted distribution and double weighted exponential distribution. Mathematical Theory and Modeling. 2013;3(5):124–34.
  • Bashir S, Rasul M. Some properties of the weighted Lindley distribution. International Journal of Economics and Business Review. 2015;3:11–17.
  • Modi K. Length-biased weighted Maxwell distribution. Pakistan Journal of Statistics and Operation Research. 2015;11:465–472.
  • Kharazmi O, Mahdavi A, Fathizadeh M. Generalized weighted exponential distribution. Communications in Statistics – Simulation and Computation. 2014;44(6):1557–1569.
  • Renyi A. On measures of entropy and information. In: Jerzy Neyman editor. Proceedings of the 4th Berkeley symposium on mathematical statistic and probability. Vol. I. Contributions to Probability and Theory. Berkeley: University of California Press; 1961. p. 547–561.
  • George D, George S. Marshall–Olkin Esscher transformed Laplace distribution and processes. Braz J Prob Stat. 2013;27(2):162–184.
  • Lindley DV. Fiducial distributions and Bayes’ Theorem. J R Stat Soc. 1958;20:102–107.
  • Maydeu A, Garci A, Forero C. Goodness-of-Fit testing. International Encyclopedia of Education. 2010;7:190–196.
  • Schwarz G. Estimating the dimension of a model. Ann Stat. 1978;6(2):461–464.
  • Akaike H. A New look at the statistical model identification. IEEE Trans Autom Control. 1974;19(6):716–723.
  • Bozdogan H. Model selection and Akaike’s information criterion (AIC): the general theory and its analytical extensions. Psychometrika. 1987;52(3):345–370.
  • Hannan EJ, Quinn BG. The determination of the order of an autoregression. J R Stat Soc B Methodol. 1979;41(2):190–195.
  • Heggland K, Lindqvist BH. A nonparametric monotone maximum likelihood estimator of time trend for repairable systems data. Reliab Eng Syst Saf. 2007;92:575–584.
  • Stephens MA. EDF statistics for goodness of Fit and some comparisons. J Am Stat Assoc. 1974;69:730–737.
  • Cramér H. On the composition of elementary errors. Scand Actuar J. 1928;1928(1):13–74.
  • Laha C, Laha RG. Handbook of methods of applied statistics. New York: John Wiley and Sons; 1967. 1; p. 392–394.
  • Murthy DNP, Min X, Jiang R. Weibull models. New York: John Wiley and Sons; 2004. 1047(22).
  • Blischke WR, Prabhakar Murthy DN. Reliability: modeling, prediction, and optimization. New York: John Wiley and Sons; 2011.