660
Views
2
CrossRef citations to date
0
Altmetric
Research Article

Analysis of uncertainty measure using unified hybrid censored data with applications

ORCID Icon, ORCID Icon &
Pages 1130-1143 | Received 27 Sep 2021, Accepted 11 Dec 2021, Published online: 31 Dec 2021

Abstract

Entropy is a measure of random variable uncertainty that reflects the anticipated quantity of information. In this paper, estimation of Shannon entropy for Lomax distribution, viz unified hybrid censored data are considered. The processes of maximum likelihood and Bayesian estimation procedures are regarded. The Bayesian estimator under balanced squared error, balanced linear exponential, and balanced general entropy loss functions are derived. The purpose of a Monte Carlo simulation study is designed to compare the accuracy of different estimators in the context of specific measures. Real data analysis is employed to confirm the proposed estimators. In summary, we discovered that, according to the findings of the study, the mean squared errors values decrease as the sample size increases. In the majority of cases, Bayesian estimates under balanced linear exponential loss function are more appropriate in terms of simulation out comes than other specified loss functions.

1. Introduction

Due to a lack of funding and time constraints in the majority of life's trials, it is better to stop the test before all of the items fail. Censored samples are the observations that arise from that situation, and there are numerous censoring strategies. The most common types of censorship are type I (T-I) and type II (T-II). The test is completed at a predetermined censoring time in T-I censoring. The test is completed at a fixed number of failures in T-II censorship. The hybrid censoring scheme (HCS) is a combination of T-I and T-II censoring methods that can be characterized as below.

In a life-testing setting, assume there are n identical items with lifetimes are identically distributed and independent. Let X1:n,X2:n,,Xn:n be the ordered failure times of these objects. When a prefixed number of items, 1rn, out of n, fail, or when a prefixed period T(0,) is reached, the test is ended. Epstein [Citation1] was the first to propose this hybrid censoring method, and there are two types: T-I HCS and T-II HCS.

The life-testing experiment in T-I HCS is completed at a random time T1=min(xr:n,T). The disadvantage of T-I HCS is that very few failures occur until the predetermined period T1. To address this flaw, Childs et al. [Citation2] suggested a recently HCS known as T-II HCS, which ensures fixed number of failures and has a termination time T2=max(xr:n,T). Although the T-II HCS ensures a specific amount of failures, detecting failures and completing the life test can take a lengthy time, which is a disadvantage. These techniques were modified by Chandrasekar et al. [Citation3], who introduced generalized T-I HCS (GT-I HCS) and generalized T-II HCS (GT-II HCS).

One fixes k,r(1,2,,n) and time T(0,) in GT-I HCS, where k<r. T=min(xr:n,T), if the kth failure appears before the time T. T=xk:n if the kth failure occurs after the time T. In GT- II HCS, r(1,2,.,n) and T1,T2(0,) are chosen so that T1<T2. T=T1 if the rth failure occurs before time T1; T=Xr:n if there is the rth failure between T1 and T2; otherwise, T=T2. Even while these two new censored sampling systems outperform the previous ones, they nevertheless have flaws. Because there is just one the pre-assigned period T in GT-I HCS, we cannot ensure r failures. There is a chance that the GT-II HCS will not see a few failures until a pre-determined time T2, and so it will have the same trouble as the T-I HCS. To overcome the shortcomings of both systems, Balakrishnan et al. [Citation4] proposed a scheme that combines GT-I HCS and GT-II HCS and is referred to the unified hybrid censoring scheme (UHCS).

Fix the integers r,k(1,2,,n) and the time points T1,T2(0,) in this scheme. T=min(max(xr:n,T1),T2), if the kth failure happened prior to time T1. T=min(xr:n,T2) if the kth failure happens between T1 and T2, and T=xk:n, if the kth failure happens after T2. Using this scheme strategy, we can confirm that the test will be finished in time T2 with at least k failures, if not, exactly k failures. See [Citation5] for more details on how well the difficulties in the previous types were resolved in this manner. As a result, we have the six cases indicated in Figure  under this UHCS:

Figure 1. Unified hybrid censoring scheme.

Figure 1. Unified hybrid censoring scheme.

Case 1: 0<xk:n<xr:n<T1<T2, if this is the case, we will end at T1,

Case 2: 0<xk:n<T1<xr:n<T2, if this is the case, we will end at xr:n,

Case 3: 0<xk,:n<T1<T2<xr:n, if this is the case, we will end at T2,

Case 4: 0<T1<xk:n<xr:n<T2, if this is the case, we will end at xr:n,

Case 5: 0<T1<xk:n<T2<xr:n, if this is the case, we will end at T2,

Case 6: 0<T1<T2<xk:n<xr:n, if this is the case, we will end at xk:n.

Entropy is a key concept in statistics and information theory, and it was first recommended in physics second law of thermodynamics. Shannon [Citation6] re-defined it and introduced the idea of entropy into information theory to quantify information uncertainty. Cover and Thomas [Citation7] built on Shannon's notion, defining differential entropy (or continuous entropy) of a continuous random variable X with probability density function f(x) as follows: (1) H(f)=f(x)logf(x)dx.(1) Entropy measurement is a crucial concept in a variety of fields. More entropy indicates that there is less information in the sample.

Shannon entropy has many applications in many fields see [Citation8,Citation9] in physics and other fields can be found in [Citation10–13].

Entropy in hybrid censoring data studied by Morabbi and Razmkhah [Citation14]. In progressive censoring of T-II, Abo-Eleneen [Citation15] looked at the entropy and the best method. Cho et al. [Citation16] used doubly GT-II HCS to calculate the entropy of the Rayleigh distribution. Liu and Gui [Citation17] investigated entropy estimate for the Lomax distribution using maximum likelihood (ML) and Bayesian methods under generalized progressively hybrid censoring. Under multiple censored data, Hassan and Zaky [Citation18] found the ML estimator of Shannon entropy for the inverse Weibull distribution. The Shannon entropy of the inverse Weibull distribution under progressive first-failure censoring was investigated by Yu et al. [Citation19]. Bantan et al. [Citation20] investigated entropy estimation of inverse Lomax distribution viz multiple censored data.

The study of inference problems associated with entropy measures has recently gained prominence. As previously stated, the UHCS may be able to overcome the restrictions of the GT-I HCS and GT-II HCS. The Lomax distribution, on the other hand, it has sparked scholarly interest in the last decade due to its possible uses in a range of domains, including lifetime data prediction. We could not discover any study on the Shannon entropy measure estimation problem using UHCS in this regard. Consequently, the main purpose of this study is to use the UHCS to derive ML and Bayesian estimators from the Lomax distribution. Three losses are used to create a Bayesian estimator for Shannon entropy. The balanced-squared error loss (BSEL) function, balanced linear-exponential (BLINEX) loss function, and the balanced general entropy (BGE) loss function are used to create a Bayesian estimate for Shannon entropy. The Markov chain Monte Carlo (MCMC) strategy is employed to solve Bayesian entropy estimators with balanced loss functions. A numerical evaluation is used to assess the behaviour of the provided estimates via UHCS. Real data is analysed for more evidence, allowing the theoretical results to be confirmed.

This paper is planned as follows: The Shannon entropy for the Lomax distribution is calculated in Section 2. The ML Shannon entropy estimator is shown in Section 3. Section 4 covers the MCMC method for calculating Bayesian estimators and derives the Bayesian estimator of entropy under various loss functions. Simulation issue and application to real data are given in Sections 5 and 6, respectively. Section 7 brings the article to a close with some last thoughts.

2. Lomax distribution

Lomax (Lo) distribution is one of the important lifetime models. It has been useful in reliability and life testing problems (see, Hassan and Al-Ghamdi [Citation21]). Lo distribution was first introduced by Lomax [Citation22] and it is useful in a variety of domains, including actuarial science and economics. The Lo distribution was applied to income and wealth data see [Citation23,Citation24]. More details about applications of Lo distribution can found in [Citation25,Citation26].

A random variable X is said to have Lo distribution with shape parameter α and scale parameter ϕ if its probability density function (pdf) is given by (2) f(x)=αϕ[1+xϕ](α+1),α>0,ϕ>0,x0.(2) The cumulative distribution function (cdf) of X is given by (3) F(x)=1[1+xϕ]α.(3) Various studies about Lo distribution can be found in the literature by several authors. Some recurrence relations between the moments of record values from the Lo distribution were studied by Balakrishnan and Ahsanullah [Citation27]. On the basis of T-II censored data, Okasha [Citation28] studied E-Bayesian estimate for the Lo distribution. Singh et al. [Citation29] explored Bayesian estimation of Lo distribution under T-II HCS using Lindley's approximation. Statistical inference of Lo distribution using some information measures can be found in [Citation30–32].

Substituting (Equation2) into (Equation1) will give the Shannon entropy of the Lo distribution H(x)=0(α/ϕ)(1+xϕ)(α+1)×log[α1ϕ(1+xϕ)(α+1)]dx.=0(α/ϕ)(1+xϕ)(α+1)×[log(α1ϕ)(α+1)log(1+xϕ)]dx.Hence, (4) H(x)=I1+I2.(4) To compute H(x) in (Equation4), we must calculate I1 and I2 I1=0αϕ(1+xϕ)(α+1)log(αϕ)dx,and I2=0α(α+1)ϕ(1+xϕ)(α+1)log(1+xϕ)dx. Using integration rules we get (5) I1=αlogαϕ01ϕ(1+xϕ)(α+1)dx=logαϕ=log(ϕ)log(α).(5) To obtain I2, let u=1+xϕ then du=dxϕ, so I2 will be I2=α(α+1)ϕ1(u)(α+1)log(u)ϕdu=α(α+1)1(u)(α+1)log(u)du,then, using integration by parts we can find the value of the integral 1(u)(α+1)log(u)du=1α2,then, (6) I2=α(α+1)(1α2)=1+1α.(6) As a result, the Shannon entropy of the Lo distribution looks like this: (7) H(x)=log(ϕ)log(α)+1α+1.(7) As a function of parameters α and ϕ, this is the needed formulation of Shannon entropy of the Lo distribution.

3. Maximum likelihood estimation

Under UHCS, the ML estimation of the Lo distribution is discussed in this section. Assume there are n identical items in a life-testing experiment. Let X1:n,X2:n,,Xn:n represent the ordered failure times of these items, with fixed integer r,k1,2,,n where k<r<n and time points T1,T2(0,) where T1<T2. Then the likelihood function of α and ϕ, for six cases of UHCS is given by (8) L(x_|α,ϕ)=n!(nD)![i=1Df(xi:n)][1F(c)]nD,(8) where D is total number of failures up to time c in the experiment, and its value in the statements given by (9) (D,c)={(D1,T1)for case 1,(r,xr:n)for case 2 and 4,(D2,T2)for case 3 and 5,(k,xk:n)for case 6,(9) where D1 and D2 refer to the number of failures prior to T1 and T2, respectively. Substituting (Equation2) and (Equation3) in (Equation8) we obtain L(x_|α,ϕ)=n!(nD)![(αϕ)Di=1D(1+xi:nϕ)(α+1)]×[1(1(1+cϕ)α)]nD,this equation can be written in the following way: (10) L(x_|α,φ)=E(αϕ)D1i=1D(1+xi:nϕ)×eα[i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ)],(10) where E=n!(nD)!. Taking the logarithm of both sides we get (11) l(x_|α,ϕ)Dlog(αϕ)(α+1)i=1Dlog(1+xi:nϕ)α(nD)log(1+cϕ),(11) where l(x_|α,ϕ) is the logarithm of the likelihood function. Taking derivatives of (Equation11) with respect to α and ϕ, we can get (12) l(x_|α,ϕ)α=Dαi=1Dlog(1+xi:nϕ)(nD)log(1+cϕ),(12) and (13) l(x_|α,ϕ)ϕ=Dϕ+(α+1)i=1Dxi:nϕ2(1+xi:nϕ)+α(nD)cϕ2(1+cϕ).(13) To find the ML estimator of α and ϕ, set Equations (Equation12) and (Equation13) equal to zero and solve them. Therefore, Equation (Equation12) can be written as (14) αˆ=A(ϕˆ),(14) where A(ϕˆ)=Di=1Dlog(1+xi:nϕˆ)+(nD)log(1+cϕˆ),and Equation (Equation13) can be written as (15) Dϕˆ+(αˆ+1)i=1Dxi:nϕˆ2(1+xi:nϕˆ)+αˆ(nD)cϕˆ2(1+cϕˆ)=0,(15) substituting from Equation (Equation14) into (Equation15) to calculate ϕˆ as shown (16) Dϕˆ+(A(ϕˆ)+1)i=1Dxi:nϕˆ2(1+xi:nϕˆ)+A(ϕˆ)(nD)cϕˆ2(1+cϕˆ)=0.(16) Therefore, we can use an iterative procedure to compute the ML estimator of ϕ and then substitute into (Equation14) to find the ML estimator of α. Then we get the ML estimator of H(x), indicated by Hˆ(x) as follows: (17) Hˆ(x)=log(ϕˆ)log(αˆ)+1αˆ+1.(17)

4. Bayesian estimation

Here, we find Bayesian estimators of the unknown parameters α, ϕ, and H(x) based on BSEL, BLINEX, and BGE loss functions. The gamma distribution is used as a conjugate prior for the class of distribution and is also a conjugate prior for Lomax distribution so we assume that α and ϕ are distributed independently as gamma (a1,b1) and gamma (a2,b2) priors, respectively. Then, the prior of α and ϕ is π1(α)(αa11eb1α),α>0,π2(ϕ)(ϕa21eϕb2),ϕ>0.Where we assumed the hyperparameters a1,a2,b1 and b2>0 are constant and known. The joint prior density of α and ϕ is given by (18) π(α,ϕ)αa11ϕa21e(b1α+b2ϕ).(18) The posterior distribution is given by (19) π(α,ϕ|x_)=L(x_|α,ϕ)π(α,ϕ)L(x_|α,ϕ)π(α,ϕ)dαdϕ,(19) where π(α,ϕ) is the joint prior distribution for the parameters α and ϕ, L(x_|α,ϕ) is the likelihood function, and π(α,ϕ|x_) is the posterior distribution for the parameters.

From (Equation10) and (Equation18) we obtain the joint posterior density function as follows: (20) π(α,ϕ|x_)=E1αD+a11ϕa2D1eb2ϕ×eα[b1+(i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ))]×1i=1D(1+xi:nϕ),(20) where E1 is the normalizing constant given by (21) E1=100L(x_|α,ϕ)π(α,ϕ)dαdϕ,(21) then, (22) π(α,ϕ|x_)αD+a11ϕa2D1eb2ϕ×eα[b1+(i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ))]×1i=1D(1+xi:nϕ).(22) The marginal posterior distributions of α and ϕ take the following forms: (23) π1(α|x_)αD+a11eαb10ϕ(a2D1)×eb2ϕ+α[i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ)]i=1D(1+xi:nϕ)dϕ,(23) and (24) π2(ϕ|x_)ϕ(a2D1)eb2ϕi=1D(1+xi:nϕ)0αD+a11×eα[b1+(i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ))]dα.(24) It is obvious from (Equation22) that the posterior density function of α given ϕ is proportional to (25) π1(α|ϕ)αD+a11eα[b1+(i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ))].(25) As a result, the posterior density function of α given ϕ is gamma distribution with shape parameter (D+a1) and scale parameter [b1+(i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ))]. Consequently, samples of α can be easily created using any gamma generating routine.

It is possible to write the posterior density function of ϕ given α as follows: (26) π2(ϕ|α)ϕ(a2D1)eb2ϕ+α[i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ)]i=1D(1+xi:nϕ).(26) Because this equation cannot be reduced analytically to well-known distributions, standard sampling methods cannot be used to sample directly. Using the MCMC method to get estimator under the following balanced loss functions.

4.1. Loss functions

Loss functions are classified according to symmetry criteria into two main types: the first one is a symmetric loss functions such as the squared error loss function. The second type of loss functions is called asymmetric loss functions, one of these functions is the entropy loss function and two types called unbalanced loss function. Balanced loss functions are appealing as they combine proximity of a given estimator δ to both a target estimator δo and the unknown parameter θ which is being estimated see [Citation33] and is defined according to Zellner's formula as follows: (27) Lρ,ω,δo(θ,ρ)=ωρ(δ,δo)+(1ω)ρ(θ,δ),(27) where 0ω1 is a weighted coefficient, ρ(θ,δ) is an arbitrary loss function, δo is a chosen prior estimator of θ, and ρ(δ,δo) an unbalanced loss function for the likelihood function. When ω=0 there are no differences between Bayesian estimators under the balanced and unbalanced loss functions.

In the BSEL loss function ρ(θ,δ)=(δθ)2, then (Equation27) takes the form Lρ,ω,δo=ω(δδo)2+(1ω)(δθ)2,and the Bayesian estimator of H(x) in this case is given by (28) HBS(x_)=ωHˆ(x)+(1ω)E(H(x)|x_).(28) (29) HBS(x_)=ωHˆ(x)+E1(1ω)00H(x)×αD+a11ϕa2D1eb2ϕeαYi=1D(1+xi:nϕ)dαdϕ,(29) where Y=(b1+[i=1Dln(1+xi:nϕ)+(nD)ln(1+cϕ)i=1D])and E1 is the normalizing constant given in (Equation21).

If we choose ρ(θ,δ)=eq(δδo)q(δθ)1,we will get the BLINEX loss function where q represents the shape parameter of the loss function and the behaviour of the BLINEX loss function and BGE loss function changes with the choice of q. Then the Bayesian estimator of H(x) in this case is (30) HBL(x_)=1qln[ωeqδo+(1ω)E(eqθ|x_)],=1qln[eqH(x)b2ϕαYαD+a11ϕa2D1i=1D(1+xi:nϕ)ωeqH(x)+E1(1ω)00×eqH(x)b2ϕαYαD+a11ϕa2D1i=1D(1+xi:nϕ)dαdϕ].(30) If we choose ρ(δ,θ)=(δoθ)q1qln(δoθ),we will get the BGE loss function where q0, and the Bayesian estimator of H(x) in this case is (31) HBG(x_)=[(H(x))qαD+a11ϕa2D1eb2ϕeαYi=1D(1+xi:nϕ)ω(H(x)q+E1(1ω)00(H(x))q×αD+a11ϕa2D1eb2ϕeαYi=1D(1+xi:nϕ)dαdϕ]1q.(31)

4.2. MCMC approach

The MCMC method is used to generate samples from posterior distributions and then compute Bayesian Shannon entropy estimators under balanced loss functions. MCMC schemes come in a wide range of options. Gibbs sampling and more general Metropolis-within-Gibbs samplers are a significant subclass of MCMC methods.

To pull samples from posterior density functions and then compute Bayesian estimators, we suggest the following MCMC technique.

Algorithm

  1. ϕ0=ϕˆ.

  2. Set l=1.

  3. α(l) is obtained from Gamma π1(α|ϕ).

  4. To generate ϕ(l) from π2(ϕ|α) using Metropolis–Hastings, see Metropolis et al. [Citation34].

  5. Calculate α(t) and ϕ(t).

  6. Steps 2–5 should be repeated N times.

  7. Obtain the Bayesian estimators of α and ϕ and compute the entropy function H(x) with respect to the balanced loss functions.

5. Simulation study

In this section, we present some simulation results mainly to compare the performances of the ML estimate (MLE) and Bayesian estimate (BE) of the Shannon entropy for Lo distribution under different losses, in terms of mean squared errors (MSEs).

  1. For given hyperparameters a1,b1,a2 and b2 generate random values of α and ϕ from Equations (Equation25) and (Equation26).

  2. For given values of n and D with the initial values of α and ϕ given in Step 1, we generate random samples from the inverse cdf of Lo distribution and then ordered them.

  3. The MLE of H(x) were calculated for certain values of r, k, T1 and T2 under six cases of UHCS, as explained in Section 3.

  4. The BE of H(x) based on BSEL, BLINEX, and BGE loss functions using the MCMC method has been given, respectively, for specific values of r, k, T1 and T2, as discussed in Section 4.

  5. If θˆ is an estimate of θ then MSE over N samples, where i=1,2,,N, is calculated as follows based on sample j, where j = 1, 2,…,N, then MSE over N samples is given by (32) MSE(θˆ)=i=1N(θˆθ)2N,θˆ=(αˆ,ϕˆ).(32)

  6. Steps 15 are repeated where, N=1000 times and used 11,000 MCMC sample and discard the first 1000 values as “burn-in” and generated a sample from Lo distribution with fixed hyperparameters equals a1=2,b1=0.5,a2=1.7 and b2=1.5, weighted coefficient ω=0.5 and the values of q=0.5,0.5. The estimate studied under two cases (α=1.2,ϕ=0.55,H0(x)=1.053), and (α=0.9,ϕ=3.55,H0(x)=3.48342) and two sample sizes where the first sample n=60,r=50,k=(20,30,40), the second sample n=100,r=80,k=(40,50,65,75), values of T1 and T2 are selected as (T1=1.8,T2=2.4), (T1=2.1,T2=5), (T1=3,T2=5) and (T1=3,T2=8). In each case the estimation of the Shannon entropy (in the first row) and MSE (in the next row) for all cases are calculated as shown in Tables  and .

Table 1. MLE and Bayesian estimate of Shannon entropy based on UHCS under balanced loss functions (α=1.2,ϕ=0.55,H0(x)=1.053) and ω=0.5 and MSE of all estimates.

Table 2. MLE and Bayesian estimate of Shannon entropy based on UHCS under balanced loss functions (α=0.9,ϕ=3.55,H0(x)=3.48342) and ω=0.5 and MSE of all estimates.

Simulated results:

Here are some observations on the Shannon entropy estimates performance as shown in Tables  and .

  • The MSE of MLE and BE of the Shannon entropy decrease when n is increased (Figure ).

  • The Bayes estimate of HBLINEX(x) at q=0.5 is superior than the other Bayes estimates for different values of n in terms of yielding more information as seen in Figure . However, when compared to other Bayes estimates, the Bayes estimate of HBGE(x) at q=0.5 is the worst, resulting in less information.

  • Figure illustrates that for q=0.5, the Bayes estimate of HBLINEX(x) and HBGE(x) is chosen over the others, allowing for better information.

  • The Bayes estimate of HBLINEX(x) at q=0.5 for n=60, are favoured over the others for different values of k. The Bayes estimate of HBGE(x) at q=0.5, on the other hand, is the poorest for various values of k. The MSE of H(x) in MLE decreases when the value of k increases (refer Figure ).

  •  indicates that the Bayes estimate of HBLINEX(x) at q=0.5 is favoured over the others, for selected values of T1 and T2 in terms of least MSE, resulting in reduced variability.

  • The amount of data obtained in the Bayesian estimate are larger than at MLE because the uncertainty in case of BE is less than the uncertainty in MLE.

  • The Shannon entropy BEs drop as T2 is increased for fixed n,r,k and T1. In addition, for fixed n,r,k and T2, the BE of Shannon entropy decreases by increasing T1 in the majority of cases.

  • Whenever T1 or T2 is small, the Bayes estimator being better for selected values of n,r and k in terms of least MSE.

  • A greater amount of data are obtained when the sample size is large because there is a small uncertainty at it.

Figure 2. The MSE of Shannon entropy estimation for different values of sample size.

Figure 2. The MSE of Shannon entropy estimation for different values of sample size.

Figure 3. The MSE of Shannon entropy BEs based on BLINEX and BGE loss functions.

Figure 3. The MSE of Shannon entropy BEs based on BLINEX and BGE loss functions.

Figure 4. The MSE of Shannon entropy estimation under different values of k at n = 60.

Figure 4. The MSE of Shannon entropy estimation under different values of k at n = 60.

Figure 5. The MSE of Shannon entropy estimates at (T1=2.1 and T2=5), (T1=3 and T2=5) and (T1=3 and T2=8) for n=100,k=40.

Figure 5. The MSE of Shannon entropy estimates at (T1=2.1 and T2=5), (T1=3 and T2=5) and (T1=3 and T2=8) for n=100,k=40.

Figure 6. MCMC plots at n=100,T1=3,T2=8,m=65,r=80.

Figure 6. MCMC plots at n=100,T1=3,T2=8,m=65,r=80.

Figures and  display a trace plot of the Shannon entropy for different loss functions with iteration numbers (1000) for n=100,T1=3,T2=8,m=65, and r=80 and for n=100,T1=3,T2=8,m=20,r=50, respectively. The estimations show that all of the generated posteriors fit the theoretical posterior density functions very well, and it is clear that a large MCMC loop produces similar and more efficient outcomes. These plots resemble a horizontal band with no long upward or downward trends, which are convergence markers.

Figure 7. MCMC plots at n=60,T1=3,T2=8,m=20,r=50.

Figure 7. MCMC plots at n=60,T1=3,T2=8,m=20,r=50.

6. Real life data

To demonstrate the proposed methodologies, we look at an application to two real data sets in this section. Nelson [Citation35] provided these statistics, which Abd Ellah [Citation36] discussed. The first data set was obtained at a voltage of 34 kV (minutes), and these data represent the time it takes for an insulating fluid between electrodes to break down (minutes). The 19 times to breakdown are recorded as

Abd Ellah [Citation36] used the Kolmogorov–Smirnov test to evaluate the correctness of the fitted model where the P-value = 0.749683 and statistic value is 0.147477. The estimated pdf and cdf are represented in Figure .

Figure 8. Estimated pdf and cdf of the Lo distribution for first data.

Figure 8. Estimated pdf and cdf of the Lo distribution for first data.

Now we will look at what happens if the data are censored. From the uncensored data set, we generate six artificially UHCD sets as follows:

Case 1: T1=20,T2=35,k=10,r=13 in this situation D=14,c=T1=20.

Case 2: T1=20,T2=35,k=13,r=15 in this situation D=15,c=xr:n=31.75.

Case 3: T1=20,T2=35,k=14,r=18 in this situation D=17,c=T2=35.

Case 4: T1=20,T2=35,k=15,r=17 in this situation D=17,c=xr:n=33.91.

Case 5: T1=20,T2=35,k=15,r=18 in this situation D=17,c=T2=35.

Case 6: T1=20,T2=35,k=18,r=19 in this situation D=18,c=xk,n=36.71.

In these cases, we used ML and a Bayesian estimate of Shannon entropy. We used 11,000 MCMC sample and discarded the first 1000 values as “burn-in” under balanced loss functions (BSLE, BLINEX, BGE) when ω=0.5. Since we do not know anything about the priors, we use a non-informative prior to compute the Bayes estimators, thus we choose a1=0,b1=0,a2=0 and b2=0, as shown in Table .

Table 3. Shannon entropy estimators for first data.

The second data were obtained from a meteorological study by Simpson [Citation37]. The data represent radar-evaluated rainfall from 52 south Florida cumulus clouds. This data set was successfully fitted for the Lo distribution using the Kolmogorov–Smirnov test with a statistic value of 0.0825474 and a P-value =0.841763. The estimated pdf and cdf are represented in Figure .

Figure 9. Estimated pdf and cdf of Lo distribution for second data.

Figure 9. Estimated pdf and cdf of Lo distribution for second data.

Now we will look at what happens if the data are censored. From the uncensored data set, we generate six artificially UHCD sets as follows:

Case 1: T1=120,T2=200,k=25,r=27 in this situation D=28,c=T1=120.

Case 2: T1=120,T2=200,k=25,r=30 in this situation D=30,c=xr:n=147.8.

Case 3: T1=120,T2=200,k=25,r=35 in this situation D=34,c=T2=200.

Case 4: T1=120,T2=200,k=29,r=31 in this situation D=31,c=xr:n=163.

Case 5: T1=120,T2=200,k=29,r=36 in this situation D=36,c=T2=200.

Case 6: T1=120,T2=200,k=35,r=37 in this situation D=35,c=xk,n=244.3.

In these cases, we used ML and a Bayesian estimate of Shannon entropy. We used 11,000 MCMC sample and discarded the first 1000 values as “burn-in” under balanced loss functions (BSLE, BLINEX, BGE) when ω=0.5. Since we do not know anything about the priors, we use a non-informative prior to compute the Bayes estimators, thus we choose a1=0,b1=0,a2=0 and b2=0, as shown in Table . Figures  and  show the trace plot and histogram of the first 1000 MCMC outputs for posterior distribution of H(x) for case 1 based on UHCS for first real and second real data respectively.

Figure 10. The trace plot and histogram of posterior samples for case 1 of UHCS for the first real data.

Figure 10. The trace plot and histogram of posterior samples for case 1 of UHCS for the first real data.

Figure 11. The trace plot and histogram of posterior samples for case 1 of UHCS for the second real data.

Figure 11. The trace plot and histogram of posterior samples for case 1 of UHCS for the second real data.

Table 4. Shannon entropy estimators for second data.

We note from the study of these applications in Tables  and  the amount of data obtained in Bayesian estimation is larger than at MLE because the uncertainty in the case of Bayesian is less than the uncertainty in MLE.

The BE based on BLINEX and BGE loss functions at q=0.5 has a large amount of data where at q=0.5 has a less amount of data because it has a large amount of uncertainty and this is the same result in simulation study.

7. Conclusion and summary

In this paper, the maximum likelihood and Bayesian estimation of Shannon entropy for Lo distribution under UHCS are considered. The Bayesian estimate of Shannon entropy for Lo distribution is obtained based on balanced loss functions (BSEL, BLINEX, BGE). The MCMC method is employed to calculate the Bayesian estimator of Shannon entropy under BSEL, BLINEX, BGE in terms of their mean squared error. Application to real data is provided.

Regarding the simulation results, we conclude that the mean squared error of ML and Bayesian estimates of the Shannon entropy decrease as the sample size increases. The entropy estimates via BLNIEX and BGE loss functions are chosen over the others at q=0.5, allowing for more information. The MLE and BE of Shannon entropy are preferred above the others when T1=1.8, providing more information for fixed values of r,k,T2. The BE has a greater amount of information than MLE since the uncertainty in BE is lower. In future work, one might study the challenge of estimating other measures of uncertainty using the E-Bayesian approach under UHCS. In this study, we used the MCMC method to calculate the Bayesian estimator of Shannon entropy under BSEL, BLINEX, BGE. Other approaches, such as Tierney–Kadane approximation and Lindley approximation will be used along with the MCMC approach under BSEL, BLINEX, BGE or another types of loss functions.

Acknowledgements

The authors would like to express their gratitude to the editor and the anonymous referees for their insightful and constructive remarks, which have significantly improve the paper's substance.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Correction Statement

This article has been republished with minor changes. These changes do not impact the academic content of the article.

Additional information

Funding

This work has no funding from any organization or any person.

References