![MathJax Logo](/templates/jsp/_style2/_tandf/pb2/images/math-jax.gif)
Abstract
In this note we propose a new semi-parametric bootstrap procedure for hypothesis tests about a statistical function and termed bootstrap warping. This procedure was motivated by empirical likelihood and bootstrap tilting techniques. The procedure is computationally efficient and has a fixed number of parameters. We show that the warping procedure has good type I error control and has monotone power as a function of sample size and shift alternatives.
1. Introduction
Let denote an i.i.d. sample drawn from an absolutely continuous population with cumulative distribution function (c.d.f.) denoted as F and corresponding quantile function denoted as
For the application described in this note we are interested in making inferences about a one-dimensional parameter of the form
where in our methodology T(F) denotes a specific smooth statistical functional for measuring expectation and having the form
(1.1)
(1.1)
In our applications we restrict to be a smooth absolutely continuous weighting function such that
is essentially a p.d.f. for
and we assume
is bounded, e.g. suppose the parameter of interest is the population mean, then the statistical functional has the well-known form
with
or alternatively j(u) = 1 and g(x) = x. The classic “bootstrap” estimator of T(F) is given by replacing the c.d.f. F with its empirical counterpart,
in Equation(1.1)
(1.1)
(1.1) or alternatively replacing Q with its empirical counterpart,
where
denotes the floor function and
denotes the ith order statistic. Substituting
into Equation(1.1)
(1.1)
(1.1) for F yields the empirical estimator of T(F), which has the well-known form
(1.2)
(1.2)
Some classic examples for include, kernel density and quantile estimators, sample moment estimators and L-estimators, e.g. see Serfling (Citation1980) for a technical overview of estimators having this form relative to their asymptotic properties.
Now, suppose we are interested in testing a hypothesis about the given statistical functional T(F) having the form versus
or without loss of generality
in a semiparametric fashion. Note that we will touch on two-sided tests later in this note. Popular nonparametric approaches for testing hypotheses of this type are given by the well-known empirical likelihood (EL) method due to Owen (Citation1988) and bootstrap tilting methods such as exponential tilting or other multinomial based resampling schemes, e.g. see Davison and Hinkley (Citation1997).
The nonparametric EL and bootstrap approaches provide the motivation for our new semiparametric testing methodology. The key idea behind the EL and bootstrap tilting approaches is to find the nonparametric maximum likelihood estimator for the probability density function (p.d.f.) f0 and c.d.f F0 given the constraint prescribed under H0, as estimated by its empirical counterpart
where
and the vi parameters corresponding to f0 sum to 1 and are bounded between 0 and 1. The common definition for the vi parameters in the continuous case for the discretized model is given as
where
and
respectively, e.g. see Owen (Citation1988) for the technical argument pertaining to this definition.
In the most common scenario the likelihood under the “unconstrained” alternative hypothesis yields the classic estimates of for simple statistics such as the sample mean. Other weights may occur for functionals corresponding to trimmed estimators, e.g. see Qin and Taso (Citation2002) with respect to the weights for trimmed mean. The weights, vi, under the null hypothesis are generally determined by minimizing a given distance measure such as the Kullback-Leibler distance
where the
vectors
and
Alternatively, one may use constrained maximum likelihood approaches for determining the vector
under H0, e.g. see Vexler and Gurevich (Citation2010) a typical model scenario. We use this idea of a discretized model as a starting point for developing an alternative based on smooth statistical functional inferential procedure using what we term statistical warping as defined in the next section.
In this Section 2 we outline the bootstrap warping procedure and follow this with simulation study in Section 3
2. Bootstrap warping and hypothesis testing
The key features of bootstrap warping, as contrasted with the EL and bootstrap tilting approaches, is that it is a semi-parametric approach and that the number of parameters in the model is reduced from n – 1 to 2, conditional on the observed data, i.e. we “warp” the observed e.d.f. versus treating each discretized segment on the continuum as a parameter. In addition, our resampling scheme follows the classic bootstrap multinomial resampling scheme with cell probabilities of
versus bootstrap tilting, which requires the weights to be determined conditional on the dataset under investigation, thus adding a layer of complexity to the computational components to these problems. The direct benefit of this parameterization is computational ease without suffering the “curse of dimensionality” associated with big data scenarios. This will be described in detail below. Additionally, in terms of future work, covariate adjustments may be made through the warping model parameters, thus extending the utility of this approach to more complex settings.
In terms of testing versus
we need to first define
relative to obtaining an empirical version of the constraint
Towards this end we define the warped empirical estimator of
based on formulation (1.1) as
(2.1)
(2.1)
(2.2)
(2.2)
where
is a weighting function defined more formally below at Equation(2.5)
(2.5)
(2.5) ,
is defined at (1.1) and we define
such that
under H0. For example, if we were interested in testing about
then
(2.3)
(2.3)
and under H0 true we would have
Comment. Note that at Equation(2.3)
(2.3)
(2.3) is used specifically to generate the null distribution for the estimator
at Equation(1.2)
(1.2)
(1.2) and is not meant to be alternative estimator for θ.
The components of the weighting function in Equation(2.2)(2.2)
(2.2) , denoted as
are defined as the c.d.f. of a Kumaraswamy distribution and given as
(2.4)
(2.4)
where
and
The choice of the Kumaraswamy distribution in terms of a weighting function is due to its numerical tractability and flexibility in terms of the relative shapes it contains, i.e. our test will be sensitive to a number of alternatives given H1 via the choice of this weighting function.
Our semi-parametric density utilized within Equation(2.2)(2.2)
(2.2) is now defined as discretized type model similar to what is used in the EL methodology and bootstrap tilting and given as
(2.5)
(2.5)
where
and
and
represents a point mass corresponding to the ith order statistic. The Kumaraswamy distribution was chosen over other candidate distributions, e.g. the beta distribution, due to its well-behaved numerical properties and relatively straightforward parameterization. See Jones (Citation2009) for a detailed description of the Kumaraswamy distribution and a description of its close relationship to the beta distribution. In essence
serves as a standard weighting function such that when α = 1 and β = 1 then
equates to
The test of interest in this note is given as versus
As in EL methods and bootstrap tilting the first step is to maximize the constrained pseudo-likelihood
(2.6)
(2.6)
with respect to α and β and under the constraint
where
is defined at Equation(2.5)
(2.5)
(2.5) . Clearly, α = 1 and β = 1 given H0 is true.
The bootstrap resampling scheme for our inferential method then is as follows:
Calculate the observed test statistic
Obtain
and
from Equation(2.6)
(2.6)
(2.6) .
Generate B nonparametric bootstrap samples of size n, i.e. generate n uniform (0,1) random variables and apply
to those randomly generated uniform variates.
Calculate
from Equation(2.2)
(2.2)
(2.2) , replacing α with
and β with
in step (2), for
Calculate the approximate one-sided bootstrap p-value
where
is the observed estimator defined at Equation(1.2)
(1.2)
(1.2) .
For a test of versus
simply reverse the inequality in step (5) above. For the test
versus
there is an added assumption of the symmetry of the distribution of
under H0. Under this assumption the two-sided p-value is given as
In general, for most of the tests of interest the test statistic will have an asymptotic normal distribution thus most two-sided tests should satisfy approximately a symmetry assumption, i.e. the statistics are based on smooth functions, which then in turn lend themselves to well-behaved and symmetric bootstrap resampling distributions.
As with similar bootstrapping methodologies for inference. e.g. see Davison and Hinkley (Citation1997), the key is that is a consistent estimator of
under H0, which in the methodology presented above holds given α = 1 and β = 1 under H0,
converges to F and
converges to T(F) given the smoothness conditions outlined earlier, which is by definition the statistical function of interest, e.g. see van der Vaart (Citation1998). The large sample proof of this concept is given by the following theorem:
Theorem 1.
Under and as
has a centered bivariate normal distribution with variance-covariance matrix
, where B is the standard maximum likelihood based information matrix associated with the Kumaraswamy density,
, and Σ is the variance-covariance matrix of a 2-dimensional random vector whose components are given by
(2.7)
(2.7)
(2.8) (2.9)
(2.8) (2.9)
where
(2.10)
(2.10)
(2.11)
(2.11)
Proof.
The technical details have been worked out in an elegant fashion for the case of a semi-parametric copula model with marginal distribution functions estimated by the empirical distribution function estimator. The result in Theorem 1 follows directly from the theoretical developments used in the copula approach in Section 4 of the copula paper Genest, Ghoudi, and Rivest (Citation1995) by simply replacing the multivariate copula function with the univariate beta density, which is essentially a special case of the higher dimension copula model. Estimates of the variance-covariance matrix are not as straightforward to obtain and we recommend bootstrap resampling for this purpose.
3. Simulation results
For our simulation study we focused on the trimmed mean with known statistical functional given as
Similar results in terms of behavior hold for moment estimators and kernel estimators and are not presented here. We centered our simulation study for the trimmed mean on the hypothesis test
versus
at type I error rate 0.05 for symmetric distributions with trimming proportions
and samples of size n = 10, 20, 50. For each simulation result we utilized 1000 Monte Carlo resamples with the number of bootstrap resamples set to B = 250. For the exponential distribution we tested
versus
with shifted exponential alternatives. For the γ trimmed mean we can simplify (2.3) such that
(3.1)
(3.1)
where
For power examinations we used shift alternatives of
and δ = 1 different from H0. The Type I error was estimated from the 1000 Monte Carlo resamples as the number of times pboot was less than the Type error of 0.05.
The results of our simulation study are presented in . We see that the type I error is controlled at the nominal level and that fluctuations about that level are primarily due to simulation error. The power is monotone in increase δ and n. As compared with an optimal scenario such as a t-test under normality with and
the one-sample t-test has power of 0.427, 0.695 and 0.967 as compared to the warping powers of 0.369, 0.656 and 0.966 for samples of size n of 10, 20 and 50, which yields relative efficiencies of 86.4%, 94.4% and 99.8%,respectively. By comparison the empirical likelihood approach yields powers of 0.414, 0.640 and 0.960 for samples of size n of 10, 20 and 50 under the same scenarios. However, it should be noted that the Type I error control for the empirical likelihood approach was inflated at 0.084 at
and n = 10, hence in turn the power value of 0.414 is inflated at this same sample size due to a much higher than desired Type I error level as compared to the warping approach.
Table 1. Type 1 error set to 0.05(δ = 0), power () at trimming proportions
and 0.2.
Acknowledgments
This work was supported by Roswell Park Cancer Institute and National Cancer Institute (NCI) grant P30CA016056, NRG Oncology Statistical and Data Management Center grant U10CA180822 and IOTN Moonshot grant U24CA232979-01. We wish to thank the reviewers for their thoughtful comments, which led to an improved version of this work.
References
- Davison, A. C., and D. V. Hinkley. 1997. Bootstrap methods and their applications. Cambridge series in statistical and probabilistic mathematics. New York, NY: Cambridge University Press.
- Genest, C., K. Ghoudi, and L. P. Rivest. 1995. A semiparametric estimation procedure of dependence parameters in multivariate families of distributions. Biometrika 82 (3):543–52. doi:https://doi.org/10.2307/2337532.
- Jones, M. C. 2009. Kumaraswamy’s distribution: A beta-type distribution with some tractability advantages. Statistical Methodology 6 (1):70–81. doi:https://doi.org/10.1016/j.stamet.2008.04.001.
- Owen, A. B. 1988. Empirical likelihood ratio confidence intervals for a single functional. Biometrika 75 (2):237–49. doi:https://doi.org/10.2307/2336172.
- Qin, G., and M. Tsao. 2002. Empirical likelihood ratio confidence intervals for the trimmed mean. Communications in Statistics – Theory and Methods 31 (12):2197–208. doi:https://doi.org/10.1081/STA-120017221.
- Serfling, R. J. 1980. Approximation theorems of mathematical statistics. New York, NY: John Wiley & Sons.
- van der Vaart, 1998. Asymptotic statistics. Cambridge series in statistical and probabilistic mathematics. Cambridge, UK: Cambridge University Press.
- Vexler, A., and G. Gurevich. 2010. Empirical likelihood ratios applied to goodness-of-fit tests based on sample entropy. Computational Statistics and Data Analysis 54 (2):531–45. doi:https://doi.org/10.1016/j.csda.2009.09.025.