3,478
Views
0
CrossRef citations to date
0
Altmetric
Theory and Methods

Confidently Comparing Estimates with the c-value

ORCID Icon, ORCID Icon &
Pages 983-994 | Received 18 Feb 2021, Accepted 30 Oct 2022, Published online: 24 Feb 2023

Abstract

Modern statistics provides an ever-expanding toolkit for estimating unknown parameters. Consequently, applied statisticians frequently face a difficult decision: retain a parameter estimate from a familiar method or replace it with an estimate from a newer or more complex one. While it is traditional to compare estimates using risk, such comparisons are rarely conclusive in realistic settings. In response, we propose the “c-value” as a measure of confidence that a new estimate achieves smaller loss than an old estimate on a given dataset. We show that it is unlikely that a large c-value coincides with a larger loss for the new estimate. Therefore, just as a small p-value supports rejecting a null hypothesis, a large c-value supports using a new estimate in place of the old. For a wide class of problems and estimates, we show how to compute a c-value by first constructing a data-dependent high-probability lower bound on the difference in loss. The c-value is frequentist in nature, but we show that it can provide validation of shrinkage estimates derived from Bayesian models in real data applications involving hierarchical models and Gaussian processes. Supplementary materials for this article are available online.

1 Introduction

Modern statistics provides an expansive toolkit of sophisticated methodology for estimating unknown parameters. However, the abundance of different estimators often presents practitioners with a difficult challenge: choosing between the output of a familiar method (e.g., a maximum likelihood estimate (MLE)) and that of a more complicated method (e.g., the posterior mean of a hierarchical Bayesian model). From a practical perspective, abandoning a familiar approach in favor of a newer alternative is unreasonable without some assurance that the latter provides a more accurate estimate. Our goal is to determine whether it is safe to abandon a default estimate in favor of an alternative, and to provide an assessment of the degree of confidence we should have in this decision.

Traditionally decisions between estimators are based on risk, the loss averaged over all possible realizations of the data with respect to a pre-specified likelihood model (Lehmann and Casella, Citation2006, chap. 4–5). We note two limitations of using risk. First, it is rare that one estimator within a given pair will have smaller risk across all possible parameter values. Instead, it is more often the case that one estimator will have smaller risk for some unknown parameter values but larger risk for other parameter values. Second, one estimator may have lower risk than another but incur higher loss on a majority of datasets; see Appendix S2 for an example in which an estimator with smaller risk has larger loss on nearly 70% of simulated datasets.

In this work we propose a framework for choosing between estimators based on their performance on the observed dataset rather than their risk. Specifically, we introduce the “c-value” (“c” for confidence in the new estimate), which we construct using a data-dependent high-probability lower bound on the difference in loss. We show that it is unlikely that simultaneously the c-value is large and the alternative estimate has larger loss than the default. For the c-value to be useful, it must meet two desiderata:

  1. The c-value must not frequently guide practitioners to incorrectly report the alternative estimate when the default estimate has smaller loss.

  2. The c-value must, in some cases, allow one to correctly identify that the alternative estimate has smaller loss.

We demonstrate that the c-value meets the first desideratum with theory showing how to use the c-value to select between two estimates in a principled, data-driven way. Critically, the c-value requires no assumptions on the unknown parameter; our guarantees hold uniformly across the parameter space. We demonstrate that the c-value can meet the second desideratum with case studies; we provide an overview of these next as motivating examples, and then proceed to present our general methodology.

Shrinkage estimates on educational testing data

We revisit Hoff’s (Citation2021) estimates of average student reading ability at several schools in the 2002 Educational Longitudinal Study. These estimates are obtained from a hierarchical Bayesian model that “shares strength” by partially pooling data across similar schools. Hoff’s (Citation2021) analysis relied on a simplifying and subjectively chosen prior. A practitioner might wonder whether the resulting estimates are more accurate than the MLE in terms of squared error loss. As we will see, a large c-value provides confidence that Hoff’s estimate is indeed more accurate. We additionally consider a clearly inappropriate prior and verify that our methodology does not always favor more complex alternative estimators. Although these estimates have a Bayesian provenance, the use of the c-value to justify these estimates does not require subjective belief in the prior.

Estimating violent crime density at the neighborhood level

Considerable empirical evidence links a community’s exposure to violent crime and adverse behavioral, mental, and physical health outcomes among its residents (Buka et al., Citation2001; Kondo et al., Citation2018). Although overall violent crimes rates in the United States have decreased over the last two decades, there is considerable variation in time trends at the neighborhood level (Balocchi and Jensen, Citation2019; Balocchi et al., Citation2022). A critical first step in understanding what drives neighborhood-level variation is accurate estimation of the actual amount of violent crime that occurs in each neighborhood.

Typically, researchers rely on the reported counts of violent crime aggregated at small spatial resolutions (e.g., at the census tract level). However, in light of sampling variability due to the relative infrequency of certain crime types in small areas, it is natural to wonder if auxiliary data could be used to improve estimates of violent crime incidence.

As a second application of our framework, we analyze the number of violent crimes reported per square mile in several neighborhoods in the city of Philadelphia. Our analysis suggests that one can obtain improved estimates of the violent crime density by using a shrinkage estimate that incorporates information about nonviolent crime incidence. Further c-value analysis reveals that leveraging spatial information on top of nonviolent incidence does not provide additional improvement.

Gaussian process kernel choice: Modeling ocean currents

Accurate estimation of ocean current dynamics is critical for forecasting the dispersion of oceanic contaminations (Poje et al., Citation2014). While it is commonplace to model ocean flow dynamics at or above the mesoscale (roughly 10 km), Lodise et al. (Citation2020) have recently advocated modeling dynamics at both the mesoscale and the submesoscale (roughly 0.1–10 km). They specifically proposed a Gaussian process model that accounts for variation across multiple resolutions to estimate ocean currents from positional data taken from hundreds of free-floating buoys.

In a third application of our framework, we find that the multi-resolution procedure produces a large c-value, indicating that accounting for variation across multiple scales enables more accurate estimates than are obtained when accounting only for mesoscale variation.

1.1 Organization of the Article and Contributions

We formally present our general framework and define the c-value in Section 2. In Section 2.1 we highlight similarities and differences between our framework and existing work on preliminary testing and post-selection inference. Our approach to computing c-values depends on the availability of high-confidence lower bounds on the difference in the losses of the two estimates that holds uniformly across the parameter space. Sections 3–5 provide these bounds for several models and classes of estimators for squared error loss. In Section 3, we illustrate our general strategy in the canonical normal means problem. Then, in Section 4, we generalize this strategy to compare affine estimates of normal means with correlated observations. Section 5 shows how to extend the framework to cover two nonlinear cases: a nonlinear shrinkage estimator and regularized logistic regression. We provide simulations validating our approach in these settings. We apply our framework to the aforementioned motivating examples in Section 6. In our discussion in Section 7, we outline ways to extend our framework beyond the estimates considered here. Software that implements the c-value computation, and code that reproduces our analyses is available at: https://github.com/blt2114/c_values.

2 Introducing the c-value

We now describe our approach for quantifying confidence in the statement that one estimate of an unknown parameter is superior to another. We begin by introducing some notation and building up to a definition of the c-value, before stating our main results. This development is very general, and we defer practical considerations to the subsequent sections. We include proofs of the results of this section in Appendix.

Suppose that we observe data y drawn from some distribution that depends on an unknown parameter θ. We consider deciding between two estimates, θ̂(y) and θ*(y), of θ on the basis of a loss function L(θ,·). Our focus is on asymmetric situations in which θ̂(·) is a standard or more familiar estimator while θ*(·) is a less familiar estimator. For simplicity, we will refer to θ̂(·) as the default estimator and θ*(·) as the alternative estimator.

We next define the “win” obtained by using θ*(y) rather than θ̂(y) as the difference in loss, W(θ,y):=L(θ,θ̂(y))L(θ,θ*(y)). While a typical comparison based on risk would proceed by taking the expectation of W(θ,y) over all possible datasets drawn for fixed θ, we maintain focus on the single observed dataset. Notably, the win is positive whenever the alternative estimate achieves a smaller loss than the default estimate. As such, if we knew that W(θ,y)>0 for the given dataset y and unknown parameter θ, then we would prefer to use the alternative θ*(y) instead of the default θ̂(y).

Since θ is unknown, determining whether W(θ,y)>0 is impossible. Nevertheless, for a broad class of estimators, we can determine whether the win is positive with high probability. To start, we construct a lower bound, b(y,α), depending only on the data and a pre-specified level α[0,1], that satisfies for all θ (1) Pθ[W(θ,y)b(y,α)]α.(1)

For values of α close to 1, b(y,α) is a high-probability lower bound on the win that holds uniformly across all possible values of the unknown parameter θ. Loosely speaking, if b(y,α)>0 for some α close to 1, then we can be confident that the alternative estimate has smaller loss than the default estimate.

To make this intuition more precise, we define a measure of confidence that θ*(y) is superior to θ̂(y). We call our measure the c-value c(y): (2) c(y):=infα[0,1]{α|b(y,α)0}.(2)

The c-value marks a meaningful boundary in the space of confidence levels; it is the largest value such that for every α<c(y), we have confidence α that the win is positive.

Remark 2.1.

An alternative definition for the c-value is c+(y)=supα[0,1]{α|b(y,α)0}. Although c+(y)=c(y) when b(y,·) is continuous and strictly decreasing in α, c+(·) may be overconfident otherwise. We detail a particularly pathological example in Appendix S3.

Our first main result formalizes the interpretation of c(y) as a measure of confidence.

Theorem 2.2.

Let b(·,·) be any function satisfying the condition in EquationEquation (1). Then for any θ and α[0,1] and c(y) as defined in EquationEquation (2), (3) Pθ[W(θ,y)0andc(y)>α]1α.(3)

The result follows directly from the definition of c(·) and the condition on b(·,·). Informally, Theorem 2.2 assures us that it is unlikely that simultaneously (A) the c-value is large and (B) θ*(y) does not provide smaller loss than θ̂(y). Just as a small p-value supports rejecting a null hypothesis, a large c-value supports abandoning the default estimate in favor of the alternative.

The strategy described above necessarily uses the data twice, once to compute the two estimates and once more to compute the c-value to choose between them. Accordingly, one might justly ask how such double use of the data affects the quality of the resulting procedure. To address this question, we formalize this two-step procedure with a single estimator, (4) θ(y,α):=1[c(y)α]θ̂(y)+1[c(y)>α]θ*(y).(4)

θ(y,α) picks between the two estimates θ̂(y) and θ*(y) based on the value c(y) and a pre-specified level α[0,1]. We can characterize the possible outcomes when using θ(·,α) with a contingency table (), where rows correspond to the estimate with smaller loss, and the columns correspond to the reported estimate.

Table 1 Contingency table with possible outcomes when using the two-stage estimator θ(·,α). θ(·,α) controls the probability of the boldface event (Theorem 2.3).

Recall that we are interested in an asymmetric situation where the alternative estimator is less familiar than the default estimator. This asymmetry makes desirable the reassurance that θ(·,α) does not incur greater loss than θ̂(·). As such, we focus on the upper right hand entry of the table. Our second main result formalizes that when we use θ(·,α) with α close to 1, the probability of the event represented by this table entry is small.

Theorem 2.3.

Let b(·,·) be any function that satisfies the condition in EquationEquation (1). Then for any θ and α[0,1], (5) Pθ[L(θ,θ(y,α))>L(θ,θ̂(y))]1α.(5)

Overview of the remainder of the article

The c-value is useful insofar as the lower bound b(y,α) is sufficiently tight and readily computable. It remains to show that such practical bounds exist. A primary contribution of this work is the explicit construction of these bounds in settings of practical interest. In what follows, we (A) illustrate one approach for constructing and computing b(y,α), (B) explore our proposed bounds’ empirical properties on simulated data, and (C) demonstrate their practical utility on real-world data.

2.1 Related Work

Hypothesis testing, p-values, and pretest estimation

Our proposed c-value bears a resemblance to the p-value in hypothesis testing, but with a few key differences. Indeed, just as a small p-value can support rejecting a simple null hypothesis in favor of a possibly more complex alternative, a large c-value can support rejecting a familiar default estimate in favor of a less familiar alternative. Furthermore both tools provide a frequentist notion of confidence based on the idea of repeated sampling. From this perspective, the two-step estimator θ(·,α) resembles a preliminary testing estimator. Preliminary testing links the choice between estimators to the outcome of a hypothesis test for the null hypothesis that θ lies in some pre-specified subspace (Wallace Citation1977).

The similarities to hypothesis testing go only so far. Notably, we consider decisions made about a random quantity, W(θ,y). Hypothesis tests, in contrast, concern only fixed statements about parameters, with nulls and alternatives corresponding to disjoint subsets of an underlying parameter space (Casella and Berger Citation2002, Definition 8.1.3). Our approach does not admit an interpretation as testing a fixed hypothesis.

Nevertheless, the connection to p-values can help us understand some limitations of the c-value. First, just as hypothesis tests may incur Type II errors (i.e., failures to reject a false null), for certain models and estimators there may be no bound b(·,·) that consistently detects improvements by the alternative estimate. Accordingly, the two stage estimator θ(·,α) does not control the probability that we report the default estimate when the alternative in fact has smaller loss. In such situations, our approach may consistently report the default estimate even though it has larger loss. Second, even if good choices of b(·,·) exist, it could be challenging to derive them analytically. This analytical challenge is reminiscent of difficulties for hypothesis testing in many models, wherein conservative p-values that are stochastically larger than uniform under the null are used when analytic quantile functions are unavailable. Third, we note that it may be tempting to interpret a c-value as the conditional probability that an alternative estimate is superior to a default; however, just as it is incorrect to interpret a p-value as a probability that the null hypothesis is true, such an interpretation for a c-value is also incorrect.

Post-selection inference

In recent years, there has been considerable progress on understanding the behavior of inferential procedures that, like θ(·,α), use the data twice, first to select amongst different models and then again to fit the selected model. Important recent work has focused on computing p-values and confidence intervals for linear regression parameters that are valid after selection with the lasso (Lockhart et al. Citation2014; Lee et al. Citation2016; Taylor and Tibshirani Citation2018) and arbitrary selection procedures (Berk et al. Citation2013). Somewhat more closely related to our focus on estimation are Tibshirani and Rosset (Citation2019) and Tian (Citation2020), which both bound prediction error after model selection. Unlike these papers, which study the effects of selection on downstream inference, we effectively perform inference on the selection itself.

3 Special Case: c-values for Estimating Normal Means

In this section, we derive a bound b(y,α) and compute the c-value in a simple case: we compare a certain class of shrinkage estimators to maximum likelihood estimates (MLE) of the mean of a multivariate normal from a single vector observation (i.e., the normal means problem). Our goal is to illustrate a simple strategy for lower bounding the win that we will later generalize to more complex estimators and models. In Section 3.1, we define the model and the estimators that we consider. In Section 3.2, we introduce our lower bound b(·,·) and present a theorem that guarantees this bound satisfies EquationEquation (1). Then, in Section 3.3, we examine the resulting c-value empirically and study the performance of the estimator θ(·,α) that chooses between the default and alternative estimators based on the c-value (EquationEquation (4)). Several details, including the proof of Theorem 3.1, are left to Appendix S4.

3.1 Normal Means: Notation and Estimates

Let θRN be an unknown vector and consider estimating θ from a noisy vector observation y=θ+ϵ where ϵN(0,IN) under squared error loss L(θ,θ̂):=||θ̂θ||2. For simplicity, we focus on the case of isotropic noise with variance one; we remove this restriction in Section 4. For our demonstration, we take the MLE θ̂(y)=y to be the default estimate. As the alternative estimator, we consider a shrinkage estimator that was first studied extensively by Lindley and Smith (Citation1972), θ*(y)=y+τ2y¯1N1+τ2,where 1N is the vector of all ones, τ>0 is a fixed positive constant, and y¯:=N11Ny is the mean of the observed yn’s. Operationally, θ*(y) shrinks each coordinate of the MLE toward the grand mean y¯.

3.2 Construction of the Lower Bound

To lower bound the win, we first rewrite θ*(y)=θ̂(y)Gy where G:=(1+τ2)1P1 and P1:=INN11N1N is the projection onto the subspace orthogonal to 1N. The win in squared error loss may then be written as (6) W(θ,y):=||θ̂(y)θ||2||θ*(y)θ||2=2ϵGyGy2.(6)

Observe that we can compute Gy directly from our data. As a result, in order to lower bound the win W(θ,y), it suffices to lower bound 2ϵGy. As we detail in Appendix S4.1, 2ϵGy follows a scaled and shifted noncentral Chi-squared distribution, 2ϵGy21+τ2[χN12(14P1θ2)14P1θ2],where χN12(λ) denotes the noncentral Chi-squared distribution with N – 1 degrees of freedom and non-centrality parameter λ. Thus, for any α(0,1) and any fixed value of P1θ2, (7) W(θ,y)21+τ2FN11(1α;14P1θ2)P1θ22(1+τ2)Gy2(7) with probability α, where FN11(1α;λ) denotes the inverse cumulative distribution function of χN12(λ) evaluated at 1α. Were P1θ2 known, the right hand side of EquationEquation (7) would immediately provide a valid bound. However, since P1θ2 is not typically known, we use the data to address our uncertainty in this quantity. We obtain our bound by forming a one-sided confidence interval for P1θ2 that holds simultaneously with EquationEquation (7).

Bound 3.1 (Normal means: Lindley and Smith estimate vs. MLE). Observe y=θ+ϵ with ϵN(0,IN) and consider θ̂(y)=y versus θ*(y)=(y+τ2y¯1N)/(1+τ2). We propose (8) b(y,α):=infλ[0,U(y,1α2)]{21+τ2FN11(1α2;λ4)λ2(1+τ2)P1y2(1+τ2)2}(8) as an α-confidence lower bound on the win, where (9) U(y,1α2):=infδ>0{δ|P1y2FN11(1α2;δ)}(9) is a high-confidence upper bound on P1θ2.

Bound 3.1 relies on a high-confidence upper bound on P1θ2, but a two-sided interval could in principle provide a valid bound as well. In Appendix S4.3 we provide an intuitive justification for the choice of an upper bound. Theorem 3.1 justifies the use of Bound 3.1 for computing c-values.

Theorem 3.1.

Define c(y):=infα[0,1]{α|b(y,α)0} for b(·,·) in Bound 3.1. Then c(y) is a valid c-value, satisfying the guarantees of Theorems 2.2 and 2.3.

Remark 3.2

(Computability of the bound). EquationEquation (8) in Bound 3.1 can be readily computed. Notably, many standard statistical software packages provide numerical approximation to noncentral χ2 quantiles. Further, the one-dimensional optimization problems in EquationEquations (8) and Equation(9) can be solved numerically.

Remark 3.3

(Unknown variance). For cases when the noise variance σ2 is unknown but a confidence interval is available, one can adapt the procedure above by replacing b(y,α) with its infimum with respect to σ2 over the confidence interval and reducing the confidence level α accordingly.

Remark 3.4.

The alternative estimator θ*(y) considered in this section is the posterior mean of θ corresponding to the hierarchical prior θ|μN(μ1N,τ2IN) with further improper hyper-prior on μ. This prior encodes a belief that θ lies close to the one-dimensional subspace spanned by 1N. Using a similar approach to the one above, we can derive lower bounds on the win for a more general class of estimators that shrink the MLE toward a pre-specified D-dimensional subspace. See Appendix S4.4 for details and an application to a real dataset on which a large computed c-value indicates an improved estimate.

3.3 Empirical Verification

To explore the empirical properties of Bound 3.1, we simulated 500 datasets with N = 50 as yN(θ,IN) for each of several values of θ. For each simulated dataset y, we computed the win W(θ,y), the proposed lower bound b(y,α), and the c-value c(y). Conveniently, for this likelihood, the distributions of W(θ,y) and b(y,α) depend on θ only through N12P1θ. Consequently, we can exhaustively assess how our procedure behaves for different θ by varying this norm. Throughout our simulation study, we fixed τ=1. With larger τ, the alternative θ* behaves more similarly to the default θ̂, but the qualitative properties of the c-value and estimators remain similar.

We first checked that the empirical probability that the win W(θ,y) exceeded the bound b(y,α) in Bound 3.1 was at least as large as the nominal probability α (). Across various choices of N12P1θ, we see that b(·,α) is conservative, typically providing higher than nominal coverage. Surprisingly, the gap between the actual and nominal coverages does not seem to depend heavily on θ, suggesting we could potentially obtain a tighter bound by calibrating b(y,α) to its actual coverage.

Fig. 1 Bound calibration and the two-stage estimator for a hierarchical normal model in simulation. (a) Empirical coverage of the lower bound b(·,α) across different levels α. Coverage is nearly identical across the parameter space. (b) Probability that the default has smaller loss but the alternative estimate is selected across the parameter space, with dashed lines reflecting nominal coverage. (c) Probability of selecting the alternative estimate. Selection probability is higher for lower thresholds α. (d) Risk profiles of the two-stage estimators for different choices of α, as well as the MLE θ̂(·) and the shrinkage estimator θ*(·). Each data point is computed from 500 replicates with N = 50.

Fig. 1 Bound calibration and the two-stage estimator for a hierarchical normal model in simulation. (a) Empirical coverage of the lower bound b(·,α) across different levels α. Coverage is nearly identical across the parameter space. (b) Probability that the default has smaller loss but the alternative estimate is selected across the parameter space, with dashed lines reflecting nominal coverage. (c) Probability of selecting the alternative estimate. Selection probability is higher for lower thresholds α. (d) Risk profiles of the two-stage estimators for different choices of α, as well as the MLE θ̂(·) and the shrinkage estimator θ*(·). Each data point is computed from 500 replicates with N = 50.

We next examined the probability that the alternative estimate is selected on the basis of a large c-value but obtains higher loss than the default estimate. Theorem 2.3 upper bounds this probability, and in we confirm this bound holds in practice across different thresholds α. additionally compares our proposed approach to using Stein’s unbiased estimate of the risk (Stein Citation1981) of θ*(·) to select between the estimates. This approach, which we label “SURE”, returns θ̂(·) if the risk estimate exceeds N and returns θ*(·) otherwise, and is akin to the focused information criterion (Claeskens and Hjort Citation2003). However, in contrast to the two-stage estimator θ(·,α), SURE does not provide tunable control over the probability that the alternative estimator θ*(·) is mistakenly returned.

In the case that P1θ/N=1.7, choosing based on SURE gives the wrong estimate 80% of the time. Moreover, in the majority of these cases it is the alternative that is incorrectly returned (, ). By contrast, the estimator that chooses based on the c-value (with a threshold α=0.95) conservatively returns the default estimate in every replicate for this P1θ/N (). While this approach provides the estimate with greater loss in 54% of cases, it incorrectly reports the alternative in 0% of cases (). This behavior is expected as Theorem 2.3 provides an upper bound of 100*(1α)%=5%. An estimator using the unbiased risk estimate satisfies no such guarantee.

Table 2 Contingency tables of simulation outcomes with P1θ/N=1.7 when using Stein’s unbiased risk estimate (SURE), θ(·,α=0.95), or θ(·,α=0.5) to choose between the default and alternative estimates.

We next checked that our computed c-values successfully detected improvements by the alternative estimate. Recall that the alternative estimate θ*(y) shrinks all components of y toward the global mean y¯. Further, recall that by construction θ(y,α)=θ*(y) if and only if c(y)>α. Intuitively, then, we would expect the alternative estimator to improve over the MLE and for the two-stage θ(·,α) to select θ*(·) when θ is close to the subspace spanned by 1N and N12P1θ is small. , which plots the probability that θ(·,α) selects θ*(·) across different values of θ and α, confirms this intuition; when N12P1θ is small, we very often obtain large c-values and select the alternative estimator.

For completeness, we also considered the risk profile of the two-stage estimator θ(·,α) (). Specifically, for different choices of θ we computed a Monte Carlo estimate of the expected squared error loss. For the most part, the risk of θ(·,α) lies between the risks of θ̂(·) and θ*(·). However, the risk of the two-stage estimator appears to exceed the risks of the default and alternative estimators for a narrow range of values of P1θ. While it is tempting to characterize this excess risk as the price we must pay for “double-dipping” into our data, we note that the bump in risk appears to be nontrivial only for very small values of α. Recall again that we recommend choosing θ*(y) in place of θ̂(y) only when c(y) is close to 1. As such, we do not expect this type of risk increase to be much of a concern in practice.

Interpreted together, , illustrate the conservatism of the two stage approach with α=0.95. For P1θ between 1 and 1.5, θ(·,α) only rarely evaluates to θ*(·) even though this estimator has lower risk and typically has smaller loss.

Unlike conventional p-values under a null hypothesis, we should not expect the distribution of informative c-values to be uniform; indeed for parameters such that the win is consistently positive or negative, c-values can concentrate near 1 or 0, respectively.

4 Comparing Affine Estimates with Correlated Noise

We now generalize the situation described in the previous section in two ways. First, we consider correlated Gaussian noise with covariance Σ, where Σ is any N × N positive definite covariance matrix rather than restricting to Σ=IN. Second, we let our default and alternative estimates, θ̂(y) and θ*(y), be arbitrary affine transformations of the data y. Though these two estimates take similar functional forms in this section, we remain concerned with asymmetric comparisons wherein θ*(y) is less familiar than θ̂(y).

Although such generalization introduces considerable analytical challenges beyond those encountered in Section 3, we nevertheless can construct an approximate lower bound on the win that works well in practice. Specifically, for Bound 3.1, we used the tractable quantile function of the noncentral χ2 to guarantee exact coverage in Theorem 3.1. Now we encounter sums of differently scaled noncentral χ2 random variables, which do not admit analytically tractable quantiles. However, by approximating these sums with Gaussians with matched means and variances, we can proceed in essentially the same manner as in Section 3 to derive an approximate lower bound on the win. After introducing the bound, we comment on the key steps in its derivation to highlight the approximations involved, but leave details of intermediate steps to Appendix S5. We conclude with a non-asymptotic bound on the error introduced by these approximations on the coverage of the proposed bound on the win.

Approximate Bound 4.1 (Correlated Gaussian likelihood: arbitrary affine estimates). Observe y=θ+ϵ with ϵN(0,Σ) and consider θ̂(y)=Ay+k versus θ*(y)=Cy+l, where A,CRN×N are matrices and k,lRN are N-vectors. We propose (10) b(y,α)=θ̂y2θ*y2+2tr[(AC)Σ]+2z1α2U(G(y)Σ2,1α2)+12Σ12(A+ACC)Σ12F2(10) as an approximate high-probability lower bound for the win. In this expression, tr[·] denotes the trace of a matrix, G(y):=(AC)y+(kl),·Σ denotes the Σ quadratic norm of a vector (vΣ:=vΣv), ·F denotes the Frobenius norm of a matrix, and zα denotes the α-quantile of the standard normal. (11) U(G(y)Σ2,1α):=infδ>0{δ | G(y)Σ2(δ+Σ12(AC)Σ12F2)+z1α2Σ12(AC)Σ(AC)Σ12F2+4Σ12(AC)Σ12OP2δ}(11) is an approximate high-confidence upper bound on G(θ)Σ2 where ·OP denotes the L2 operator norm of a matrix.

To derive Approximate Bound 4.1 we again start by rewriting the alternative estimate as θ*(y)=θ̂(y)G(y), where now G(·) is an affine transformation of y, G(y):=(AC)y+(kl). We next write the squared error win of using θ*(y) in place of θ̂(y) as (12) W(θ,y)=2ϵG(y)+(θ̂(y)y2θ*(y)y2)(12) and observe that it suffices to obtain a high-probability lower bound for this first term. For tractability, we approximate the distribution of ϵG(y) by a normal with matched mean and variance. As we will soon see, this approximation is accurate when N is large and AC is well conditioned; in this case ϵG(y) may be written as the sum of many of uncorrelated terms of similar size. The mean and variance may be expressed as (13) E[ϵG(y)]=tr[(AC)Σ],var[ϵG(y)]=G(θ)Σ2+Σ12(A+ACC)Σ12F22.(13)

With these moments in hand, we form a probability α lower bound approximately as (14) W(θ,y)θ̂(y)y2θ*(y)y2+2tr[(AC)Σ]+2z1αG(θ)Σ2+12Σ12(A+ACC)Σ12F2.(14)

However, as before, in order to use this approximate bound we require a simultaneous upper bound on a norm of a transformation of the unknown parameter, in this case G(θ)Σ2. We compute one by considering the test statistic G(y)Σ2 and again appealing to approximate normality. In particular we characterize the dependence of the distribution of this statistic on G(θ)Σ2 through its mean and variance. We find its mean as (15) E[G(y)Σ2]=G(θ)Σ2+Σ12(AC)Σ12F2(15) and upper bound its variance by (16) var[G(y)Σ2]2Σ12(AC)Σ(AC)Σ12F2+4Σ12(AC)Σ12OP2G(θ)Σ2.(16)

Using the two quantities above and an appeal to approximate normality, we propose the approximate high-confidence upper bound, U(G(y)Σ2,1α), in EquationEquation (11). As before, by splitting our α across these two bounds we obtain the desired expression, EquationEquation (10) in Approximate Bound 4.1.

Approximation Quality

Due to the two Gaussian approximations, Approximate Bound 4.1 does not provide nominal coverage by construction. Our next result shows that little error is introduced when N is large enough and the problem is well conditioned.

Theorem 4.1

(Berry–Esseen bound). Let α(0,1) and consider b(·,α) in Approximate Bound 4.1. If both A and C are symmetric, then (17) Pθ[W(θ,y)b(y,α)]α102NC1·κ(Σ12(AC)Σ12)2(17) where κ(·) denotes the condition number of its matrix argument (i.e., the ratio of its largest to smallest singular values) and C11.88 is a universal constant (Berry, Citation1941, Theorem 1).

Remark 4.2.

Theorem 4.1 is a special case of a more general result that we provide in Appendix S5.4, which does not require A and C to be symmetric. We highlight this special case here because the bound takes a simpler form from which the dependence on the conditioning of AC is clearer, and because this condition is satisfied for many important estimates. Notably A and C are symmetric in all applications discussed in this article.

Though Theorem 4.1 provides an expected O(N12) drop in approximation error, the bound itself may be too loose to be useful in practice. In Section 6.1 we show in simulation that Approximate Bound 4.1 provides sufficient coverage even without this correction. This conservatism likely owes to slack from (A) the operator norm bound in EquationEquation (16) and (B) the union bound ensuring that the confidence interval for G(θ)Σ2 and the quantile in EquationEquation (14) hold simultaneously.

Remark 4.3

(Fast computation of b(y,α) A naive approach to computing b(y,α) in EquationEquation (10) involves finding U(G(y)Σ2,1α2) with a binary search. For more rapid computation, we can recognize U(G(y)Σ2,1α2) as the root of a quadratic. Specifically, define γ:=G(y)Σ2Σ12(AC)Σ12F2,η:=zα2, ρ:=2Σ12(AC)Σ(AC)Σ12F2, and ν:=4Σ12(AC)Σ12OP2; then from EquationEquation (11) we have that the δ that achieves the supremum satisfies γ=δ+ηρ+νδ. Rearranging, we find that U(G(y)Σ2,1α2) is the larger root of x2(2γ+η2ν)x+(γ2η2ρ)=0.

5 Extending the Reach of the c-value

Up to this point, we focused on estimating normal means with fixed affine estimators. Now we extend our c-value framework in two important directions, which we support with both theoretical and empirical results. In Section 5.1, we derive c-values for a nonlinear shrinkage estimator of normal means. We then move beyond Gaussian likelihoods in Section 5.2 and derive c-values for regularized logistic regression. In contrast to the earlier cases, these settings introduce nonlinear estimates and non-Gaussian models. To gain analytical tractability, we approximate the estimates by linear transformations of a statistic that is asymptotically Gaussian. This approximation allows us to derive bounds b(y,α) that we show have the correct coverage in an asymptotic regime. Our approach provides a template that can be followed for other nonlinear estimates and models for which the MLE is asymptotically Gaussian. We defer all proofs and details of synthetic data experiments to Appendices S6 and S7.

5.1 Empirical Bayes Shrinkage Estimates

Many Bayesian estimates are affine in the data for fixed settings of prior parameters. But when prior parameters are chosen using the data, the resulting empirical Bayesian estimates are not affine in general. We next explore computation of approximate high-confidence lower bounds on the win of empirical Bayesian estimators. In particular, we consider an approach that essentially amounts to ignoring the randomness in estimated prior parameters and computing the bound as if the prior were fixed. For simplicity, we focus on a particularly simple empirical Bayesian estimator for the normal means problem that coincides with the James–Stein estimator (Efron and Morris Citation1973). We find that, in the high-dimensional limit, bounds obtained with this naive approach achieve at least the desired nominal coverage. Finally, we show in simulation that the approximate bound has favorable finite sample coverage properties.

Empirical Bayes for estimation of normal means

Consider a sequence of real-valued parameters θ1,θ2,, and corresponding observations ynindepN(θn,1). For each NN, let ΘN:=[θ1,θ2,,θN] and YN:=[y1,y2,,yN] denote the first N parameters and observations, respectively.

We consider the MLE for ΘN (i.e., YN) as our default, which we denote by Θ̂N(YN)=YN, and we take the James–Stein estimate as our alternative; we compare on the basis of squared error loss. We write the James–Stein estimate on the first N data points as ΘN*(YN):=(1(1+τ̂N2(YN))1)YN, where τ̂N2(YN):=YN2/(N2)1. ΘN*(YN) corresponds to the Bayes estimate under the prior θniidN(0,τ̂N2) (Efron and Morris Citation1973). For this comparison, the win is WN(YN,ΘN):=Θ̂N(YN)ΘN2ΘN*(YN)ΘN2, and Appendix S6 details the associated bound bN(YN,α) obtained with Bound S4.1. In the following theorem, we lower bound the win by applying our earlier machinery for Bayes rules with fixed priors. We find that the desired coverage is obtained in the high-dimensional limit.

Theorem 5.1.

For each NN, let τN2:=N1n=1Nθn2. If the sequence τ1,τ2, is bounded, then for any α[0,1],  limNP[WN(YN,ΘN)bN(YN,α)]α.

The key step in the proof of Theorem 5.1 is establishing an Op(N12) rate of convergence of τ̂N2τN2 to zero; under this condition the empirical Bayes estimate and bound converge to the analogous estimates and bounds computed with the prior variance fixed to τN2. Accordingly, we expect similar results to hold for other models and empirical Bayes estimates when the standard deviations of the empirical Bayes estimates of the prior parameters drop as Op(N12).

Remark 5.2.

Theorem 5.1 easily extends to cover the case in which we consider a sequence of random (rather than fixed) parameters drawn iid from a Bayesian prior, which is a more classical setup for guarantees of empirical Bayesian methods; see, for example, Robbins (Citation1964). Specifically, our proof goes through in this Bayesian setting so long as the sequence τ12,τ22, is bounded in probability. This condition is satisfied, for example, when the θn are iid from any prior with a finite second moment.

To check finite sample coverage, we performed a simulation and evaluated calibration of the associated c-values (Figure S4 in Appendix S6). Despite the empirical Bayes step, the c-values appear to be similarly conservative to those computed with the exact bound in . Furthermore, this calibration profile does not appear to be sensitive to the magnitude of the unknown parameter.

5.2 Logistic Regression

In this section we illustrate how to compute an approximate high-confidence lower bound on the win in squared error loss with a logistic regression likelihood. Our key insight is that by appealing to limiting behavior, we can tackle the non-Gaussianity using the machinery developed in Section 4.

Notation and estimates

Consider a collection of M data points with random covariates XM:=[x1,x2,,xM]RM×N and responses YM:=[y1,y2,,yM]{1,1}M. For the mth data point, assume (18) ymindepp(·|xm;θ):=(1+exp{xmθ})1δ1+(1+exp{xmθ})1δ1,(18) where θRN is an unknown parameter of covariate effects and δ1 and δ1 denote Dirac masses on 1 and –1, respectively.

In this section, we choose the MLE as our default, θ̂(XM,YM):=arg maxθlogp(YM|XM;θ). And we choose our alternative to be a Bayesian maximum a posteriori (MAP) estimate under a standard normal prior (θN(0,IN)): θ*(XM,YM):=arg maxθ{logp(YM|XM;θ)12θ2}.

While a first choice for a Bayesian estimate might be the posterior mean, the MAP is an effective and widely used alternative to the MLE in practice. Furthermore, θ*(XM,YM) is also of interest as an L2 regularized logistic regression estimate.

Approximating θ* by an affine transformation

In moving away from a Gaussian likelihood we forfeit prior-to-likelihood conjugacy. In previous sections, conjugacy provided analytically convenient expressions for Bayes estimates. In order to regain analytical tractability, we appeal to a Gaussian approximation of the likelihood, defined with a second order Taylor approximation of the log-likelihood around the MLE. Under this approximation, θ̂(XM,YM)N(θ,Σ˜M), where Σ˜M:=θ2logp(YM|XM;θ)|θ=θ̂(XM,YM). As such, we regain conjugacy, and we obtain an approximate Bayes estimate as an affine transformation of the MLE, (19) θ˜*(XM,YM)=[IN+Σ˜M]1θ̂(XM,YM).(19)

As we show in Appendix S7, θ˜*(XM,YM) is a very close approximation of θ*(XM,YM), with distance decreasing at an Op(M2) rate.

An approximate bound and an asymptotic guarantee

We leverage the form in EquationEquation (19) to compute Approximate Bound 4.1 as a lower bound on the win in squared error of using the MAP estimate in place of the MLE. In particular, we take y:=θ̂(XM,YM) as the data in Approximate Bound 4.1 (this corresponds to A = IN and k = 0) and approximate the distribution of ϵ:=θ̂(XM,YM)θ as N(0,Σ˜M). Further, to compute the bound, we approximate θ*(XM,YM) by θ˜*(XM,YM) as in EquationEquation (19), corresponding to C=[IN+Σ˜M]1 and l=0.

While the precise coverage of this bound is difficult to analyze, our next result reveals favorable properties in the large sample limit.

Theorem 5.3.

Consider a sequence of random covariates x1,x2, and responses y1,y2, distributed as in EquationEquation (18). For each MN, let WM:=θ̂(XM,YM)θ2θ*(XM,YM)θ2 be the win of using the MAP estimate in place of the MLE. Finally, let bM(α) be the level-α approximate bound on WM described above. If x1,x2, are iid with finite third moment and with positive definite covariance, then for any α(0,1),limMPθ[WMbM(α)]α.

Theorem 5.3 guarantees that in the large sample limit, bM(·) has at least nominal coverage. We provide a proof of the theorem and demonstrate its favorable empirical properties in simulation in Appendix S7.

6 Applications

We now demonstrate our approach on the three applications introduced in Section 1. Our goal in this section is to demonstrate how one can compute and interpret c-values in realistic workflows. In analogy to hypothesis testing, where a p-value cutoff of 0.05 is standard for rejecting a null, we require a c-value of at least 0.95 to accept the alternative estimate; with this threshold, we expect to incorrectly reject the default estimate in at most 5% of our decisions. This choice, instead of 0.5 for example, reflects the presumed asymmetry of the comparisons; we demand strong support to adopt the alternative over the default. For all applications, we provide substantial additional details in Appendix S8.

6.1 Estimation from Educational Testing Data and Empirical Bayes

In this section we apply our methodology to a model and dataset considered by (Hoff Citation2021, sec. 3.2), in which the goal is to estimate the average student reading ability at different schools in the 2002 Educational Longitudinal Study. At each of N = 676 schools, between 5 and 50 tenth grade students were given a standardized test of reading ability. We let y=[y1,y2,,yN] denote the average scores, and for each school, indexed by n, model ynindepN(θn,σn2), where θ=[θ1,θ2,,θN] denotes the school-level means and each σn is the school-level standard error; specifically σn:=σ/Nn where σ denotes a student-level standard deviation and Nn is the number of students tested at school Nn. For convenience, we let Σ:=diag([σ12,σ22,,σN2]) so that we may write yN(θ,Σ). The goal is to estimate the school-level performances θ.

Following Hoff (Citation2021), we perform small area inference with the Fay-Herriot model (Fay and Herriot Citation1979) to estimate θ under the assumption that similar schools may have similar student performances. Specifically, we consider a vector of D = 8 attributes of each school X=[x1,x2,,xN]; these include participation levels in a free lunch program, enrollment, and other characteristics such as region and school type. We model the school-level mean as a priori distributed as θN(Xβ,τ2IN) where β is an unknown D-vector of fixed effects and τ2 is an unknown scalar that describes variation in θ not captured by the covariates. Following Hoff (Citation2021), we take an empirical Bayesian approach and estimate β,τ, and σ with lme4 (Bates et al. Citation2015). We then compare the posterior mean—which is affine in y for fixed β,τ, and σ—as an alternative to the MLE as a default; we use Approximate Bound 4.1. Specifically, we take θ*(y):=E[θ|y;β,τ,σ]=[IN+τ2Σ]1y+[IN+τ2Σ1]1Xβ and θ̂(y)=y. We compute a large c-value (c = 0.9926); its closeness to one strongly suggests that θ*(y) is more accurate than θ̂(y).

We should not always expect to obtain a large c-value for any alternative estimate, however. We next describe a case where we expect the alternative estimate to be less accurate than the default, and we check that we obtain a small c-value. In particular, we now let our alternative estimate be the posterior mean under the same model as above but with the covariates, X, randomly permuted across schools. In this situation, the responses y have no relation to the covariates, and we should not expect an improvement. Indeed, on this dataset we compute a c-value of exactly zero. However, we recall that just as a large p-value in hypothesis testing does not provide support that a null hypothesis is true, a small c-value does not provide direct support that the alternative estimate is less accurate than the default.

We provide additional details for all parts of this application in Appendix S8.1. There, we demonstrate in a simulation study that our bounds remain substantially conservative for these estimators and model even with an empirical Bayes step.

6.2 Estimating Violent Crime Density in Philadelphia

As a second application, we consider estimating the areal density of violent crimes (i.e., counts per square mile) reported in each of Philadelphia’s N = 384 census tracts. Following Balocchi et al. (Citation2022), we work with the inverse hyperbolic sine transformed density. Letting yn be the observed transformed density of reported violent crimes in census tract n, we model ynindepN(θn,σy2) where θn represents the underlying transformed density and σy2 is the noise variance. While one might interpret θn as the true density of violent crime in census tract n, we note that the implicit assumption of zero-mean error in each tract may not be realistic. Namely, systematic biases may impact the rates at which police receive and respond to calls and file incident reports in different parts of the city. Unfortunately, we are unable to probe this possibility with the available data. Nevertheless, our goal is to estimate the vector of unknown rates, θ=[θ1,θ2,,θN] from y=[y1,y2,,yN]. The observations y are a simple proxy of transformed violent crime density, but they are noisy. So it is natural to wonder if we might obtain a more accurate estimate of θ.

plots the transformed densities of both violent and nonviolent crimes reported in October 2018 in each census tract. Immediately, we see that, for any particular census tract, the observed densities of the two types of crime are similar. Further, we observe considerable spatial correlation in each plot. It is tempting to use a Bayesian hierarchical model that exploits this structure in order to produce more accurate estimates of θ. In this application, we consider iteratively refining an estimate of θ by (A) incorporating the observed nonviolent crime data and then by (B) carefully accounting for the observed spatial correlation. At each step of our refinement, we use a c-value to decide whether to continue. Before proceeding, we make a remark about our sequential approach.

Fig. 2 Transformed densities of reported (a) violent and (b) nonviolent crimes in each census tract in Philadelphia in October 2018.

Fig. 2 Transformed densities of reported (a) violent and (b) nonviolent crimes in each census tract in Philadelphia in October 2018.

Remark 6.1.

Consider using c-values and a chosen level α to choose one of three estimates (say θ̂(y),θ*(y), and θ°(y)) in two stages. Suppose we first choose θ*(y) over θ̂(y) only if the associated c-value is greater than α. Second, only if we chose θ*(y), we next choose θ°(y) over θ*(y) only if the new c-value associated with those estimates exceeds α. Then a union bound guarantees that θ°(y) will be incorrectly chosen with probability at most 2(1α).

We begin by seeing if we can improve upon the MLE, θ̂(y)=y, by leveraging the auxiliary dataset of transformed nonviolent crimes in each tract, z1,z2,,zN. To this end, we model these auxiliary data analogously to y; in each tract n, we let ηn be the unknown transformed density and independently model znindepN(ηn,σz2). We next introduce a hierarchical prior that captures the apparent similarity between θ and η within each tract. Specifically, for each tract n we decompose θn=μn+δny and ηn=μn+δnz, where μn is a shared mean for the transformed densities of violent and nonviolent reports and δny and δnz represent deviations from the shared mean specific to each crime type. Rather than encode explicit prior beliefs about μn, we express ignorance in these quantities with an improper uniform prior. Additionally, we model δny,δnziidN(0,σδ2). We fix the values of σy,σz, and σδ using historical data. We then compute the posterior mean of θ as an alternative estimate, θ*(y). Thanks to the Gaussian conjugacy of this model, θ*(y) is affine in the data y, and a closed form expression is available. See Appendix S8.2 for additional details. The resulting c-value exceeded 0.999, suggesting that we should be highly confident that θ*(y) is a more accurate estimate of θ than θ̂(y).

We next consider additionally sharing strength amongst spatially adjacent census tracts. To this end, consider a second model with spatially correlated variance components: θn=μn+δny+κny and ηn=μn+δnz+κnz. The additional terms κy=[κ1y,κ2y,,κNy] and κz=[κ1z,κ2z,,κNz] capture a priori spatial correlations; we model κy,κziidN(0,K), where K is an N×N covariance matrix determined by a squared exponential covariance function (Rasmussen and Williams Citation2006, chap. 4) that depends on the distance between the centroids of the census tracts. Once again, we exploit conjugacy in this second hierarchical model to derive the posterior mean θ°(y) in closed form. As θ°(y) is also an affine transformation of y, we can use Approximate Bound 4.1 to compute the c-value for comparing θ°(y) to θ*(y). The c-value for this comparison is only 0.843, providing much weaker support for using θ°(y) over θ*(y). Because this c-value is less than 0.95, we conclude our analysis content with θ*(y) as our final estimate.

6.3 Gaussian Process Kernel Choice: Modeling Ocean Currents

Accurate understanding of ocean current dynamics is important for forecasting the dispersion of oceanic contaminations, such as after the Deepwater Horizon oil spill (Poje et al. Citation2014). Lodise et al. (Citation2020) have recently advocated for a statistical approach to inferring ocean currents from observations of free-floating, GPS-trackable buoys. Their approach seeks to provide improved estimates by incorporating variation at the submesoscale (roughly 0.1–10 km) in addition to more commonly considered mesoscale variation (roughly 10 km and above). In this section we apply our methodology to assess if this approach provides improved estimates relative to a baseline including only mesoscale variation.

In our analysis, we consider a segment of the Carthe Grand Lagrangian Drifter (GLAD) deployment dataset (Özgökmen Citation2013). Specifically, we model a set of 50 buoys with velocities estimated at 3 hr intervals over one day (N = 400 observations total). Each observation n consists of latitudinal and longitudinal ocean current velocity measurements yn=[yn(1),yn(2)]R2 and associated spatio-temporal coordinates [latn,lonn,tn]. Following Lodise et al. (Citation2020), we model each measurement as a noisy observation of an underlying time varying vector-field distributed independently as ynindepN(F(latn,lonn,tn),σϵ2I2), where F:R3R2 denotes the time evolving vector-field of ocean currents and σϵ2 is the error variance. Our goal is to estimate F at the observation points θ:=[θ1,θ2,,θN], where for each n,θn=[θn(1),θn(2)]=F(latn,lonn,tn).

Following Lodise et al. (Citation2020), we place a Gaussian process prior on F to encode expected spatio-temporal structure while allowing for variation at multiple scales. Specifically, we model FGP(0,k(·,·)), where (20) k(θn(i),θn(i))=k1(θn(i),θn(i))+k2(θn(i),θn(i)),i{1,2}.(20)

Here k1 and k2 are squared exponential kernels with spatial and temporal length-scales that reflect mesoscale and submesoscale variations, respectively; see Appendix S8.3 for details. For simplicity, we model the latitudinal and longitudinal components of F independently. We take the posterior mean of θ under this model as the alternative estimate, θ*(y).

As a baseline, we consider an analogous estimate with covariance function k(θn(i),θn(i))=k1(θn(i),θn(i))+k2(θn(i),θn(i))1[n=n], which maintains the same marginal variance but excludes submesoscale covariances. We take the posterior mean under this model as the default estimate θ̂(y). Both θ*(y) and θ̂(y) may be written as affine transformations of y.

Using Approximate Bound 4.1, we compute a c-value of 0.99981. This large c-value allows us to confidently conclude that modeling both mesoscale and submesocale variation can yield more accurate estimates of ocean currents than mesocale modeling alone.

7 Discussion

We have provided a simple method for quantifying confidence in improvements provided by a wide class of shrinkage estimates without relying on subjective assumptions about the parameter of interest. Our approach has compelling theoretical properties, and we have demonstrated its utility on several data analyses of recent interest. However, the scope of the current work has several limitations. The present article has explored the use of the c-value only for problems of moderate dimensionality (N between 20 and 700). Loosely speaking, we suspect c-values may be underpowered to robustly identify substantial improvements provided by estimates in lower dimensional problems. Further investigation into such dimension dependence is an important direction for future work. In addition, our approach depends crucially on a high-probability lower bound that is inherently specific to the underlying model of the data, a loss function, and the pair of estimators. In the present work, we have shown how to derive and compute this bound for models with general Gaussian likelihoods, when accuracy may be measured in terms of squared error loss, and when both estimates are affine transformations of the data. We have provided a first step to extending beyond simple Gaussian models with the application to logistic regression; while we have not yet explored the efficacy of this extension on real data, we view our work as an important starting point for generalizing to broader model classes and estimation problems. We believe that further extensions to the classes of models, estimates, and losses for which c-values can be computed provide fertile ground for future work.

One direction we believe is promising is to construct the bound b(y,α) in a model and loss agnostic manner using, for example, the parametric bootstrap. Constructing an informative c-value is possible because in some cases the distribution of the win depends on the unknown parameter only through some low-dimensional projection (or at least approximately so). We suspect that this phenomenon may extend to more complex models and estimates. In such cases, when this low-dimensional characteristic sufficiently captures the distribution of the win and is estimated well enough, a parametric bootstrap may present a powerful solution. In particular, one would begin by forming an initial estimate of the parameter, and simulate a collection of bootstrap datasets by sampling data from the likelihood parameterized by the initial estimate, compute the win for each simulated dataset, and return for each b(y,α) the 1α quantile of this distribution. We expect that this method may work in many important settings; indeed, much of modern statistics and nonlinear methods are predicated on the assumption that low-dimensional structure (e.g., sparsity) exists and may be inferred. We leave further development of this more flexible approach, including an investigation of the theoretical properties, to follow-up work.

Supplementary Materials

The readme in the github provides a list of the computational resources and experimental code. And the supplementary text includes a table of contents.

Supplemental material

Supplemental Material

Download Zip (50.7 MB)

Supplemental Material

Download PDF (3.3 MB)

Acknowledgments

The authors thank Jonathan H. Huggins for the suggestions to consider Berry–Esseen bounds and the extension to logistic regression, Lorenzo Masoero and Hannah Diehl for insightful comments on the manuscript, and Matthew Stephens and Lucas Janson for useful early conversations.

Additional information

Funding

This work was supported in part by an ARPA-E project with program director David Tew, and an NSF CAREER Award. We are grateful to the Office of Naval Research for partial support under grant N00014-20-1-2023 (MURI ML-SCOPE) to the Massachusetts Institute of Technology. BLT is supported by NSF GRFP. SKD is supported by the Wisconsin Alumni Research Foundation.

References

  • Balocchi, C., Deshpande, S. K., George, E. I., and Jensen, S. T. (2022), “Crime in Philadelphia: Bayesian Clustering with Particle Optimization,” Journal of the American Statistical Association, DOI: 10.1080/01621459.2022.2156348.
  • Balocchi, C., and Jensen, S. T. (2019), “Spatial Modeling of Trends in Crime Over Time in Philadelphia,” The Annals of Applied Statistics, 13, 2235–2259.
  • Bates, D., Mächler, M., Bolker, B., and Walker, S. (2015), “Fitting Linear Mixed-Effects Models using lme4,” Journal of Statistical Software, 67, 1–48.
  • Berk, R., Brown, L., Buja, A., Zhang, K., and Zhao, L. (2013), “Valid Post-Selection Inference,” The Annals of Statistics, 41, 802–837.
  • Berry, A. C. (1941), “The Accuracy of the Gaussian Approximation to the Sum of Independent Variates,” Transactions of the American Mathematical Society, 49, 122–136.
  • Buka, S. L., Stichick, T. L., Birdthistle, I., and Earls, F. (2001), “Youth Exposure to Violence: Prevalance, Risks, and Consequences,” American Journal of Orthopsychiatry, 71, 298–310. DOI: 10.1037/0002-9432.71.3.298.
  • Casella, G., and Berger, R. L. (2002), Statistical Inference, Pacific Grove, CA: Duxbury.
  • Claeskens, G., and Hjort, N. L. (2003), “The Focused Information Criterion,” Journal of the American Statistical Association, 98, 900–916.
  • Efron, B., and Morris, C. (1973), “Stein’s Estimation Rule and its Competitors — An Empirical Bayes Approach,” Journal of the American Statistical Association, 68, 117–130.
  • Fay, R. E., and Herriot, R. A. (1979), “Estimates of Income for Small Places: An Application of James-Stein Procedures to Census Data,” Journal of the American Statistical Association, 74, 269–277.
  • Hoff, P. D. (2021), “Smaller p-values via Indirect Information,” Journal of the American Statistical Association, 117, 1254–1269.
  • Kondo, M. C., Andreyeva, E., South, E. C., MacDonal, J. M., and Branas, C. C. (2018), “Neighborhood Interventions to Reduce Violence,” Annual Review of Public Health, 39, 253–271.
  • Lee, J. D., Sun, D. L., Sun, Y., and Taylor, J. E. (2016), “Exact Post-Selection Inference, with Application to the Lasso,” Annals of Statistics, 44, 907–927.
  • Lehmann, E. L., and Casella, G. (2006), Theory of Point Estimation, New York: Springer.
  • Lindley, D. V., and Smith, A. F. (1972), “Bayes Estimates for the Linear Model,” Journal of the Royal Statistical Society, Series B, 34, 1–41.
  • Lockhart, R., Taylor, J., Tibshirani, R. J., and Tibshirani, R. (2014), “A Significance Test for the Lasso,” Annals of Statistics, 42, 413–468. DOI: 10.1214/13-AOS1175.
  • Lodise, J., Özgökmen, T., Gonçalves, R. C., Iskandarani, M., Lund, B., Horstmann, J., Poulain, P.-M., Klymak, J., Ryan, E. H., and Guigand, C. (2020), “Investigating the Formation of Submesoscale Structures Along Mesoscale Fronts and Estimating Kinematic Quantities using Lagrangian Drifters,” Fluids, 5, 1–38.
  • Özgökmen, T. M. (2013), “GLAD Experiment CODE-style Drifter Trajectories (Low-Pass Filtered, 15 Minute Interval Records),” Northern Gulf of Mexico near DeSoto Canyon, July–October 2012. Harte Research Institute, Texas A&M University-Corpus Christi. Available at https://data.gulfresearchinitiative.org/data/R1.x134.073:0004
  • Poje, A. C., Özgökmen, T. M., Lipphardt, B. L., Haus, B. K., Ryan, E. H., Haza, A. C., Jacobs, G. A., Reniers, A., Olascoaga, M. J., Novelli, G., Griffa, A., Beron-Vera, F. J., Chen, S. S., Coelho, E., Hogan, P. J., Kirwan, A. D. J., Huntley, H. S., and Mariano, A. J. (2014), “Submesoscale Dispersion in the Vicinity of the Deepwater Horizon Spill,” Proceedings of the National Academy of Sciences, 111, 12693–12698.
  • Rasmussen, C. E., and Williams, C. K. (2006), Gaussian Processes for Machine Learning, Cambridge, MA: MIT Press.
  • Robbins, H. (1964), “The Empirical Bayes Approach to Statistical Decision Problems,” The Annals of Mathematical Statistics, 35, 1–20.
  • Stein, C. M., (1981), “Estimation of the Mean of a Multivariate Normal Distribution,” The Annals of Statistics, 9, 1135–1151.
  • Taylor, J., and Tibshirani, R. (2018), “Post-Selection Inference for Penalized Likelihood Models,” Canadian Journal of Statistics, 46, 41–61.
  • Tian, X. (2020), “Prediction Error after Model Search,” Annals of Statistics, 48, 763–784.
  • Tibshirani, R., and Rosset, S. (2019), “Excess Optimism: How Biased is the Apparent Error of an Estimator Tuned by SURE?” Journal of the American Statistical Association, 114, 697–712.
  • Wallace, T. D. (1977), “Pretest Estimation in Regression: A Survey,” American Journal of Agricultural Economics, 59, 431–443.

Appendix

Proof of Theorem 2.2

Proof. The result follows directly from the definition of c(y) and the conditions on b(·,·). More explicitly, Pθ[W(θ,y)0 and c(y)>α]Pθ[W(θ,y)0 and b(y,α)>0]Pθ[W(θ,y)<b(y,α)]1α,where the first line follows from the definition of the c-value and the final line follows from EquationEquation (1). □

Proof of Theorem 2.3

Proof. The condition L(θ,θ(y,α))>L(θ,θ̂(y)) can occur only when both (A) 0>W(θ,y) and (B) θ(·,α) evaluates to θ*(·) rather than θ̂(·). Event (B) implies c(y)>α and therefore b(y,α)>0. By transitivity, b(y,α)>0 and 0>W(θ,y)b(y,α)>W(θ,y). By assumption, the event b(y,α)>W(θ,y) occurs with probability at most 1α. □