780
Views
0
CrossRef citations to date
0
Altmetric
Smoothing Methods

Inference and Computation for Sparsely Sampled Random Surfaces

, &
Pages 1361-1374 | Received 17 Mar 2021, Accepted 22 Jan 2022, Published online: 21 Mar 2022

Abstract

Nonparametric inference for functional data over two-dimensional domains entails additional computational and statistical challenges, compared to the one-dimensional case. Separability of the covariance is commonly assumed to address these issues in the densely observed regime. Instead, we consider the sparse regime, where the latent surfaces are observed only at few irregular locations with additive measurement error, and propose an estimator of covariance based on local linear smoothers. Consequently, the assumption of separability reduces the intrinsically four-dimensional smoothing problem into several two-dimensional smoothers and allows the proposed estimator to retain the classical minimax-optimal convergence rate for two-dimensional smoothers. Even when separability fails to hold, imposing it can be still advantageous as a form of regularization. A simulation study reveals a favorable bias-variance tradeoff and massive speed-ups achieved by our approach. Finally, the proposed methodology is used for qualitative analysis of implied volatility surfaces corresponding to call options, and for prediction of the latent surfaces based on information from the entire dataset, allowing for uncertainty quantification. Our cross-validated out-of-sample quantitative results show that the proposed methodology outperforms the common approach of pre-smoothing every implied volatility surface separately. Supplementary materials for this article are available online.

1 Introduction

The term random surfaces refers to continuous data on a two-dimensional domain. Such datasets consist of multiple independent replications of some underlying process X=(X(t,s),tT,sS), forming a random sample X1,,XN, and can be tackled in the context of functional data analysis (FDA), see Ramsay and Silverman (Citation2005, Citation2007). In practice, surface-valued data cannot be observed as infinite-dimensional objects, one instead observes only a finite number of measurements per surface. The classical split in literature (see Zhang and Wang Citation2016, for a comprehensive overview) consists of two measurement regimes: the dense sampling and the sparse sampling. The former considers the setting where the surface data are recorder densely enough (usually on a grid) so they can be worked with as if they were truly infinite dimensional, possibly after a pre-smoothing step, while retaining the same asymptotic properties as truly infinite dimensional data (Hall, Müller, and Wang Citation2006). The sparse regime, however, requires a different approach. Here, the surfaces are observed only on a small (varying) number of irregular locations, which also vary across the surfaces and are corrupted with measurement errors. If such observations are gridded, which is common for computational reasons, the resulting matrix-valued samples contain many missing entries. Densely observed random surfaces arise frequently in medical imaging, linguistics (Pigoli et al. Citation2018), or climate studies (Gneiting, Genton, and Guttorp Citation2006). Examples of sparsely observed random surfaces include longitudinal studies, where only a part of a functional profile is measured at each visit (Lopez et al. Citation2020), geolocalized data (Yarger, Stoev, and Hsing Citation2020; Zhang and Li Citation2020; Wang, Wong, and Zhang Citation2020), or financial data (Fengler Citation2009; Kearney, Cummins, and Murphy Citation2018).

Regardless of the sampling regime, any functional data analysis will begin by estimating two fundamental objects: the mean function μ(t,s)=EX(t,s) and the covariance kernel c(t,s,t,s)=cov(X(t,s),X(t,s)). These ought to be estimated nonparametrically, since availability of replicated observations should allow so. While the mean has the same dimension as the underlying process itself, the covariance is an object of a higher dimension, and its estimation thus suffers from the curse of dimensionality. The size of a general covariance often poses both computational and statistical issues, for example, restricting the resolution of the grid one can handle (Aston, Pigoli, and Tavakoli Citation2017; Masak, Sarkar, and Panaretos Citation2020).

Separability of the covariance is a popular nonparametric assumption to reduce the statistical and computational burden, which arises when working on a multi-dimensional domain. Well-known and often coupled together with parametric models in the field of spatial statistics (see e.g., De Iaco, Myers, and Posa Citation2002, and references therein), separability has lately attracted a lot of attention also from the nonparametric viewpoint of FDA (Bagchi and Dette Citation2020; Constantinou, Kokoszka, and Reimherr Citation2017; Masak and Panaretos Citation2019; Dette, Dierickx, and Kutta Citation2020). Intuitively, it allows one to decouple the dimensions, regarding the random surfaces as an enlarged ensemble of random curves instead. Specifically, separability assumes that the full covariance can be written as a product of covariance kernels corresponding to individual dimensions, that is:c(t,s,t,s)=a(t,t)b(s,s)for some bivariate kernels a=a(t,t) and b=b(s,s). Thus separability decomposes the four-dimensional covariance kernel into two kernels, each of the same dimensionality as the mean function.

Though separability cannot always be justified in the context of applied problems (Rougier Citation2017), it is nevertheless frequently imposed because of the computational advantages it entails (Gneiting, Genton, and Guttorp Citation2006; Genton Citation2007; Pigoli et al. Citation2018). And, while wrongfully assuming separability introduces a bias, it may still lead to improved estimation and prediction due to an implicit bias-variance tradeoff. What is certain is that separability greatly reduces computational costs of both the estimation and prediction tasks.

To the best of our knowledge, the aforementioned properties of separability have been mostly explored in the case of densely observed data. In the sparse regime, there is no procedure for the nonparametric estimation of a separable covariance. Methodology designed for the densely observed surfaces does not apply, and methods designed for sparsely observed curves cannot feasibly be extended to surfaces, as we will later argue. Since sparse data are generally associated with higher computational complexity (extra costs associated with smoothing) as well as higher statistical complexity (variance is inflated due to sparse measurements burdened by noise), it appears that the gains stemming from separability could be even larger in the sparse regime.

The goal of this article is to leverage the separability assumption to reduce complexity of covariance estimation down to that of mean estimation when working under the sparse regime. The method of choice for sparsely observed functional data on one-dimensional domains is the PACE approach (Yao, Müller, and Wang Citation2005), which is based on kernel regression smoothers. A naive generalization of PACE to a two-dimensional domain would entail a computationally infeasible local linear smoothing step in a four-dimensional space. Instead, we demonstrate how to make careful use of separability to collapse the four-dimensional smoother into several two-dimensional surface smoothers. Consequently, the estimation of the mean, the covariance, and the noise level all become of similar computational complexity. While this complexity depends on a specific smoothing algorithm, our methodology can benefit from any speed-ups (such as those proposed by Chen and Jiang Citation2017) just as well as the naive generalization of PACE. Reducing the dimensionality of the problem always cuts down computations, regardless of the specific smoothing algorithm used (see Appendix I.2, supplementary materials).

Second, our asymptotic theory shows that the estimators match the minimax optimal convergence rates for two-dimensional smoothing problems. Moreover, our simulation study demonstrates the enormous computational gains offered by separability, the statistical gains when the ground-truth is separable, and a favorable bias-variance tradeoff when the truth is in fact not separable.

Finally, we illustrate the benefits of our approach on a qualitative analysis as well as prediction of implied volatility surfaces corresponding to call options. Here, one surface corresponds to a fixed asset (e.g., a stock), for which the right to buy it for an agreed-upon strike price (first dimension) at a future time to expiration (second dimension) is traded. The value of implied volatility at given strike and time to expiration is derived directly from observed market data (the option prices), by the well-known and commonly used Black-Scholes formula (Black and Scholes Citation1973; Merton Citation1973). The implied volatilities are preferred over the option prices, since they are dimensionless quantities, allow for a direct comparison of different assets, and are well familiar to practitioners. Since the options are traded only for a finite number of strikes and times to expiration, which vary across different surfaces, the observed data consist of a sparse ensemble. The interpolation of such an ensemble is a typical objective in financial mathematics as the latent implied volatility surfaces are of interest for other tasks such as prediction (forecasting) of option prices (Hull Citation2006). The common practice is to interpolate or smooth the measurements for every surface independently. For example, Cont and Da Fonseca (Citation2002) used pre-smoothing by local polynomial regression, evaluating an individual smoother for every single surface. This approach may, however, pose issues when the available sparse measurements for a given surface are concentrated only on a subset of the domain, which is often the case. Once the predicted surfaces are fed into a subsequent predictive models, the naively extrapolated parts of the surface are given the same weight as the more reliable interpolated parts. Instead, we advocate for the idea of “borrowing strength” for the purpose of predicting the latent surfaces via best linear unbiased prediction using the information from the entire dataset, which also allows for uncertainty quantification (Yao, Müller, and Wang Citation2005). Under separability, we only need to use two-dimensional surface smoothing, which is the case of the pre-smoothing approach as well. At the same time, the proposed methodology outperforms the pre-smoothing approach in terms of prediction error.

2 Methodology

2.1 Model and Observation Scheme

We assume the existence of iid latent surfaces XnL2([0,1]2),n=1,,N, which are mean-square continuous with continuous sample paths. We denote the mean function as μ=μ(t,s), whereμ(t,s)=EX1(t,s),t,s[0,1],and the covariance kernel as c=c(t,s,t,s), wherec(t,s,t,s)=E[(X1(t,s)μ(t,s))(X1(t,s)μ(t,s))],t,s,t,s[0,1].

We think of the first dimension as being temporal, denoted by variable t, and the second dimension as being spatial, denoted by variable s, though this convention is only made for the purposes of presentation.

The crucial assumption in this paper is that of separability of the covariance:

  1. The covariance kernel of the random surfaces X1,,XN satisfies(1) c(t,s,t,s)=a(t,t)b(s,s),t,s,t,s[0,1],(1)

for a purely temporal covariance a=a(t,t) and some purely spatial covariance b=b(s,s).

The process XL2([0,1]2) is separable if it is, for example, an outer product of two independent univariate processes (a purely temporal one and a purely spatial one). In that case, apart from the covariance, the mean function also separates into a product of a purely temporal and a purely spatial functions. However, a process can have a separable covariance even when it is not itself separable (Rougier Citation2017), for example, the mean function may not be separable. We do not assume separability of the latent process itself, we only assume separability of its covariance in the sense of (1).

We work under the sparse sampling regime, where every surface is observed only at a finite number of irregularly distributed locations, and those measurements are corrupted by independent additive errors. For the n-th latent surface Xn , the number of measurements Mn as well as the locations of the measurements {(tnm,snm)|m=1,,Mn}[0,1]2 are considered random, and the observations are given by the errors-in-measurements model (Yao, Müller, and Wang Citation2005; Li and Hsing Citation2010; Zhang and Wang Citation2016):(2) Ynm=Xn(tnm,snm)+εnm,m=1,,Mn,n=1,,N,(2) where εnm are iid with E εnm=0 and var(εnm)=σ2>0 being the noise level.

2.2 Motivation

In this section, we assume for simplicity that the mean is zero, and provide a heuristic description of how one might estimate the separable covariance (1). Note that we havecov(Ynm,Ynm)=a(tnm,tnm)b(snm,snm)+σ21[m=m],n=1,,N,m,m=1,,Mn.

Consider the “raw covariances” Gnmm:=YnmYnm. Ignoring the assumption of separability, one could attempt to “lift” the PACE approach (Yao, Müller, and Wang Citation2005) up to higher dimensions. This amounts to plotting the raw covariances as a scatterplot in a four-dimensional domain, discarding the diagonal covariances burdened by noise, and using a surface smoother to obtain an estimator of the covariance. We refer to this procedure as 4D smoothing, see Appendix I.2, supplementary materials for details. However, there are two issues with 4D smoothing. The curse of dimensionality results in an estimator of poor quality, unless surfaces are observed relatively densely and many replications are available. And, especially when the latter is true, the computational costs of smoothing in a higher dimension can be excessive. We make the assumption of separability mainly to cope with these two issues, which is often the case in the literature already when working with fully observed data (Gneiting, Genton, and Guttorp Citation2006; Pigoli et al. Citation2018). We do not see separability as a critical modeling assumption, but rather as a regularization, which possibly introduces a bias. Separability reduces both the statistical and the computational complexity of the covariance estimation task. This is always important when working with random surfaces, whatever their mode of observation, but becomes particularly crucial when working with sparsely observed surfaces.

In the following, we provide a heuristic on how separability can be used to our advantage in the sparse observation regime. Assuming zero mean for now, we haveEGnmm=EYnmYnm=a(tnm,tnm)b(snm,snm)+σ2 1[m=m].

Imagine for a moment that the spatial kernel b=b(s,s) is known, and consider the set of values(3) {YnmYnmb(snm,snm)|m,m=1,,Mn,mm,n=1,,N}.(3)

The expectation of every point in this set is a(tnm,tnm), so we can chart these points in a scatterplot as(tnm,tnm,YnmYnmb(snm,snm)),and use a two-dimensional surface smoother to obtain an estimator of a=a(t,t).

Correspondingly, if one knew the temporal kernel a=a(t,t) instead, the set of values (3) (with a in the denominator instead of b) could be arranged against snm and snm to obtain an estimator of b=b(s,s). When neither the temporal kernel a nor the spatial b is known, one can start with a fixed b and iterate between updates of a and b, smoothing a scatterplot once per every single update.

However, there are two issues with such an approach. First, for small denominators, the corresponding points on the scatterplot are not reliable, and using them as they are can have a severe negative impact on estimation quality. Second, unless very few observations per surface are available, the procedure above is computationally very demanding, see Appendix I.5, supplementary materials.

To cope with these issues, we use weights for the surface smoother and use gridding, that is, split the domain into disjoint intervals and work on a grid. While gridding can significantly reduce computations already on a univariate domain (Yao, Müller, and Wang Citation2005), the gains are much bigger in higher dimensions. In the following section, we introduce our methodology in full from the theoretical perspective, while computational aspects are deferred to Appendix I.2, supplementary materials.

2.3 Estimation of the Model Components

We use local linear regression surface smoothers (Fan and Gijbels Citation1996) to formalize the heuristic described in the previous section and estimate the components of the model from Section 2.1, that is, the mean μ=μ(t,s), the temporal kernel a=a(t,t), the spatial kernel b=b(s,s), and the noise level σ2.

By applying a surface smoother to the set {(xk,yk,zk)|k=1,,M}R3 with given weights {wk|k=1,,M}, we understand calculating γ̂0=γ̂0(x,y) as the minimizer of the weighted sum of squares(γ0̂,γ1̂,γ2̂)=arg minγ0,γ1,γ2k=1MK(xxkh1)K(yykh2)wk[zkγ0γ1(xxk)γ2(yyk)]2for every fixed (x,y)[0,1]2, where K(·) is a smoothing kernel function and h1,h2>0 are bandwidths. Throughout the paper, we use the Epanechnikov kernel, use cross-validation to select the bandwidths, and mention weights only when they are not all equal.

First, we estimate the mean by applying the surface smoother to the set(4) {(tnm,snm,Ynm)|m=1,,Mn,n=1,,N}.(4)

Denote the resulting estimator by μ̂=μ̂(t,s).

Next, consider the “raw” covariances Gnmm=[Ynmμ̂(tnm,snm)][Ynmμ̂(tnm,snm)]. We begin by applying the surface smoother to the set{(tnm,tnm,Gnmm)|m,m=1,,Mn, mm,n=1,N}to obtain a preliminary estimator of a=a(t,t), denoted by â0=â0(t,t). Then we use this preliminary estimator to calculate a proxy of b=b(s,s), like we described in the previous section. Namely, we apply the surface smoother to the set(5) {(snm,snm,Gnmmâ0(tnm,tnm))|m,m=1,,Mn,mm,n=1,N}(5) using weights {â02(tnm,tnm)} to obtain b̂0=b̂0(s,s). If the denominator in the set above is ever zero, we remove the corresponding point from the set. Note that since the weights are equal exactly to the denominators squared, it makes sense even formally that such a point is never considered for the surface smoother. As the next step, we refine our estimator of a by applying the surface smoother to the set(6) {(tnm,tnm,Gnmmb̂0(snm,snm))|m,m=1,,Mn,mm,n=1,N}(6) using weights {b̂02(snm,snm)} from which we obtain â=â(t,t). Finally, we refine the estimator of b by applying the surface smoother to set (5) with â0 replaced by â, resulting in the estimator b̂=b̂(s,s).

Finally, once both the mean and the separable covariance have been estimated, it remains to estimate the noise level σ2, which is of interest for example, for the purposes of prediction. We apply the surface smoother to the set {(tnm,snm,Gnmm)|m=1,,Mn,n=1,N} to obtain V̂=V̂(t,s). Note that since EGnmma(tnm,tnm)b(snm,snm)+σ2, we can estimate σ2 byσ̂2=41/43/41/43/4[V̂(t,s)â(t,t)b̂(s,s)]dtds,where (similarly to Yao, Müller, and Wang Citation2005) we integrate only along the middle part of the domain to mitigate boundary issues.

The workflow of the estimation scheme described above is visualized in . The main novelty of our approach lies in the part where the separable covariance is estimated. Separability allows us to reduce dimensionality of the problem. Hence only two-dimensional smoothing is required, while a straightforward multi-dimensional generalization of for example, the PACE approach (Yao, Müller, and Wang Citation2005) not using separability would require four-dimensional smoothing to estimate the covariance.

Fig. 1 Workflow of the proposed estimation procedure. We estimate firstly the mean from the data {Ynm}, then the separable covariance (in several steps) from the raw covariances {Gnmm}, and finally the noise level. A surface smoother over a 2D domain is used in every step (once per a single thin arrow).

Fig. 1 Workflow of the proposed estimation procedure. We estimate firstly the mean from the data {Ynm}, then the separable covariance (in several steps) from the raw covariances {Gnmm′}, and finally the noise level. A surface smoother over a 2D domain is used in every step (once per a single thin arrow).

The estimation of the separable terms can be viewed as an iterative procedure, where either a or b is kept fixed while the other term is being updated. The initializing estimate â0 is also obtained in this way, starting from b̂01. A natural question is whether one should iterate this process until convergence, or simply stop after a single step, and use for example, â0 as the estimator of a. The approach we advocate for uses exactly two steps (for both a and b). The reason is the following. As will be shown in Section 3, the asymptotic distribution of b̂ does not depend on â0, and similarly for â. This fact follows from separability. However, given the motivation in the previous section, one can anticipate the finite sample performance of â to be better than that of â0, which is verified in our simulation study. Also, one can think of the first step as estimating the optimal weights consistently, and the second step as using those consistently estimated weights to produce the estimators. Naturally, one can use more than two steps, and the optimal number of steps can be chosen in a data-driven way using cross-validation, cf Appendix I.3, supplementary materials. Still, we recommend using exactly two steps due to the conceptual simplicity.

One of the distinctive features of our methodology is the use of this explicit weighing scheme, where the weights for each of the two covariance kernels depend on the other covariance kernel. We explain why this is done in the following example. A more precise justification for the specific (quadratic) form of the weights, which requires some additional background, is deferred to Appendix I.1, supplementary materials.

Example 1.

Assume X1 is observed at four locations only, as depicted in . First, let us describe the 4D smoothing estimator at a fixed location (t,s,t,s): X1 contributes to ĉ(t,s,t,s) only if there is a pair of two locations, where X1 is observed, such that one of the locations is close to (t, s), and the other is close to (t,s). In this case, closeness in time, resp. space, is measured by the bandwidths ht , resp. hs . In (left), only a single pair of locations contributes to estimation at (t,s,t,s).

Fig. 2 A single random surface is observed at four locations (depicted by “x”), and these observations contribute differently to estimation of the covariance at a fixed location (t,s,t,s) in the case of 4D smoothing (left) and one step of the proposed approach leading to the estimator of a(t,t) (right). The gray areas depict active neighborhoods and the dashed lines depict the contributing raw covariances (products of the values at the connected locations).

Fig. 2 A single random surface is observed at four locations (depicted by “x”), and these observations contribute differently to estimation of the covariance at a fixed location (t,s,t′,s′) in the case of 4D smoothing (left) and one step of the proposed approach leading to the estimator of a(t,t′) (right). The gray areas depict active neighborhoods and the dashed lines depict the contributing raw covariances (products of the values at the connected locations).

Second, let us contrast this to a single step in the proposed estimating procedure. Assume that b=b(s,s) is fixed in the current step, and we are estimating a=a(t,t). In this step, the spatial dimensions are not explicitly considered, smoothing is conducted only in the temporal dimensions. Hence the product of any two locations, where X1 is observed, contributes to â, as long as the locations are close to (t, s) and (t,s) in the temporal domain. In the situation displayed in (right), this leads to four contributing raw covariances. In other words, when estimating the temporal part of the covariance at (t,s,t,s), we can consider even raw covariances, which are spatially far from (t,s,t,s). This is meaningful due to separability. The adopted weighting scheme then ensures that raw covariances arising from points which are spatially distant are appropriately weighted.

To sum up, 4D smoothing can be understood as local averaging over information about c(t,s,t,s) captured in raw covariances. Under separability, however, the proposed methodology borrows information in a different manner, always allowing for more freedom in one dimension or the other, depending on which dimension is currently held fixed.

2.4 Marginalization on a Grid

By marginalization, we mean preparation of the raw covariances for 2D smoothing step, that is, charting the raw covariances either in time or in space and weighting them as in formulas (6) and (5). In this section, we will show that scatter points can be pooled together during the marginalization process to save computations during the subsequent smoothing step, when the data are observed (or rounded to) a common grid. The associated computational advantages are discussed in Appendix I.2.

Assume for the remainder of this section that data from the measurement scheme (2) arrive as matrices Y1,,YNRd1×d2 with only some of their entries known, that is, most of the entries are missing. The marginal covariance kernels a=a(t,t) and b=b(s,s) are replaced by matrices ARd1×d1 and BRd2×d2, respectively. We assume again for simplicity that the mean μ=μ(t,s) is zero. The raw covariances then form a tensor Gn=YnYnRd1×d2×d1×d2 with entries Gn[i,j,i,j]=Yn[i,j]Yn[i,j].

Again, like before, assume that B is fixed for now, and we are using the surface smoother on the discrete equivalent to set (6), that is(7) {(i,i,Gn[i,j,i,j]B[j,j])|Yn observed at (i,j)(i,j),n=1,,N},(7) to obtain Â. Assume that Yn was observed at locations at (i,j1),(i,j2) and (i,j), where no two locations are the same. As explained in Appendix I.1, supplementary materials, it is equivalent for the surface smoother to replace the corresponding two values from (7), that isGn[i,j1,i,j]B[j1,j]andGn[i,j2,i,j]B[j2,j] with weights B[j1,j] and B[j2,j], by a single value(8) (B[j1,j]Gn[i,j1,i,j]+B[j2,j]Gn[i,j2,i,j])/(B2[j1,j]+B2[j2,j])(8) with the aggregated weight B2[j1,j]+B2[j2,j].

Let yn,i (resp. yn,i) denote the ith (resp. ith) column of Yn. Let qi denote the identifier of whether the entries of yn,i were observed (and similarly qi), that is

Then, value (8) can be calculated as yn,iByn,i/qiB2qi with the aggregated weight given by qiB2qi, where B2 is the entry-wise square of B. Naturally, this can be generalized to the case when arbitrary number of entries in the ith and ith columns of Yn are observed. But more importantly, it can be also generalized to account for different pairs of columns of Yn at the same time.

Let Qn be formed by vectors qi, that is, Qn[i,j]= 1[Y[i,j] is observed]. Then the contribution of the n-th surface Yn into set (7) can be calculated at once as(9) YnB˜Yn/QB˜2Q(9) where B˜ is B with the diagonal values replaces by zeros and B˜2 is the entry-wise square of B˜. The diagonal values of B are replaced by zeros as described in order to discard products of the type Y2[i,j], which are burdened with noise.

The situation is analogous in the other step, when A is fixed and B is calculated. The whole procedure of estimating the separable covariance based on gridded sparse measurements is outlined in Algorithm I.1 in Appendix I.2, supplementary materials, where we also provide some implementation details.

2.5 Prediction

Another objective of our methodology is the recovery of the latent surfaces based on the sparse and noisy observations thereon. The prediction method we are going to present in this section follows the principle of borrowing strength across the entire dataset, an expression framed by Yao, Müller, and Wang (Citation2005). Specifically, consider the training dataset {Ynm:m=1,,Mn, n=1,,N} composed of the observations made on random surfaces X1,,XN via (2), and a new random surface X observed under the same sampling protocol:(10) Ym=X(tm,sm)+εm,m=1,,M.(10)

We assume that X comes from the same population as X1,,XN. In fact, it may be set (together with its sparse measurements) to one of the training surfaces X1,,XN if the task is to predict one of those.

Our prediction method is calibrated on all the observations, and this information is used for the prediction of X. This contrasts to the pre-smoothing step often used in the functional data literature (Ramsay and Silverman Citation2005, Citation2007), typically in the dense regime, where the prediction (usually some kind of a smoother) of X is based only on {Ym:m=1,,M}.

Let X be a random surface with the mean μ(t,s), t,s[0,1], and the covariance kernel a(t,t)b(s,s),t,t,s,s[0,1], observed through sparse measurements (10). Then the best linear unbiased predictor of the latent surface X given the sparsely observed data Y=(Y1,,YM), denoted as Π(X|Y), is given by (see Henderson Citation1975):(11) Π(X(t,s)|Y)=μ(t,s)+cov(X(t,s),Y)[var(Y)]1(YEY),t,s[0,1],(11) where(12) cov(X(t,s),Y)=(a(t,tm)b(s,sm))m=1MRM,t,s[0,1],(12) (13) var(Y)=(a(tm,tm)b(sm,sm)+σ21[m=m])m,m=1MRM×M,EY=(μ(tm,sm))m=1MRM.(13)

Formula (11) contains the unknown mean surface μ(·,·), covariance kernels a(·,·) and b(·,·) as well as the measurement error variance σ2. In reality, we estimate these quantities from the training dataset, say sparsely observed measurements on X1,,XN, and plug-in the estimates μ̂,â,b̂, and σ̂2 into (11). We shall denote such predictor as Π̂(X|Y). We show in the following section (Theorem 2) that the predictor Π̂(X|Y) converges to its theoretical counterpart Π(X|Y) as the number of training samples N grows to infinity.

In the rest of this section, we turn our attention to the construction of prediction bands under a Gaussian assumption.

  1. The random surface X is a Gaussian random element in L2([0,1]2) and the measurement error ensemble {εm}m=1M is a Gaussian random vector.

First note that under the Assumption (A2), the best linear unbiased predictor (11) actually corresponds to the conditional expectation E[X(t,s)|Y]. Furthermore, the conditional covariance structure is given for t,t,s,s[0,1] by(14) cov(X(t,s),X(t,s)|Y)=a(t,t)b(s,s)cov(X(t,s),Y)[var(Y)]1[cov(X(t,s),Y)].(14)

Moreover, we denote cov̂(X(t,s),X(t,s)|Y) the empirical counterpart to (14), where the unknown quantities a, b and σ2 are replaced by their estimators.

For (t,s)[0,1]2 and α(0,1), the (1α)-prediction interval for X(t,s) is given by(15) Π̂(X(t,s)|Y)±u1α/2cov̂(X(t,s),X(t,s)|Y)(15) where u1α/2 is the (1α/2)-quantile of the standard Gaussian law. The point-wise prediction band is then constructed by connecting the intervals (15) for all (t,s)[0,1]2.

The construction of the simultaneous prediction band is more involved, and we shall use the technique proposed by Degras (Citation2011). Define the conditional correlation kernel(16) corr̂(X*(t,s),X*(t,s)|Y*)=cov̂(X*(t,s),X*(t,s)|Y*)var̂(X*(t,s)|Y*)var̂(X*(t,s)|Y*)(16) if the division on the right-hand is well defined and zero otherwise, and where we define var̂(X*(t,s)|Y*)=cov̂(X*(t,s),X*(t,s)|Y*).

Then, construct the simultaneous prediction band by connecting the intervals(17) Π̂(X(t,s)|Y)±û1αcov̂(X(t,s),X(t,s)|Y),(17) where û1α is the (1α)-quantile of the law of(18) Ŵ=supt,s[0,1]2|Ẑ(t,s)|(18) with Ẑ being a Gaussian element with cov(Ẑ(t,s),Ẑ(t,s))=corr̂(X(t,s),X(t,s)|Y) for t,t,s,s[0,1]. Therefore, by the definition, P(sup(t,s)[0,1]2|Ẑ(t,s)|û1α)=1α. Numerical calculation of this quantile is explained by Degras (Citation2011), who also concludes that û1α<u1α/2. Therefore the point-wise prediction band is always enveloped by the simultaneous prediction band, as expected.

The asymptotic coverage of the point-wise prediction band (15) and the simultaneous prediction band (17) is verified in Theorem 2 in the following section.

3 Asymptotic Properties

In this section, we establish consistency and convergence rates of the estimators μ̂=μ̂(t,s),â=â(t,t), and b̂=b̂(s,s), as well as the measurement error variance estimator σ̂2.

To obtain the asymptotic properties of the proposed estimators, we need a standard set of assumptions refining our observation scheme introduced in Section 2.1, specifying smoothness of the mean surface and the separable covariance, and providing finiteness of moments and rates for the bandwidths. These are specified in the Appendix (Section II.1), supplementary materials, but one problem-specific assumption is provided here.

  1. The true value of the covariance kernel b(·,·), of the separable model (1) satisfies(19) Θ=def0101b(s,s)fs(s)fs(s)dsds0(19)

where fs(s)=01f(t,s)(t,s)dt is the marginal density of the random location s11.

While the assumptions, which are deferred to the appendix, are standard in the smoothing literature, Assumption (B10) might be surprising, especially considering it is not symmetric between a(·,·) and b(·,·). The reason behind this asymmetry is that our estimation methodology starts by smoothing the raw covariances Gnmm against (tnm,tnm) in order to produce the preliminary estimator â0(·,·). The condition (19) ensures that the estimator â0(·,·) converges to a nonzero quantity, see the constant Θ in Theorem 1. By contrast, this issue is not present in the follow-up steps. Due to positive semidefinitness of b(·,·), the constant Θ can only be zero if all eigenfunctions of b(·,·) are orthogonal to the marginal sampling density fs . This cannot happen for example, unless b(·,·) is exactly low-rank, and with all the eigenfunctions changing signs. From the practical perspective, the condition is merely a technicality.

The mean surface asymptotic theory is presented as the following proposition. Note that the separability Assumption (A1) is not required here.

Proposition 1.

Under the Assumptions (B1)–(B6):sup(t,s)[0,1]2|μ̂(t,s)μ(t,s)|=OP(logNNh2+h2)asN.

Our main asymptotic result, the consistency and the convergence rates for the separable model components (1), is presented in the following theorem.

Theorem 1.

Under the Assumptions (A1), (B1)–(B10):(20) sup(t,t)[0,1]2|â(t,t)Θa(t,t)|=OP(logNNh2+h2),(20) (21) sup(s,s)[0,1]2|b̂(s,s)1Θb(s,s)|=OP(logNNh2+h2),(21) as N, where Θ is defined in (19).

The separable decomposition (1) is not identifiable, because a constant can multiply one component while dividing the other, that is, a(t,t)b(s,s)=[λa(t,t)][(1/λ)b(s,s)] for any λ(0,). Therefore we can only aim to recover the covariance kernels a(·,·) and b(·,·) up to a multiplicative constant and its reciprocal, respectively. The number Θ in statements (20) and (21) plays the role of such a constant and depends on the initialization of the algorithm, in our case on the fact that the first estimator â0 smooths the raw covariances Gnmm without any weighting. Still, the product â(t,t)b̂(s,s), estimates consistently the covariance structure c(t,s,t,s)=a(t,t)b(s,s), which is summarized in the following corollary.

Corollary 1.

Under the Assumptions (A1), (B1)–(B10):sup(t,s,t,s)[0,1]4|â(t,t)b̂(s,s)a(t,t)b(s,s)|=OP(logNNh2+h2)asN.

The asymptotic behavior of the noise level σ2 is given next.

Proposition 2.

Under the Assumptions (A1), (B1)–(B11):σ̂2=σ2+OP(logNNh2+h2)asN.

This completes the asymptotic theory for our estimators, and we now turn to prediction. The following theorem shows that the predictor Π̂(X|Y) defined in Section 2.5 converges—as the sample size grows to infinity—to its oracle counterpart (11), which assumes the knowledge of the true second-order properties of the data. Moreover, the theorem also proves the asymptotic coverage of the point-wise and simultaneous prediction bands (15) and (17).

Theorem 2.

Under the Assumptions (A1), (B1)–(B11), we have, conditionally on Y:(22) sup(t,s)[0,1]2|Π̂(X(t,s)|Y)Π(X(t,s)|Y)|=oP(1),asN.(22)

Assuming further (A2) and fixing α(0,1), it holds as N that(t,s)[0,1]2P(|Π̂(X(t,s)|Y)X(t,s)|u1αvar̂(X(t,s)|Y) | Y)1α,P(sup(t,s)[0,1]2[var̂(X(t,s)|Y)]1/2|Π̂(X(t,s)|Y)X(t,s)|û1α | Y )1α.

The rates established in this section manifest the statistical consequences of separability. Corollary 1 shows that the complete covariance structure is estimated with the rate OP((logN)/(Nh2)+h2), which is the known optimal minimax convergence rate (Fan and Gijbels Citation1996) for two-dimensional nonparametric regression. By steps similar to our proofs presented in Appendix II, supplementary materials, it could be shown that the empirical covariance smoother yields the convergence rate OP((logN)/(Nh4)+h2). The empirical covariance smoother’s convergence rate is thus slower than the one found in Corollary 1, achieved via the separable model.

4 Simulation Study

We explore the finite sample performance of the proposed methodology by means of a moderate simulation study (total runtime of about 1000 CPU hours). Computational efficiency (relatively small runtimes) is achieved by working on a 20 × 20 grid, like described in Appendix I.2, supplementary materials. Every surface X1,,X100 is first sampled fully on this grid as a zero-mean matrix-variate Gaussian with covariance C (to be specified), superposed with noise (zero-mean iid Gaussian entries with variance σ2), and then sub-sampled in a way that only a fraction of the entries, selected at random, is retained. The covariance C is always standardized to have trace one, and σ2 is chosen such that the gridded white noise process is also trace one.

Methods compared

We compare the proposed separable estimator ĉ=â·b̂ against the nonseparable empirical estimator obtained by local linear smoothing in four dimensions (4D smoothing), and also against the best separable approximation (BSA, Genton Citation2007) obtained from the fully observed and noise-free surfaces. We also compare the proposed estimator against its one-step version (ĉ=â0·b̂0, cf ), while it is shown in Appendix I.5, supplementary materials that more than two steps lead to a similar performance as the proposed two-step estimator.

Covariance choices

We consider four specific choices for the covariance: (a) a separable covariance with Fourier basis eigenfunctions and power decay of the eigenvalues; (b) Brownian sheet covariance, which is also separable; (c) the parametric covariance introduced by Gneiting (Citation2002), which is nonseparable; and (d) a superposition of the Fourier covariance from (a) with another separable covariance, having shifted Legendre polynomial eigenfunctions and power decay of the eigenvalues. We discuss these choices in detail in Appendix I.4, supplementary materials. However, for the purpose of presenting the simulation results, it is only important to point out the following. First, while the Fourier setup (a) and the Brownian setup (b) are separable, the Gneiting setup (c) and the Fourier-Legendre setup (d) are nonseparable. Second, while the Brownian setup (b) and Gneiting setup (c) lead to rather flat covariances, the Fourier setup (a) and the Fourier-Legendre setup (d) lead to quite wiggly (though infinitely smooth) covariances.

Sparsity

In all simulation setups, we consider different percentages of the entries observed p{2,5,10,20,40,70}. Since the grid size is 20 × 20, this means for example, that for p = 2 we have 2/100·202=8 observations per surface. For all the setups and percentages, we report relative estimation errors ||ĈC||2/||C||2, where Ĉ is an estimator obtained by one of the four methods above (4D smoothing, one-step, proposed, or BSA). The results are shown in . Every reported error was calculated as an average of 100 Monte Carlo runs.

Fig. 3 Relative estimation errors depending on percentages of the surfaces observed p for 4 ground truth covariance choices (a)–(d) and 4 methods compared. BSA provides a baseline, having access to full surfaces and hence not depending on p. For 4D smoothing, only results for small p are reported.

Fig. 3 Relative estimation errors depending on percentages of the surfaces observed p for 4 ground truth covariance choices (a)–(d) and 4 methods compared. BSA provides a baseline, having access to full surfaces and hence not depending on p. For 4D smoothing, only results for small p are reported.

Error components

The reported estimation errors can be thought of having four components: (i) asymptotic bias, which is zero if the true covariance is separable, that is, in cases (a) and (b); (ii) error due to finite number of samples (N = 100); (iii) error due to sparse observations (i.e., not observing the full surfaces); and (iv) noise contribution. BSA errors are always free of the latter two, providing a baseline. The effect of not observing the full surfaces is displayed for different values of p. Finally, the noise contamination prevents smoothing approaches to reach the performance of BSA even for p large. Although our methodology explicitly handles noise, the finite sample performance is better with noise-free data, which only BSA has the access to.

Results

There is a number of comments to be made about the results in :

  1. In the setups where the covariance is flat, that is, (b) and (c), the one-step and the proposed approaches work the same, and 4D smoothing also works relatively well. These two setups are simple in a sense, because information can be borrowed quite efficiently via smoothing, regardless of whether the truth is separable or not. Still, the proposed approach using separability does not perform worse than 4D smoothing even in the nonseparable case (c), having the advantage of being much faster to obtain. For p = 10 the proposed estimator takes only a couple of seconds while 4D smoothing takes about 40 min even at this relatively small size of data.

  2. When the true covariance is wiggly, the proposed methodology clearly outperforms 4D smoothing, both for the separable truth (a) and the nonseparable truth (d). The reason is that smoothing procedures are not very efficient in this case, and borrowing strength via separability is imperative.

  3. The reason why error curves for 4D smoothing are only calculated up to p = 10 is the computational cost of smoothing in higher dimensions. In fact, performing 4D smoothing when p = 10 took more time than calculating of all the remaining results combined (cross-validations included). The runtimes are reported in Appendix I.5, supplementary materials.

Remark 1.

We used a standard cross-validation strategy to choose bandwidths for the separable model. However, for 4D smoothing, cross-validation is not feasible—using it would increase the total runtime of our simulation study to over 1000 CPU days. Therefore, we chose bandwidths for 4D smoothing based on those cross-validated for a separable model. This approach is taken throughout the article. See Appendices I.2 and I.5 for details and an empirical demonstration of the effectiveness of this choice.

5 Data Analysis: Implied Volatility Surfaces

A call option is a contract granting its holder the right, but not the obligation, to buy an underlying asset, for example, a stock, for an agreed-upon strike price at a defined expiration time. The ratio between the strike price and the current asset price is called moneyness, and the call option is traded for a certain market price C*(log(m),τ), which depends on the moneyness m and the time to expiration τ. The Black–Sholes–Merton model (Black and Scholes Citation1973; Merton Citation1973) is commonly used to bijectively transform the observed market prices into implied volatilities, which can be thought of as an outlook of future fluctuations in the asset price. The advantage of considering the implied volatilities as opposed to the market prices of options themselves is that the implied volatilities tend to be smoother and comparable across different assets. Thus we can take advantage of the functional data analysis framework.

Since only certain combinations of the strike price and time to expiration are traded for a given asset, and these combinations differ across the assets, the implied volatility surfaces given as triplets of the moneyness, time to expiration and the corresponding implied volatility, that is (m,τ,σIV(m,τ)), can be treated as sparsely observed random surfaces. In this section, we consider the options data offered by DeltaNeutral (https://www.historicaloptiondata.com/content/free-data, retrieved on July 15, 2020), resulting in a dataset of N = 196 random surfaces. We take logarithms of moneyness and implied volatilities, and to reduce computational costs we round the log-moneyness and the time to expiration to fall on a common 50×50 grid.

For details on implied volatilities and Black–Scholes–Merton model, as well as the data preprocessing steps, see Appendix I, supplementary materials.

shows two observations in our sample. The snapshot of Qualcomm Inc (QCOM) features a cummulation of observation at the short expiration. Here, it happens five times that two raw observations fall in the same pixel on the common 50 × 50 grid. In these few cases we calculate the average of the two observations in each pair. Due to smoothness, the option prices (and hence the implied volatilities) attain very similar values and thus this rounding and averaging does not change the conclusions of our analysis. We have observed this fact by fitting the model on finer grids while the estimates remained similar.

Fig. 4 Two sample snapshots of the considered log implied volatility surfaces corresponding to the call options on the stocks of Dell Technologies Inc on January 19, 2006 (a) and Qualcomm Inc on February 07, 2018 (b), and the mean surface of the implied volatility gained from pooling all the data together (c).

Fig. 4 Two sample snapshots of the considered log implied volatility surfaces corresponding to the call options on the stocks of Dell Technologies Inc on January 19, 2006 (a) and Qualcomm Inc on February 07, 2018 (b), and the mean surface of the implied volatility gained from pooling all the data together (c).

The estimated mean surface is displayed on the right-hand side of . The mean surface captures the typical feature of the implied volatility surfaces: the volatility smile (Hull Citation2006). The implied volatility is typically greater for the options with moneyness away from 1, while this aspect is more significant for shorter times to expiration.

displays the estimates of the separable covariance components by our methodology presented in Section 2.3. The moneyness component demonstrates the highest marginal variability at the center of the covariance surface, meaning that the log implied volatility oscillates the most for the options with the log-moneyness around 0 (i.e., moneyness 1). The marginal variance is lower as the log-moneyness departs from 0. The eigendecomposition plot of the moneyness covariance kernel shows that the most of variability is explained by a nearly constant function with a small bump at the log-moneyness 0. The second leading eigenfunction adjusts the peak at the log-moneyness around 0 to a greater extent than the first eigenfunction. The covariance kernel corresponding to the time to expiration variable is smoother and demonstrates slightly higher marginal variability at shorter expiration. This phenomenon is well known for implied volatility (Hull Citation2006). The eigendecomposition of this covariance kernel indicates that the log implied volatility variation is mostly driven by the constant function while the second leading eigenfunction adjusts the slope of the surface for varying time to expiration.

Fig. 5 Top-left: The estimated covariance kernel â=â(t,t) corresponding to the time to expiration variable. Top-right: The three leading eigenfunctions of the spectral decomposition of the covariance kernel â=â(t,t). Bottom-left and botton-right: The same as above but for the estimated covariance kernel b̂=b̂(s,s) corresponding to the log-moneyness variable.

Fig. 5 Top-left: The estimated covariance kernel â=â(t,t′) corresponding to the time to expiration variable. Top-right: The three leading eigenfunctions of the spectral decomposition of the covariance kernel â=â(t,t′). Bottom-left and botton-right: The same as above but for the estimated covariance kernel b̂=b̂(s,s′) corresponding to the log-moneyness variable.

demonstrates our prediction techniques presented in Section 2.5 together with the 95% simultaneous prediction band. We recall that the prediction band aims to capture the latent smooth random surface itself, while our raw observations are modeled by adding an error term. Therefore, the raw data are not guaranteed to be covered in the band.

Fig. 6 Two views on prediction based on the call options written on the stock of Dell Technologies Inc on January 19, 2006. The circles depict the available sparse observations, the ribbons depict the predicted latent surface by the method of Section 2.5, where the covariance structure was assumed separable, and finally the transparent gray surfaces depict the 95% simultaneous prediction band for the latent log implied volatility surface.

Fig. 6 Two views on prediction based on the call options written on the stock of Dell Technologies Inc on January 19, 2006. The circles depict the available sparse observations, the ribbons depict the predicted latent surface by the method of Section 2.5, where the covariance structure was assumed separable, and finally the transparent gray surfaces depict the 95% simultaneous prediction band for the latent log implied volatility surface.

5.1 Quantitative Comparison

The prediction method outlined in Section 2.5 requires as an input the pairwise covariances regardless whether they have been estimated by the separable estimator â(t,t)b̂(s,s) or the 4D smoother ĉ(t,s,t,s). As the benchmark for our comparison, we choose the locally linear kernel smoother (Fan and Gijbels Citation1996) applied individually for each surface as such smoothers constitute a usual preprocessing step (Cont and Da Fonseca Citation2002). We will refer to this predictor as pre-smoothing. In this section, we demonstrate that the predictive performance is comparable for both covariance estimator strategies (the separable and the 4D smoother), and that both of such approaches are superior to pre-smoothing. Moreover, the separable smoother is substantially faster than the 4D smoother.

We compare the prediction error by performing a 10-fold cross-validation, where the covariance structure is fitted always on varying 90% of the surfaces, with the remaining 10% used for out-of-sample prediction. In the set that is held out for prediction, we select some of the sparse observations and predict them based on the remaining observations on that surface. We use the following hold-out patterns:

  1. Leave one chain out. Since the options are quotes always for a range of strikes, they constitute features known as option chains (see ) where multiple option prices (or equivalently implied volatilities) are available for a fixed time to expiration. For those surfaces that include at least two such chains, we remove gradually each chain and predict it based on the other chains. Therefore the number of prediction tasks performed on a single surface is equal to the number of chains observed per that surface.

  2. Predict in-the-money. Predict implied volatilities for below-average moneyness (i.e., moneyness m1) based on the out-of-the-money observations (moneyness m > 1).

  3. Predict out-of-the-money. Predict implied volatilities for above-average moneyness (i.e., moneyness m1) based on the in-the-money observations (moneyness m < 1).

  4. Predict short maturities. Predict the implied volatility for options with the time to maturity τ<183 [days] based on the implied volatility of the options with the time to maturity τ183 [days].

  5. Predict long maturities. Predict the implied volatility for options with the time to maturity τ>183 [days] based on the implied volatility of the options with the time to maturity τ183 [days].

All the prediction strategies are performed only for those surfaces where both the discarded part and the kept part are nonempty. We measure the prediction error on surface with the index n (in the test partition within the K-fold cross-validation) by the following root mean squared error (RMSE) criterion, relative to the pre-smoothing benchmark:(23) mMndiscarded((Π̂method(X(tnm,snm)|Ynkept))Ynm)2mMndiscarded((Π̂presmooth(X(tnm,snm)|Ynkept))Ynm)2(23) where Ynm,m=1,,Mn are the implied volatility observations on the n-th surface, Mndiscarded{1,,Mn} denotes the set of observations’ indices discarded for the n-th surface, Ynkept are the vectorized implied volatility observations that were kept to be conditioned on. The predictor Πpresmooth denotes the pre-smoothing based on the observations Ynkept). The predictors Πmethod for method being either the separable smoother or the 4D smoother constitute the proposed predictors in this article with the covariance structure estimated by either of the two smoothers. These predictors are always trained only on the training partition (90% of the surfaces) within the 10-fold cross-validation scheme. Note that this out-of-sample comparison is adversarial for the proposed approach, because when predicting a fixed surface, the measurements on that surface (and another 10% of measurements total) are not used for the mean and covariance estimation. In practice, we naturally use all available information. However, for the hold-out comparison study here, that would require frequent refitting of the covariance, which would not be computationally feasible, in particular for 4D smoothing.

presents the results under the five hold-out patterns in form of boxplots created from the relative errors (23). We see that the prediction errors based on estimated covariances, be it the separable smoother or the 4D smoother, are typically smaller than the presmoothing benchmark with the only exception of the 4D smoother in the hold-out pattern (b). The predictive performances of the separable and the 4D smoother are comparable, but they differ a lot in terms of runtime. It typically takes 30 seconds to calculate the separable smoother (including a cross-validation based selection of the smoothing bandwidths), while the 4D smoother takes around 3 hr. The latter runtime is moreover without considering any automatic selection of the bandwidths, because such would be computationally infeasible. Hence we use the bandwidths selected by the separable model, as described in Remark 1. The calculations are performed on a quite coarse grid of size 20 × 20. The calculations on a dense grid, such as 50 × 50 used in the qualitative analysis in Section 5 are not feasible for the 4D smoother.

Fig. 7 Boxplots of root mean squared errors relative to the pre-smoothing benchmark (23) for prediction method of Section 2.5 with the covariance estimated by 4D smoothing or the proposed separable approach, and different hold-out patterns: (a) leave one chain out; (b) predict in-the-money; (c) predict out-of-the-money; (d) predict short maturities; and (e) predict long-maturities. Numbers inside the boxes provide numerical values of the median. For a given method, RMSE value 3 means that the given method is 3-times worse than the benchmark, while RMSE value 1/3 corresponds to 3-fold improvement.

Fig. 7 Boxplots of root mean squared errors relative to the pre-smoothing benchmark (23) for prediction method of Section 2.5 with the covariance estimated by 4D smoothing or the proposed separable approach, and different hold-out patterns: (a) leave one chain out; (b) predict in-the-money; (c) predict out-of-the-money; (d) predict short maturities; and (e) predict long-maturities. Numbers inside the boxes provide numerical values of the median. For a given method, RMSE value 3 means that the given method is 3-times worse than the benchmark, while RMSE value 1/3 corresponds to 3-fold improvement.

Therefore we conclude that the separable smoother approach enjoys a better predictive performance than the pre-smoothing benchmark and—while having similar predictive performance as the predictor based on 4D smoothing—is computationally much faster than the said competitor. In fact, it requires two-dimensional smoothers only, just as the pre-smoothing benchmark.

6 Conclusions and Future Directions

In practice, covariances are often nonseparable (Gneiting, Genton, and Guttorp Citation2006; Aston, Pigoli, and Tavakoli Citation2017; Bagchi and Dette Citation2020; Dette, Dierickx, and Kutta Citation2020; Pigoli et al. Citation2018; Rougier Citation2017), and assuming separability induces a bias. The variance stemming from sparse measurements and noise contamination is, however, often of a of larger magnitude, thus sanctioning separability as a means to achieve a better bias-variance tradeoff. Moreover, separability entails faster computation and lower storage requirements. As demonstrated above, these advantages are pronounced in the sparse regime.

When data are observed fully and no smoothing is used (i.e., simple averages are calculated at the grid points instead), calculation of the estimator takes about 0.03 sec in case of separability and 0.12 sec in general, with our grid size 20 and sample size N = 100. With larger grid size, the acceleration gained by assuming separability naturally magnifies. But still, separability is assumed mostly to save memory rather than time, in the fully observed case (Aston, Pigoli, and Tavakoli Citation2017). On the other hand, with sparsely observed data and kernel smoothing deployed, the speed-up factor stemming from separability can be of the order of hundreds already with rather small grid sizes. The point we advocate is that—with sparse observations, kernel smoothing deployed, and compared to observations on a grid—the computational savings offered by separability are the same in terms of memory, but are much more profound in terms of runtime. This is due to the extra costs associated with smoothing.

Our analysis of implied volatility surfaces provides some insights into the statistical dependencies of such sparsely observed data, and our quantitative comparison demonstrates how the proposed prediction method outperforms the pre-smoothing benchmark. We have formed our dataset collecting options across various symbols and timestamps with the reasoning that the volatility across these come also from one population, which is debatable. This study should be seen rather as a proof-of-concept on how to “borrow strength” across the dataset, and how this approach can be used to decrease the prediction error.

Having said that, we believe that our methodology can provide even better results when considering a more homogeneous population such as the time series of the implied volatility surfaces related to a single fixed symbol/asset. In this case, pre-smoothing is usually required (Cont and Da Fonseca Citation2002; Kearney, Cummins, and Murphy Citation2018) for forecasting of such time series. Our methodology could avoid the pre-smoothing step in this case by predicting the surfaces while borrowing the information across the entire dataset. Furthermore, our methodology could be easily tailored to predicting principal components scores by conditional expectation (similarly to Yao, Müller, and Wang Citation2005), which could be used for forecasting by a vector autoregression. Combining separable covariances and time series of sparsely observed surfaces hints at other directions of future work. For example, Rubín and Panaretos (Citation2020) showed how to estimate the spectral density operator nonparametrically from sparsely observed functional time series. Estimating separable spectral density operator for sparsely observed surface-valued time series seems to be within reach and is likely to provide predictions that benefit from the information across time, and thus, reducing the prediction error even more.

While our methodology is aimed at sparsely observed surfaces, it can be also used for unbalanced regimes, where some surfaces are observed only at few locations while others at many more. It is natural to ask whether an additional “per-subject” weighting scheme such as the one proposed by Li and Hsing (Citation2010) could be generalized to the multi-dimensional case and lead to a better performance. The motivation here is to take into account different levels of information contained in a single point on a sparsely observed surface compared to a single point on a densely observed surface. We provide a detailed discussion of this direction in Appendix III, supplementary materials.

Supplemental material

Supplemental Material

Download Zip (4.9 MB)

Acknowledgments

The authors thank the editor, the associate editor, and two anonymous referees for their valuable feedback.

Supplementary Materials

Appendix:The Appendix contains all the proofs and some additional results, as referenced from the main paper. (appendix.pdf)

R code:The available R code implements the methodology proposed in the main paper, and can be used to reproduce the reported results. (see README.txt)

Additional information

Funding

This research was supported by a Swiss National Science Foundation grant.

References

  • Aston, J. A., Pigoli, D., and Tavakoli, S. (2017), “Tests for Separability in Nonparametric Covariance Operators of Random Surfaces,” The Annals of Statistics, 45, 1431–1461. DOI: 10.1214/16-AOS1495.
  • Bagchi, P., and Dette, H. (2020), “A Test for Separability in Covariance Operators of Random Surfaces,” Annals of Statistics, 48, 2302–2322.
  • Black, F., and Scholes, M. (1973), “The Pricing of Options and Corporate Liabilities,” Journal of Political Economy, 81, 637–654. DOI: 10.1086/260062.
  • Chen, L.-H., and Jiang, C.-R. (2017), “Multi-Dimensional Functional Principal Component Analysis,” Statistics and Computing, 27, 1181–1192. DOI: 10.1007/s11222-016-9679-5.
  • Constantinou, P., Kokoszka, P., and Reimherr, M. (2017), “Testing Separability of Space-Time Functional Processes,” Biometrika, 104, 425–437.
  • Cont, R., and Da Fonseca, J. (2002), “Dynamics of Implied Volatility Surfaces,” Quantitative Finance, 2, 45–60. DOI: 10.1088/1469-7688/2/1/304.
  • De Iaco, S., Myers, D. E., and Posa, D. (2002), “Nonseparable Space-Time Covariance Models: Some Parametric Families,” Mathematical Geology, 34, 23–42. DOI: 10.1023/A:1014075310344.
  • Degras, D. A. (2011), “Simultaneous Confidence Bands for Nonparametric Regression with Functional Data,” Statistica Sinica, 21, 1735–1765. DOI: 10.5705/ss.2009.207.
  • Dette, H., Dierickx, G., and Kutta, T. (2020), “Quantifying Deviations from Separability in Space-Time Functional Processes,” arXiv preprint arXiv:2003.12126.
  • Fan, J., and Gijbels, I. (1996), Local Polynomial Modelling and Its Applications: Monographs on Statistics and Applied Probability (Vol. 66), Boca Raton, FL: CRC Press.
  • Fengler, M. R. (2009), “Arbitrage-Free Smoothing of the Implied Volatility Surface,” Quantitative Finance, 9, 417–428. DOI: 10.1080/14697680802595585.
  • Genton, M. G. (2007), “Separable Approximations of Space-Time Covariance Matrices,” Environmetrics: The Official Journal of the International Environmetrics Society, 18, 681–695. DOI: 10.1002/env.854.
  • Gneiting, T. (2002), “Nonseparable, Stationary Covariance Functions for Space–Time Data,” Journal of the American Statistical Association, 97, 590–600. DOI: 10.1198/016214502760047113.
  • Gneiting, T., Genton, M. G., and Guttorp, P. (2006), Geostatistical Space-Time Models, Stationarity, Separability, and Full Symmetry, Boca Raton, FL: Chapman and Hall/CRC, pp. 151–175.
  • Hall, P., Müller, H.-G., and Wang, J.-L. (2006), “Properties of Principal Component Methods for Functional and Longitudinal Data Analysis,” The Annals of Statistics, 34, 1493–1517. DOI: 10.1214/009053606000000272.
  • Henderson, C. R. (1975), “Best Linear Unbiased Estimation and Prediction Under a Selection Model,” Biometrics, 31, 423–447. DOI: 10.2307/2529430.
  • Hull, J. (2006), Options, Futures, and Other Derivatives (Pearson International ed.), New York: Pearson/Prentice Hall.
  • Kearney, F., Cummins, M., and Murphy, F. (2018), “Forecasting Implied Volatility in Foreign Exchange Markets: A Functional Time Series Approach,” The European Journal of Finance, 24, 1–18. DOI: 10.1080/1351847X.2016.1271441.
  • Li, Y., and Hsing, T. (2010), “Uniform Convergence Rates for Nonparametric Regression and Principal Component Analysis in Functional/Longitudinal Data,” The Annals of Statistics, 38, 3321–3351. DOI: 10.1214/10-AOS813.
  • Lopez, G., Eisenberg, D. P., Gregory, M. D., Ianni, A. M., Grogans, S. E., Masdeu, J. C., Kim, J., Groden, C., Sidransky, E., and Berman, K. F. (2020), “Longitudinal Positron Emission Tomography of Dopamine Synthesis in Subjects with gba1 Mutations,” Annals of Neurology, 87, 652–657. DOI: 10.1002/ana.25692.
  • Masak, T., and Panaretos, V. M. (2019), “Spatiotemporal Covariance Estimation by Shifted Partial Tracing,” arXiv preprint arXiv:1912.12870.
  • Masak, T., Sarkar, S., and Panaretos, V. M. (2020), “Principal Separable Component Analysis via the Partial Inner Product,” arXiv preprint arXiv:2007.12175.
  • Merton, R. C. (1973), “Theory of Rational Option Pricing,” The Bell Journal of Economics and Management Science, 4, 141–183. DOI: 10.2307/3003143.
  • Pigoli, D., Hadjipantelis, P. Z., Coleman, J. S., and Aston, J. A. (2018), “The Statistical Analysis of Acoustic Phonetic Data: Exploring Differences Between Spoken Romance Languages,” Journal of the Royal Statistical Society, Series C, 67, 1103–1145. DOI: 10.1111/rssc.12258.
  • Ramsay, J., and Silverman, B. (2007), Applied Functional Data Analysis: Methods and Case Studies, Springer Series in Statistics, New York: Springer.
  • Ramsay, J. O., and Silverman, B. W. (2005), Functional Data Analysis, New York: Springer.
  • Rougier, J. (2017), “A Representation Theorem for Stochastic Processes with Separable Covariance Functions, and Its Implications for Emulation,” arXiv preprint arXiv:1702.05599.
  • Rubín, T., and Panaretos, V. M. (2020), “Sparsely Observed Functional Time Series: Estimation and Prediction,” Electronic Journal of Statistics, 14, 1137–1210. DOI: 10.1214/20-EJS1690.
  • Wang, J., Wong, R. K., and Zhang, X. (2020), “Low-Rank Covariance Function Estimation for Multidimensional Functional Data,” Journal of the American Statistical Association, 1–14. DOI: 10.1080/01621459.2020.1820344.
  • Yao, F., Müller, H.-G., and Wang, J.-L. (2005), “Functional Data Analysis for Sparse Longitudinal Data,” Journal of the American Statistical Association, 100, 577–590. DOI: 10.1198/016214504000001745.
  • Yarger, D., Stoev, S., and Hsing, T. (2020), “A Functional-Data Approach to the Argo Data,” arXiv preprint arXiv:2006.05020.
  • Zhang, H., and Li, Y. (2020), “Unified Principal Component Analysis for Sparse and Dense Functional Data Under Spatial Dependency,” arXiv preprint arXiv:2006.13489.
  • Zhang, X., and Wang, J.-L. (2016), “From Sparse to Dense Functional Data and Beyond,” The Annals of Statistics, 44, 2281–2321. DOI: 10.1214/16-AOS1446.