378
Views
0
CrossRef citations to date
0
Altmetric
Articles

FragmGAN: generative adversarial nets for fragmentary data imputation and prediction

&
Pages 15-28 | Received 25 Dec 2022, Accepted 12 Oct 2023, Published online: 27 Oct 2023

Abstract

Modern scientific research and applications very often encounter ‘fragmentary data’ which brings big challenges to imputation and prediction. By leveraging the structure of response patterns, we propose a unified and flexible framework based on Generative Adversarial Nets (GAN) to deal with fragmentary data imputation and label prediction at the same time. Unlike most of the other generative model based imputation methods that either have no theoretical guarantee or only consider Missing Completed At Random (MCAR), the proposed FragmGAN has theoretical guarantees for imputation with data Missing At Random (MAR) while no hint mechanism is needed. FragmGAN trains a predictor with the generator and discriminator simultaneously. This linkage mechanism shows significant advantages for predictive performances in extensive experiments.

1. Introduction

Modern scientific research and applications very often encounter data from multiple data sources, and for each data source, various variables can be collected for data analysis. Such increasing data sources bring big opportunities for predicting people's behaviours with huge potential social and commercial benefits. However, these different data sources usually can not be available for every sample, which leads to ‘fragmentary data’ and brings big challenges to data imputation and label prediction. To be more specific, we introduce two motivating examples that represent the most typically practical scenarios for fragmentary data.

Internet Loan: A leading company of wealth management is exploring its internet loan business and trying to predict the applicants' income for risk management purpose. There are five possibly available data sources (Table ). (i) Card: the credit card information; (ii) Shopping: the shopping history at internet; (iii) Mobile: the monthly bill of mobile phone; (iv) Bureau: the credit report from the Central Bank; (v) Fraud: the information from an anti-fraud platform. However, some applicants are not willing to provide their shopping or mobile information, not all the applicants have credit reports, and many of them are never included in the database of the anti-fraud platform. As a result, there are 10 ‘response patterns’ in the Internet Loan data as shown in Table , where ‘√’ means the data source is available for the applicants with the corresponding response pattern.

Table 1. The response patterns of the Internet Loan data.

ADNI: The Alzheimer's Disease Neuroimaging Initiative http://adni.loni.usc.edu is a widely used data by researchers for the Alzheimer's disease which has four data sources. (i) CSF: cerebrospinal fluid; (ii) PET: positron emission tomography; (iii) MRI: magnetic resonance imaging; (iv) Gene: the gene expression. As show in Table , it has 8 different response patterns corresponding to different data availability for each data source.

Table 2. The response patterns of the ADNI data.

Such kind of fragmentary data, also known as ‘block-wise missing data’ in the statistics literature, are very common in the area of risk management, marketing research, social sciences, medical studies and so on. Data imputation and label prediction are two main goals for the analysis of such data. But the extremely high missing rate and complicated missing patterns bring big challenges to the achievement of the goals.

Some work has been done to deal with fragmentary data in both areas of statistics and computer sciences in recent years. From the statistics perspective, methods based on model averaging (Fang et al., Citation2019), factor models (Zhang et al., Citation2020), generalized methods of moments (Xue & Qu, Citation2021), iterative least squares (Lin et al., Citation2021) and integrative factor regression (Li & Li, Citation2021) are proposed. These statistical methods provide useful theoretical properties but exhibit notable shortcomings. (i) They depend on certain statistical models, for example, linear regression models. (ii) They are not flexible in handling mixed data types that include continuous and categorical variables. (iii) Only a couple of methods consider imputation and prediction at the same time.

From the computer science perspective, GAIN (Yoon et al., Citation2018) first uses a Generative Adversarial Net (GAN) to impute data Missing Completed At Random (MCAR), which means the missingness occurs entirely at random without depending on any of the variables. MisGAN (Li et al., Citation2019) trains a mask generator along with the data generator for imputation. GAMIN (Yoon & Sull, Citation2020) proposes a generative adversarial multiple imputation network for highly missing data. HexaGAN (Hwang et al., Citation2019) deals with missing data imputation, conditional generation and semi-supervised learning together. GRAPE (You et al., Citation2020) proposes a graph-based framework for data imputation and label prediction. MIWAE (Mattei & Frellsen, Citation2019) and Not-MIWAE (Ipsen et al., Citation2021) propose imputation methods based on variational auto-encoding (VAE) framework instead of GAN. Ma and Chen (Citation2019) proposes a matrix completion algorithm when the data is missing not at random. However, these generative methods have various drawbacks. For instance, some of them (Ipsen et al., Citation2021; Mattei & Frellsen, Citation2019; Yoon & Sull, Citation2020; You et al., Citation2020) do not have the theoretical guarantee that the imputed data has the same distribution as the original data. Some of them (Hwang et al., Citation2019; Li et al., Citation2019; Yoon et al., Citation2018) only have theoretical results for data MCAR, which is highly unlikely in the practice. Most of them either consider data imputation and label prediction separately or only consider data imputation.

In this paper, by leveraging the structure of response patterns, we propose a ‘FragmGAN’ for fragmentary data imputation and prediction. The main contributions are as follows.

  • FragmGAN is a unified framework based on GAN to deal with fragmentary data imputation and label prediction at the same time. It's flexible in the sense that (i) it's applicable to both continuous and categorical data and label, and (ii) users can adjust the relative importance of the task of imputation to prediction by an ‘adjusting factor’.

  • FragmGAN has theoretical guarantees for imputation with data Missing At Random (MAR), which is much more general than MCAR and will be defined in Section 3.2. Also, the theoretical results do not need a hint mechanism that is required by GAIN.

  • Using similar technical skills, we extend the theoretical results of GAIN to MAR.

  • Other than the generator and discriminator, FragmGAN trains a predictor simultaneously. This linkage mechanism shows significant advantages for predictive performances in extensive experiments.

2. Related work

There are lots of discriminative and generative imputation methods that will be considered in our experiments, including Expectation Maximization (García-Laencina et al., Citation2010), matrix completion (Mazumder et al., Citation2010), MICE (van Buuren & Groothuis-Oudshoorn, Citation2011), MissForest (Stekhoven & Buhlmann, Citation2011) and Auto-Encoder (Gondara & Wang, Citation2017).

There are several other GAN based imputation methods. CollaGAN (Lee et al., Citation2019) proposes a collaborative GAN for missing data imputation but it focuses on image data. WGAIN (Friedjungová et al., Citation2020), CGAIN (Awan et al., Citation2021), PC-GAIN (Wang et al., Citation2021) and S-GAIN (Neves et al., Citation2021) extend GAIN in various ways. IFGAN (Qiu et al., Citation2020) conducts missing data imputation using a feature-specific GAN and MCFlow (Richardson et al., Citation2020) proposes a Monte Carlo flow method for data imputation but no theoretical result is provided. When all the variables are assumed to be categorical, theoretical results of GAN based methods are extended to an uncommon concept of Extended Always Missing At Random (Deng et al., Citation2020).

Although they are not our main interest, we also mention some other VAE based imputation methods including VAEAC (Ivanov et al., Citation2019), variational inference of deep subspaces (Dalca et al., Citation2019), iterative imputation using AE dynamics (Smieja et al., Citation2020), VAE using pattern-set mixtures (Ghalebikesabi et al., Citation2021) and VSAE (Gong et al., Citation2021). Some of them only focus on image data. A common disadvantage of VAE based methods is the lack of theoretical guarantee for imputation. Some results of empirical comparison of GAN and VAE based methods are presented in Camino et al. (Citation2019).

3. GAN-Based fragmentary data imputation

We first formulate the problem and discuss the method and theory of fragmentary data imputation in this section. The problem of label prediction will be addressed in Section 4.

Throughout the paper we usually use bold type letters to denote vectors and use the regular letters for scalars. The upper-case letters are used for random variables and the corresponding lower-case letters are their realizations. Abusing notation slightly, we use a generic notation p() or p(|) to denote the distribution/probability or conditional distribution/probability for various continuous/categorical variables as long as there is no ambiguity.

3.1. Imputation method

Let X=(X1,,Xd) be the d-dimensional data vector of interested variables that could take continuous or categorical values. Note that d is the number of variables but not the number of data sources since each data source may have multiple variables.

Define the mask vector M=(M1,,Md){0,1}d such that Mi=1 means Xi is observed and Mi=0 means Xi is missing, i=1,,d. So what we actually observe is X~=MX=(M1X1,,MdXd),where ⊙ denotes element-wise multiplication.

Assume overall there are K (a fixed number) possible response patterns in the data and define W=(W1,,WK) as the pattern indicator, where Wk=1 if the sample belongs to the kth response pattern and Wk=0 otherwise, k=1,,K. Note that k=1KWk=1. In the fragmentary data setting, M can actually only take K (rather than 2d) different values and there is a one-to-one mapping between M and W. In the two motivating examples, K = 10 and 8 respectively.

Generator

Let Z=(Z1,,Zd) be a d-dimensional noise vector that is independent of all other variables. It is typically taken as Gaussian white noise. We then feed X~=MX, Z and W into the generator G and obtain X¯=G(MX,(1M)Z,W),where G is a function from Rd×Rd×{W(1),,W(K)} to Rd, and {W(1),,W(K)} are the K possible values of W. X¯ is the generated data vector but we are only interested in the missing variables. So the complete data vector after imputation is Xˆ=MX~+(1M)X¯=MX+(1M)X¯.Our target is to make sure the distribution of Xˆ is the same as the distribution of X, i.e., p(Xˆ)=p(X). The randomness of Z makes our method a random imputation method rather than fixed imputation. Although we focus on single imputation in the paper, but by modelling the distribution of the data, we are able to make multiple imputation capture the uncertainty for the imputation value (Rubin, Citation2004; van Buuren & Groothuis-Oudshoorn, Citation2011).

Discriminator

The discriminator D tries to figure out which part of Xˆ is from the generator. The vanilla GAIN (Yoon et al., Citation2018) aims to distinguish each component of Xˆ is real (observed) or fake (imputed). It's a hard task since d is usually a large number. Consequently, a hint mechanism, which reveals all but one of the components of M to D, is required for GAIN to solve the model identifiability problem and make sure the generated distribution is what we want.

In the fragmentary data setting, each sample should exactly belong to one of the K response patterns. By leveraging this informative structure, our discriminator D just needs to figure out which pattern Xˆ belongs to. So D is a function from Rd to [0,1]K (instead of [0,1]d in GAIN) such that Wˆ=D(Xˆ)=(Wˆ1,,WˆK)is the predicted probability vector for W, where Wˆk is the predicted probability that Xˆ is from the kth response pattern and k=1KWˆk=1. Note that the output layer of D has a softmax form.

We train the discriminator D to maximize the probability of correctly predicting W. On the other hand, the generator G is trained to minimize the probability of D correctly predicting W. The objective function is defined to be the negative cross-entropy loss (1) V(G,D)=E(Xˆ,W)[k=1KWklogDk(Xˆ)],(1) where Dk(Xˆ) is just Wˆk. Note that the objective function depends on G through Xˆ. Then the minimax problem is given by (2) minGmaxDV(G,D).(2)

Remark 3.1

The key difference of our imputation method to GAIN is that we use a different objective function by taking the response patterns into consideration. This adjustment makes sure the model is identifiable even no hint mechanism is used as we show in the next subsection.

3.2. Theoretical results

Most previous theoretical results for GAN-based imputation methods including GAIN (Yoon et al., Citation2018), MisGAN (Li et al., Citation2019) and HexaGAN (Hwang et al., Citation2019) are established under the MCAR assumption, which means the missingness occurs entirely at random without depending on any of the variables. This is a very restrictive assumption and rarely satisfied in the real world. In contrast, our theoretical results will be established under the MAR assumption.

Assume X can be decomposed into (Xo,Xm), where Xo is an always observed subvector of X, and Xm could be missing. The missing mechanism is characterized (Little & Rubin, Citation2014) into three types.

  • Missing Completed At Random (MCAR): M is independent of X.

  • Missing At Random (MAR): p(M|X)=p(M|Xo), or equivalently, M is conditionally independent of Xm given Xo.

  • Missing Not At Random (MNAR): p(M|X) depends on Xm.

Remark 3.2

For a random vector X, it could be ambiguous for the definition of MAR. Another way to define MAR is p(M|X)=p(M|{Xi:Mi=1,i=1,,d}). However, since M appears in both sides of the condition, there is no way to generate a group of independently and identically distributed samples satisfying this equation, unless there exists an always observed subvector Xo such that p(M|X)=p(M|Xo). This is the reason why we use the MAR definition as above.

The complete data vector Xˆ can be decomposed into (Xˆo,Xˆm) correspondingly. Note that Xˆo=Xo. So p(Xˆ)=p(Xˆo)p(Xˆm|Xˆo)=p(Xo)p(Xˆm|Xo).To verify that the solution to the minimax problem (Equation2) satisfies p(Xˆ)=p(X), we just need to show p(Xˆm|Xo)=p(Xm|Xo). First we present a lemma.

Lemma 3.1

Let xˆ is a realization of Xˆ such that p(xˆ)>0. For a fixed generator G, the kth component of the optimal discriminator D(xˆ) to the minimax problem (Equation2) is given by Dk(xˆ)=p(W=wk0|xˆ)for k=1,,K, where wk0=(0,,1,,0) is a K-dimensional vector with only the kth element being 1, and W=wk0 means that the sample belongs to the kth response pattern.

Proof.

All proofs are provided in A.1.

We now rewrite (Equation1) by substituting D to obtain the objective function for G to minimize C(G)=V(D,G)=E(Xˆ,W)[k=1KWklogp(W=wk0|Xˆ)].

Theorem 3.2

A global minimum for C(G) is achieved if and only if (3) p(xˆm|xo,W=wk0)=p(xˆm|xo)(3) for each k{1,,K} and xˆ=(xo,xˆm) such that p(xˆ)>0 and p(xo|W=wk0)>0.

It's worthy to mention that Lemma 3.1 and Theorem 3.2 do not depend on the MAR assumption and they are generally true even under MNAR.

Theorem 3.2 tells us that the optimal generator will generate data so that the conditional distributions of Xˆm given Xo across different response patterns are the same. But it does not guarantee p(Xˆm|Xo)=p(Xm|Xo) yet.

To further explore, we assume the first response pattern is the case that all the variables are observed, i.e., Mi=1 for all i{1,,d}. Note the first response patterns in the two motivating examples are exactly the case. Then given W=w10, there is no missing variable and we have Xˆm=Xm. So following (Equation3), we have (4) p(Xˆm|Xo)=p(Xˆm|Xo,W=w10)=p(Xm|Xo,W=w10).(4)

Under the MAR assumption, M is conditionally independent of Xm given Xo, and so is W since there is a one-to-one mapping between M and W. Therefore (5) p(Xm|Xo,W=w10)=p(Xm|Xo).(5) Combining (Equation4) and (Equation5) gives us the final theorem that provides theoretical guarantees for our proposed imputation method.

Theorem 3.3

Under the MAR assumption, the density solution to (Equation3) is unique and satisfies p(xˆm|xo)=p(xm|xo).So the distribution of Xˆ is the same as the distribution of X.

This theorem tells us that the optimal solution is uniquely identified and it is the one we need. Compared to GAIN (Yoon et al., Citation2018), our method does not need a hint mechanism for model identifiability. An intuitive explanation is that we just need to classify each sample into one of the K response patterns. It requires much fewer model parameters than GAIN, in which d binary classifiers need to be modelled if the hint mechanism is not applied. Note that we only require K to be a fixed number and it could be as large as 2d. So theoretically FragmGAN can be applied to any kind of dataset with arbitrary missing patterns when the data vector is low-dimensional, for example, d = 4, and the associated computation is not too heavy. However, FragmGAN is not necessary better than GAIN in such cases. The most suitable scenario for FragmGAN is when K is relatively small compared to 2d.

Our theoretical results are established under MAR assumption while the vanilla GAIN (Yoon et al., Citation2018) assumes MCAR. However, we find that GAIN (with hint) also guarantees that p(Xˆ)=p(X) under the MAR assumption, which is consistent to a recent theoretical result (Deng et al., Citation2020) considering a special case that all the variables are categorical. We provide a direct proof of this conclusion of GAIN in A.2.

4. A unified framework for imputation and prediction

Many previous methods including GAIN (Yoon et al., Citation2018) consider label prediction as a post-imputation problem, that is, they first impute the data and then develop a prediction model as if the data were fully observed. The disconnection between imputation and prediction mostly likely damages the accuracy of prediction. In this section we propose a unified framework that considers data imputation and label prediction together. The key idea is to train a predictor P with the generator and discriminator simultaneously.

Predictor

Let Y be the interested q-dimensional label that could be continuous or categorical. Unlike the semi-supervised learning, the label Y is assumed to be available for all the training samples. A predictor P is a function from Rd to Rq such that Yˆ=P(Xˆ) is a predicted value of Y.

To evaluate the prediction performance of P, we define a loss function L(Y,P(Xˆ)) where L is from Rd×Rd to R. The explicit form of L depends on the data type of Y and is very flexible. For example, if Y is continuous, we may use L(Y,P(Xˆ))=YP(Xˆ)2. If Y{0,1} is a binary scalar and the predicted value is the probability of being 1, then we may use L(Y,P(Xˆ))=YlogP(Xˆ)(1Y)log(1P(Xˆ)).

To train G, D and P together, define the linked objective function as (6) U(G,D,P)=γV(G,D)+(1γ)E(Y,Xˆ)L(Y,P(Xˆ)),(6) where V(G,D) is from (Equation1) and γ[0,1] is an ‘adjusting factor’ that controls the relative importance of data imputation to label prediction.

The second part of (Equation6) does not involve D, so the target of D is still to maximize V(G,D). The first part of (Equation6) does not involve P, so the target of P is to minimize the predictive loss E(Y,Xˆ)L(Y,P(Xˆ)). Both parts of (Equation6) involve G, but fortunately they both require G to minimize. So the minimax optimization problem is given by (7) minPminGmaxDU(G,D,P).(7)

The choice of γ is quite flexible. If the user is just interested in data imputation, he can take γ=1 and U(G,D,P) is reduced to V(G,D). If the user is mainly interested in label prediction, he may use a cross-validation procedure to choose an appropriate γ or simply take γ=0.5 which works quite well as shown in the experiments. Note that γ=0 is not a good choice since it will lead to overfitting. If the user cares about both imputation and prediction, he may decide γ by the relative importance of the two tasks in his mind.

The pseudo code to implement (Equation7) is given in Algorithm 1. Several issues are discussed as follows.

First, although the hint mechanism is not required for our theoretical results, it is still empirically helpful. So we also use the same hint mechanism as Yoon et al. (Citation2018) in implementation. The impact of including the hint mechanism or not will be checked in the experiments.

Second, the generator also generates data even for the observed variables, which can be used to check the generation performance. An extra loss function LM:{0,1}d×Rd×RdR defined as αi=1dMiLM(X~i,X¯i) is added to V(G,D) for training G, where LM:R×RR is a user-specified loss function depending on the variable type of Xi. The algorithm result is not sensitive to the choice of hyper-parameter α. Actually, as long as α is relatively large (α=10 in the experiments), its main effect is to force X¯i=Xi for the variable with Mi=1.

Third, when γ=1, Algorithm 1 actually implements (Equation2) and the post-imputation prediction.

5. Experiments

In this section we check the imputation and prediction performance of FragmGAN in multiple datasets.

First we consider five UCI datasets (Lichman, Citation2013) used in GAIN (Yoon et al., Citation2018): Breast, Spam, Letter, Credit and News. Since the original datasets do not have any missing value, we randomly remove part of data by variable groups to make it fragmentary. Unless otherwise stated, the miss rate is 20%. By designing the removing strategy, we can make it MCAR or MAR. Specifically, we manually set several response patterns in advance. For MCAR, each sample is assigned to the patterns totally at random. For MAR, the probability of each sample being assigned to each pattern depends on the always observed covariate vector Xo. For this group of datasets, we are able to check the performance of data imputation along with label prediction since the true data values are known.

Then we consider two datasets Internet Loan and ADNI for the motivating examples introduced in Section 1. The miss rates of them are 46.6% and 22.3%, respectively. More details of these two datasets are provided in A.3. Since the missing values are unknown, we can only check the label prediction performance for these two datasets.

For the purpose of comparison, we consider MICE, MissForest, matrix completion (Matrix), Auto-Encoder (AE), Expectation Maximization (EM) and MisGAN that have been mentioned in Section 1. For the prediction task for Internet Loan and ADNI, we also consider two statistical methods: Model Averaging (Fang et al., Citation2019) and FR-FI (Zhang et al., Citation2020).

The hyperparameters of FragmGAN and some implementation details are provided in A.3.

For each dataset, we randomly split it into a training set (80%) and a test set (20%) by response patterns. All the methods are fitted in the training set and then applied to the test set. The imputation and prediction performances are evaluated at the test set. We repeat this experiment 10 times and report the averages and standard deviations of the evaluation criteria (RMSE or AUC). In each table, the best result for each dataset is marked in bold type.

5.1. Results for the UCI datasets

Imputation Performance. Table  reports the RMSEs of the imputation errors for the UCI datasets. We take γ=1 for FragmGAN since imputation is the focus here. For both FragmGAN and GAIN, we consider two versions with or without the hint mechanism.

Table 3. Imputation performance for UCI datasets in terms of RMSE (Average ± Std) of imputation error.

As we can see from Table , FragmGAN outperforms all the other methods in most cases. For the two cases that FragmGAN is not the best (Breast and Letter with MAR, in which MissForest performs the best), it performs the second best. Both FragmGAN and GAIN perform better than their corresponding versions without hint, indicating that the hint mechanism really helps empirically. This is expected since the hint mechanism provides useful information to the discriminator. Note that the results here can not be directly compared to the results in the paper of GAIN (Yoon et al., Citation2018) since here we consider fragmentary data with certain response patterns while the missing data in GAIN is generated totally at random.

To check the imputation performance under different miss rates, we take the dataset Credit and generate missing data with miss rate from 10% to 80%. Figure  presents the RMSEs of imputation errors under different miss rates. We can see that FragmGAN consistently performs the best. Again, both FragmGAN and GAIN perform better than their corresponding versions without hint. FragmGAN outperforms GAIN in both versions with or without hint.

Figure 1. RMSE of imputation error of the Credit data under different miss rates. Left: MCAR. Right: MAR.

Figure 1. RMSE of imputation error of the Credit data under different miss rates. Left: MCAR. Right: MAR.

Overall speaking, FragmGAN performs quite well in data imputation in the sense that it has smaller RMSE of imputation error compared to the competitors. Specifically it is better than GAIN, indicating that considering the structure of response patterns in the algorithm is really useful.

Prediction Performance. Table  reports the AUCs for the prediction performance in the datasets Breast, Spam, Credit and News. The dataset Letter is not considered here since it does not have a binary label. We include hint for both FragmGAN and GAIN. The adjusting factor γ is taken as 1 or 0.5 for FragmGAN. Note that when γ=1, FragmGAN first imputes data and then makes the prediction as if the data were fully observed. When γ=0.5, the imputation and prediction are considered simultaneously.

Table 4. Prediction performance for UCI datasets in terms of AUC (Average ± Std).

As we can see, FragmGAN with γ=0.5 outperforms the other methods in all the cases. This result shows that the linkage mechanism of training generator and predictor together can improve the prediction performance as we expected. Also note that although FragmGAN with γ=1 performs worse than FragmGAN with γ=0.5, it still performs better than all the other methods.

5.2. Results for the motivating examples

For the dataset Internet Loan, the original label is the applicant's income, which is a continuous variable. In the analysis we use log(income) as the label Y. For the dataset ADNI, the original label Y is the score of Mini-Mental State Examination (MMSE) taking value from 0 to 30, in which higher score means better cognitive function. In the real analysis, we consider two labels. (i) The normalized MMSE which can be considered as a continuous variable. (ii) A binary label Y = 1 if MMSE≥28 and Y = 0 otherwise.

Prediction Performance. Table  reports the RMSEs for the continuous label prediction and AUCs for the binary label prediction. The last two methods (Model Averaging and FR-FI) rely on linear regression models so they are not applicable to the binary label prediction. For the proposed FragmGAN, we take γ=1, 0.75, 0.5 and 0.25, indicating different relative importance of imputation to prediction. Also, we use a 5-fold cross-validation to select the best (for label prediction) γ. The CV criterion is defined as the averaged prediction performance in the leave-out samples.

Table 5. Prediction performance for the two motivation examples (Average ± Std).

Table  shows that FragmGAN with the γcv selected by cross-validation outperforms all the other methods in all the three cases, indicating that cross-validation is a good way to choose γ. FragmGAN with γ=0.5 always performs the second best. Note that γ controls the relative importance of data imputation to label prediction. As γ decreases from 1 to 0.25, the prediction performances first increase and then decrease. This result confirms two points that we have made: first, the linkage mechanism of training generator and predictor together can improve the prediction performance; second, a small γ close to 0 will lead to overfitting and damage the label prediction performance at the test data. Base on the results, we believe γ=0.5 is a reasonable choice if the users are not willing to apply cross validation to choose the best γ due to the computational burden.

6. Concluding remarks

Fragmentary data is becoming more and more popular in many areas and it is not easy to handle. By leveraging the structure in the response patterns, we propose a unified and flexible GAN based framework to deal with data imputation and label prediction simultaneously. An adjusting factor γ is used to adjust the relative importance of imputation to prediction. Theoretical guarantees for imputation are provided under the MAR assumption. Extensive experiments confirm the superiority of our proposed FragmGAN. It has wide application prospects especially in personal internet credit investigation, individual patient data (IPD) meta analysis in medical research and so on.

Based on the theoretical explorations and results of the experiments, we provide several practical suggestions for implementing FragmGAN in the practice.

  • Always use the hint mechanism although it is not required for the theoretical results.

  • Use γ=1 if there is no label in the analysis or you are only interested in data imputation.

  • If you are interested in label prediction, use cross-validation to select the best γ or simply use γ=0.5.

Possible future work includes the following. (i) We find that the best performances of data imputation (requires γ=1) and label prediction (requires a γ between 0 and 1) are not achieved at the same time. This is an interesting phenomenon and further investigation may lead to some thought-provoking results. (ii) In this paper we assume that the label is always available in the training data. We may explore FragmGAN under a semi-supervised setting in which some labels are not available. A naive method is to consider the label as part of the data vector and apply GAIN or FragmGAN directly. But this method is trying to recover the distribution of X and the conditional distribution of Y|X, which may not be desirable as we just mentioned in the first point. (iii) In this paper we assume that data are MAR. The extension of the results to the more general case of MNAR is a difficult but interesting task.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

The authors gratefully acknowledge the research support from the National Key R&D Program of China [Grant Numbers 2021YFA1000100 and 2021YFA1000101], National Natural Science Foundation of China [Grant Numbers 72331005, 12071143 and 11831008] and the Basic Research Project of Shanghai Science and Technology Commission [Grant Number 22JC1400800].

References

  • Awan, S. E., Bennamoun, M., Sohel, F., Sanfilippo, F., & Dwivedi, G. (2021). Imputation of missing data with class imbalance using conditional generative adversarial networks. Neurocomputing, 453(17), 164–171. https://doi.org/10.1016/j.neucom.2021.04.010
  • Camino, R. D., Hammerschmidt, C. A., & State, R. (2019). Improving missing data imputation with deep generative models. arXiv:1902.10666v1.
  • Dalca, A. V., Guttag, J., & Sabuncu, M. R. (2019). Unsupervised data imputation via variational inference of deep subspaces. arXiv:1903.03503v1.
  • Deng, G., Han, C., & Matteson, D. S. (2020). Learning to rank with missing data via generative adversarial networks. arXiv:2011.02089v2.
  • Fan, J., & Lv, J. (2008). Sure independence screening for ultrahigh dimensional feature space (with discussions). Journal of Royal Statistical Society Series B, 70(5), 849–911. https://doi.org/10.1111/j.1467-9868.2008.00674.x
  • Fang, F., Lan, W., Tong, J., & Shao, J. (2019). Model averaging for prediction with fragmentary data. Journal of Business & Economic Statistics, 37(3), 517–527. https://doi.org/10.1080/07350015.2017.1383263
  • Friedjungová, M., Vasata, D., Balatsko, M., & Jirina, M. (2020). Missing features reconstruction using a Wasserstein generative adversarial imputation network. In International Conference on Computational Science (ICCS 2020). pp. 225–239.
  • García-Laencina, P. J., Sancho-Gómez, J. -L., & Figueiras-Vidal, A. R. (2010). Pattern classification with missing data: a review. Neural Computing and Applications, 19(2), 263–282. https://doi.org/10.1007/s00521-009-0295-6
  • Ghalebikesabi, S., Cornish, R., Holmes, C., & Kelly, L. (2021). Deep generative missingness pattern-set mixture models. In Proceedings of The 24th International Conference on Artificial Intelligence and Statistics (AISTATS 2021). pp. 3727–3735.
  • Gondara, L., & Wang, K. (2017). Multiple imputation using deep denoising autoencoders. arXiv:1705.02737.
  • Gong, Y., Hajimirsadeghi, H., He, J., Durand, T., & Mori, G. (2021). Variational selective autoencoder: Learning from partially-observed heterogeneous data. In Proceedings of The 24th International Conference on Artificial Intelligence and Statistics (AISTATS 2021). pp. 2377–2385.
  • Hwang, U., Jung, D., & Yoon, S. (2019). HexaGAN: Generative adversarial nets for real world classification. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019). pp. 2921–2930.
  • Ipsen, N. B., Mattei, P. -A., & Frellsen, J. (2021). NOT-MIWAE: Deep generative modelling with missing not at random data. In International Conference on Learning Representations (ICLR 2021).
  • Ivanov, O., Figurnov, M., & Vetrov, D. (2019). Variational autoencoder with arbitrary conditioning. In International Conference on Learning Representations (ICLR 2019).
  • Lee, D., Kim, J., Moon, W.-J., & Ye, J. C. (2019). CollaGAN: Collaborative gan for missing image data imputation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019). pp. 2487–2496.
  • Li, Q., & Li, L. (2021). Integrative factor regression and its inference for multimodal data analysis. Journal of the American Statistical Association, https://doi.org/10.1080/01621459.2021.1914635.
  • Li, S. C. -X., Jiang, B., & Marlin, B. (2019). MisGAN: Learning from incomplete data with generative adversarial networks. In International Conference on Learning Representations (ICLR 2019).
  • Lichman, M. (2013). UCI machine learning repository. http://archive.ics.uci.edu/ml.
  • Lin, H., Liu, W., & Lan, W. (2021). Regression analysis with individual-specific patterns of missing covariates. Journal of Business & Economic Statistics, 39(1), 179–188. https://doi.org/10.1080/07350015.2019.1635486
  • Little, R. J., & Rubin, D. B. (2014). Statistical analysis with missing data. (2nd ed.).John Wiley & Sons.
  • Ma, W., & Chen, H. G. (2019). Missing not at random in matrix completion: The effectiveness of estimating missingness probabilities under a low nuclear norm assumption. In 33rd Conference on Neural Information Processing Systems (NeurIPS 2019).
  • Mattei, P.-A., & Frellsen, J. (2019). MIWAE: Deep generative modelling and imputation of incomplete data sets. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019). pp. 4413–4423.
  • Mazumder, R., Hastie, T., & Tibshirani, R. (2010). Spectral regularization algorithms for learning large incomplete matrices. Journal of Machine Learning Research, 11(Aug), 2287–2322.
  • Neves, D. T., Naik, M. G., & Proenca, A. (2021). SGAIN, WSGAIN-CP and WSGAIN-GP: Novel GAN methods for missing data imputation. In International Conference on Computational Science (ICCS 2021). pp. 98–113.
  • Qiu, W., Huang, Y., & Li, Q. (2020). IFGAN: Missing value imputation using feature-specific generative adversarial networks. In IEEE International Conference on Big Data (Big Data2020). pp. 4715–4723.
  • Richardson, T. W., Wu, W., Lin, L., Xu, B., & Bernal, E. A. (2020). MCFlow: Monte carlo flow models for data imputation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020). pp. 14205–14214.
  • Rubin, D. B. (2004). Multiple imputation for nonresponse in surveys. John Wiley & Sons.
  • Smieja, M., Kolomycki, M., Struski, L., Juda, M., & Figueiredo, M. A. T. (2020). Iterative imputation of missing data using auto-encoder dynamics. In International Conference on Neural Information Processing (ICONIP 2020).
  • Stekhoven, D. J., & Buhlmann, P. (2011). MissForest — non-parametric missing value imputation for mixed-type data. Bioinformatics, 28(1), 112–118. https://doi.org/10.1093/bioinformatics/btr597
  • van Buuren, S., & Groothuis-Oudshoorn, K. (2011). MICE: multivariate imputation by chained equations in R. Journal of Statistical Software, 45(3), 1–67.
  • Wang, Y., Li, D., Li, X., & Yang, M. (2021). PC-GAIN: pseudo-label conditional generative adversarial imputation networks for incomplete data. Neural Networks, 141(Sep), 395–403. https://doi.org/10.1016/j.neunet.2021.05.033
  • Xue, F., & Qu, A. (2021). Integrating multi-source block-wise missing data in model selection. Journal of the American Statistical Association, 116(536), 1914–1927. https://doi.org/10.1080/01621459.2020.1751176
  • Yoon, J., Jordon, J., & van der Schaar, M. (2018). GAIN: Missing data imputation using generative adversarial nets. In Proceedings of the 35th International Conference on Machine Learning (ICML 2018). pp. 5689–5698.
  • Yoon, S., & Sull, S. (2020). GAMIN: Generative adversarial multiple imputation network for highly missing data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020). pp. 8456–8464.
  • You, J., Ma, X., Ding, D., Kochenderfer, M., & Leskovec, J. (2020). Handling missing data with graph representation learning. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020).
  • Zhang, Y., Tang, N., & Annie, Q. (2020). Imputed factor regression for high-dimensional blockwise missing data. Statistica Sinica, 30(2), 631–651.

Appendices

Appendix 1.

A.1. Proofs of the proposed FragmGAN

1.1. Proof of Lemma 3.1

Proof.

Let πk=P(W=wk0). (A1) V(G,D)=E(Xˆ,W)[k=1KWklogDk(Xˆ)]=k=1KP(W=wk0)xˆp(xˆ|W=wk0)logDk(xˆ)dxˆ=xˆ[k=1Kπkp(xˆ|W=wk0)logDk(xˆ)]dxˆ.(A1) Note that k=1KDk(xˆ)=1. By the fact that k=1Kcklogxk with k=1Kxk=1 achieving its maximum when xk=ckk=1Kck, (EquationA1) is maximized (for fixed G) when Dk(xˆ)=πkp(xˆ|W=wk0)k=1Kπkp(xˆ|W=wk0)=p(W=wk0|xˆ):=Dk(xˆ)for the xˆ such that p(xˆ)>0.

1.2. Proof of Theorem 3.2

Proof.

(A2) C(G)=V(D,G)=E(Xˆ,W)[k=1KWklogp(W=wk0|Xˆ)]=k=1Kπkxˆp(xˆ|W=wk0)logp(W=wk0|xˆ)dxˆ=k=1Kπkxˆp(xˆ|W=wk0)logp(xˆ|W=wk0)πkp(xˆ)dxˆ=k=1Kπkxˆp(xˆ|W=wk0)logp(xˆ|W=wk0)p(xˆ)dxˆ+k=1Kπkxˆp(xˆ|W=wk0)logπkdxˆk=1Kπkxˆp(xˆm|xo,W=wk0)p(xo|W=wk0)logp(xˆm|xo,W=wk0)p(xo|W=wk0)p(xˆm|xo)p(xo)dxˆ(A2) (A3) =k=1Kπkxop(xo|W=wk0)[xˆmp(xˆm|xo,W=wk0)logp(xˆm|xo,W=wk0)p(xˆm|xo)dxˆm]dxo+k=1Kπkxop(xo|W=wk0)[xˆmp(xˆm|xo,W=wk0)logp(xo|W=wk0)p(xo)dxˆm]dxo,(A3) where ‘∝’ means equation holds by ignoring terms unrelated to G, and (EquationA2) holds since xˆp(xˆ|W=wk0)dxˆ=1 is a constant and xˆ=(xo,xˆm). Note that logp(xo|W=wko)p(xo) is unrelated to xˆm and xˆmp(xˆm|xo,W=wk0)dxˆm=1, so the second term of (EquationA3) is unrelated to G. Following (EquationA3), we have C(G)k=1Kπkxop(xo|W=wk0)KL(p(xˆm|xo,W=wk0)||p(xˆm|xo))dxo,where KL(||) denotes the KL divergence. Its minimum is achieved when p(xˆm|xo,W=wk0)=p(xˆm|xo) for each k{1,,K} and (almost) every x such that p(xˆ)>0 and p(xo|W=wk0)>0.

1.3. Proof of Theorem 3.3

Proof.

Actually we have proved this theorem in the statements between Theorem 3.2 and Theorem 3.3 in Section 3.2.

A.2. Extend theoretical results of GAIN (Yoon et al., Citation2018) to missing at random

We first rewrite the formulation of GAIN under MAR with our notation (just a little bit different from the original GAIN paper).

The original data X=(Xo,Xm)Rd, dim(Xo)=do, dim(Xm)=dm and do+dm=d. Denote M{0,1}dm as the response indicator for Xm. We assume Xm is missing at random, i.e., p(M|X)=p(M|Xo). Let Z=(Z1,,Zdm) be a dm-dimensional noise vector. Denote X¯m=G(Xo,MXm,(1M)Z,M)Rdm,Xˆm=MXm+(1M)X¯mRdm,and Xˆ=(Xo,Xˆm)Rd is the complete data after imputation. Let HRdm be the hint vector. The discriminator D is a function from Rd×Rdm to [0,1]dm such that Mˆ=D(Xˆ,H)=(Mˆ1,,Mˆdm)is the predicted probability vector for M. The minimax problem is (A4) minGmaxDV(G,D)=minGmaxDE(Xˆ,M,H)[MlogD(Xˆ,H)+(1M)log(1D(Xˆ,H))],(A4) where log is element-wise logarithm and dependence on G is through Xˆ.

The proof of Lemma 1 in GAIN (Yoon et al., Citation2018) does not depend on the decomposition of X. So the result still holds: the optimal D for given G is given by Di(xˆ,h)=p(Mi=1|xˆ,h)for i{1,,dm}.

Denote Hti={h:p(h|mi=t)>0} for t{0,1} and i{1,,dm}. Substituting D=(D1,,Ddm) into V(G,D) in (EquationA4), we have the objective function for G (to minimize): (A5) C(G)=E(Xˆ,M,H)[i:Mi=1logp(mi=1|Xˆ,H)+i:Mi=0logp(mi=0|Xˆ,H)]=xˆhi=1dm[p(xˆ,h,mi=1)logp(mi=1|xˆ,h)+p(xˆ,h,mi=0)logp(mi=0|xˆ,h)]dhdxˆ=i=1dmt{0,1}Htixˆp(xˆ,h,mi=t)logp(mi=t|xˆ,h)dhdxˆ=i=1dmt{0,1}Htixˆp(xˆ,h,mi=t)logp(xˆ,mi=t|h)p(xˆ|h)dhdxˆ=i=1dmt{0,1}Htixˆp(xˆ,h,mi=t)logp(xˆ,mi=t|h)p(mi=t|h)p(xˆ|h)p(mi=t|h)dhdxˆ=i=1dmt{0,1}Htixˆp(xˆ,h,mi=t)logp(xˆ|h,mi=t)p(mi=t|h)p(xˆ|h)dhdxˆ=i=1dmt{0,1}Htixˆp(xˆ,h,mi=t)logp(xˆ|h,mi=t)p(xˆ|h)dhdxˆ+i=1dmt{0,1}Htixˆp(xˆ,h,mi=t)logp(mi=t|h)dhdxˆ.(A5) Note that logp(mi=t|h) and xˆp(xˆ,h,mi=t)dxˆ=p(h,mi=t) are not related to xˆ. So the second term of (EquationA5) is not related to G and hence C(G)i=1dmt{0,1}Htixˆp(xˆ,h,mi=t)logp(xˆ|h,mi=t)p(xˆ|h)dhdxˆ=i=1dmt{0,1}Htip(h,mi=t)[xˆp(xˆ|h,mi=t)logp(xˆ|h,mi=t)p(xˆ|h)dxˆ]dh=i=1dmt{0,1}Htip(h,mi=t)[xˆp(xˆm|xo,h,mi=t)p(xo|h,mi=t)logp(xˆm|xo,h,mi=t)p(xo|h,mi=t)p(xˆm|xo,h)p(xo|h)dxˆ]dh=i=1dmt{0,1}Htip(h,mi=t)xop(xo|h,mi=t)[xˆmp(xˆm|xo,h,mi=t)logp(xˆm|xo,h,mi=t)p(xˆm|xo,h)dxˆm]dxodh+i=1dmt{0,1}Htip(h,mi=t)xop(xo|h,mi=t)[xˆmp(xˆm|xo,h,mi=t)logp(xo|h,mi=t)p(xo|h)dxˆm]dxodhi=1dmt{0,1}Htip(h,mi=t)xop(xo|h,mi=t)[xˆmp(xˆm|xo,h,mi=t)logp(xˆm|xo,h,mi=t)p(xˆm|xo,h)dxˆm]dxodh=i=1dmt{0,1}Htip(h,Mi=t)xop(xo|h,mi=t)KL(p(xˆm|xo,h,mi=t)||p(xˆm|xo,h))dxodh,which achieves its minimum when (A6) p(xˆm|xo,h,mi=t)=p(xˆm|xo,h)(A6) for t{0,1} and i{1,,dm}.

Let B=(B1,,Bdm){0,1}dm be a random vector taking value bi0 with probability 1dm, where bi0=(1,,1,0,1,,1) is a dm-dimensional vector with only the ith element being 0, i=1,,dm. The hint vector H=BM+0.5(1B). Note that Hi=t means Mi=t for t{0,1} and Hi=0.5 implies nothing about Mi. With this H, Di(xˆ,h)=t for h such that hi=t and t{0,1}.

For any m=(m1,,mdm){0,1}dm and i{1,,dm}, let m0, m1{0,1}dm be any two vectors such that they are the same as m on the jth element for ji, and the ith components of m0 and m1 are 0 and 1, respectively. So m=m0 if mi=0 and m=m1 if mi=1. Define a realization of the hint vector H as h such that hj=mj if ji and hj=0.5 if j = i. Since p(h|mi=t)>0, by (EquationA6) we have (A7) p(xˆm|xo,h,mi=0)=p(xˆm|xo,h,mi=1).(A7) Note (A8) p(xˆm|xo,h,mi=t)=p(xˆm|xo,B=bi0,m=mt)=p(xˆm|xo,m=mt),(A8) where the first equation holds since {h,mi=t} is equivalent to {B=bi0,m=mt}, and the second equality holds due to the independence of B to the other variables. Combing (EquationA7) and (EquationA8), we have p(xˆm|xo,m0)=p(xˆm|xo,m1).

Let 1=(1,,1) and m be any vector in {0,1}dm. There exists a sequence of vectors m1,,mL such that ml and ml+1 only differ on one component and m1=m and mL=1. By the arguments above, we have p(xˆm|xo,m)=p(xˆm|xo,m1)==p(xˆm|xo,mL)=p(xˆm|xo,1).Note that p(xˆm|xo,1)=p(xm|xo,1). And by the MAR assumption we have p(xm|xo,1)=p(xm|xo). So p(xˆm|xo)=m{0,1}dmp(M=m)p(xˆm|xo,m)=m{0,1}dmp(M=m)p(xm|xo)=p(xm|xo),which implies p(Xˆ)=p(X) as we need.

A.3. Dataset details and hyper-parameters

The details of the five UCI datasets can be found in UCI machine learning repository (Lichman, Citation2013) and the paper of GAIN (Yoon et al., Citation2018).

For the datasets Internet Loan and ADNI, the number of variables at each data source and sample size of each response pattern are presented in below tables.

Table B1. The number of variables at each data source and sample size of each response pattern for the Internet Loan data.

Table B2. The number of variables at each data source and sample size of each response pattern for the ADNI data.

The original ADNI data is available at http://adni.loni.usc.edu. The number of variables from the last three sources in the original data is larger. We use feature screen methods (Fan & Lv, Citation2008) to screen out the 10 most important variables for each source for our experiment.

In all experiments, the depth of generator, discriminator and predictor in FragmGAN, GAIN and Auto-Encoder is set to be 3. The number of hidden nodes in each layer for generator and discriminator is 2d, d and d, respectively. The number of hidden nodes in each layer for predictor is d, d/2 and 1, respectively. The activation function is ReLu except for the output layer that uses sigmoid. The training batch sizes kG, kD and kP are all 64. The α in LM is 10. For the cross-validation of FragmGAN, we search the value of γ on the grid of {0.40,0.41,,0.59,0.60}.

We use PyTorch to implement FragmGAN, GAIN, Auto-Encoder and MisGAN. We use Python to implement

MICE (package ‘fancyimpute’, https://github.com/iskandr/fancyimpute),

MissForest (package ‘missingpy’, https://github.com/softmechanics/missingpy),

EM (package ‘impyute’, https://github.com/eltonlaw/impyute), and Matrix (https://www.cnblogs.com/wuliytTaotao/p/10814770.html).

We use R to implement Model Averaging and FR-FI.