Authors
Xing Han, Jing Hu, Joydeep Ghosh
Publication date
2023/6/18
Conference
2023 International Joint Conference on Neural Networks (IJCNN)
Pages
1-8
Publisher
IEEE
Description
We propose a new, principled approach to tackling missing data problems that can reduce both bias and variance of any (stochastic) gradient descent-based predictive model that is learned on such data. The proposed method can use an arbitrary (and potentially biased) imputation model to fill in the missing values, as it corrects the biases introduced by imputation with a control variates method, leading to an unbiased estimation for gradient updates. Theoretically, we prove that our control variates approach improves the convergence of stochastic gradient descent under common missing data settings. Empirically, we show that our method yields superior performance as compared to the results obtained using competing imputation methods, on various applications, across different missing data patterns.
Scholar articles
X Han, J Hu, J Ghosh - 2023 International Joint Conference on Neural …, 2023