Abstract
We propose a debiased stochastic gradient descent algorithm for online statistical inference with high-dimensional data. Our approach combines the debiasing technique developed in high-dimensional statistics with the stochastic gradient descent algorithm. It can be used to construct confidence intervals efficiently in an online fashion. Our proposed algorithm has several appealing aspects: as a one-pass algorithm, it reduces the time complexity; in addition, each update step requires only the current data together with the previous estimate, which reduces the space complexity. We establish the asymptotic normality of the proposed estimator under mild conditions on the sparsity level of the parameter and the data distribution. Numerical experiments demonstrate that the proposed debiased stochastic gradient descent algorithm attains nominal coverage probability. Furthermore, we illustrate our method with analysis of a high-dimensional text dataset.
Original language | English |
---|---|
Pages (from-to) | 93-108 |
Number of pages | 16 |
Journal | Biometrika |
Volume | 111 |
Issue number | 1 |
DOIs | |
Publication status | Published - 1 Mar 2024 |
Keywords
- High-dimensional statistics
- Online learning
- Some key words: Confidence interval
- Stochastic gradient descent
ASJC Scopus subject areas
- Statistics and Probability
- General Mathematics
- Agricultural and Biological Sciences (miscellaneous)
- General Agricultural and Biological Sciences
- Statistics, Probability and Uncertainty
- Applied Mathematics