Online Inference with Debiased Stochastic Gradient Descent

Han Ruijian, Luo Lan, Lin Yuanyuan, Huang Jian

Research output: Journal article publicationJournal articleAcademic researchpeer-review

4 Citations (Scopus)

Abstract

We propose a debiased stochastic gradient descent algorithm for online statistical inference with high-dimensional data. Our approach combines the debiasing technique developed in high-dimensional statistics with the stochastic gradient descent algorithm. It can be used to construct confidence intervals efficiently in an online fashion. Our proposed algorithm has several appealing aspects: as a one-pass algorithm, it reduces the time complexity; in addition, each update step requires only the current data together with the previous estimate, which reduces the space complexity. We establish the asymptotic normality of the proposed estimator under mild conditions on the sparsity level of the parameter and the data distribution. Numerical experiments demonstrate that the proposed debiased stochastic gradient descent algorithm attains nominal coverage probability. Furthermore, we illustrate our method with analysis of a high-dimensional text dataset.

Original languageEnglish
Pages (from-to)93-108
Number of pages16
JournalBiometrika
Volume111
Issue number1
DOIs
Publication statusPublished - 1 Mar 2024

Keywords

  • High-dimensional statistics
  • Online learning
  • Some key words: Confidence interval
  • Stochastic gradient descent

ASJC Scopus subject areas

  • Statistics and Probability
  • General Mathematics
  • Agricultural and Biological Sciences (miscellaneous)
  • General Agricultural and Biological Sciences
  • Statistics, Probability and Uncertainty
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'Online Inference with Debiased Stochastic Gradient Descent'. Together they form a unique fingerprint.

Cite this