|
SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES
Algorithms with gradient clipping for stochastic optimization with heavy-tailed noise
M. Yu. Danilova Moscow Institute of Physics and Technology, Moscow, Russia
Abstract:
This article provides a review of the results of several research studies, in which open questions related to the high-probability convergence analysis of stochastic first-order optimization methods under mild assumptions on the noise were gradually addressed. In the beginning, we introduce the concept of gradient clipping, which plays a pivotal role in the development of stochastic methods for successful operation in the case of heavy-tailed distributions. Next, we examine the importance of obtaining the highprobability convergence guarantees and their connection with in-expectation convergence guarantees. The concluding sections of the article are dedicated to presenting the primary findings related to minimization problems and the results of numerical experiments.
Keywords:
convex optimization, stochastic optimization, first-order methods.
Citation:
M. Yu. Danilova, “Algorithms with gradient clipping for stochastic optimization with heavy-tailed noise”, Dokl. RAN. Math. Inf. Proc. Upr., 514:2 (2023), 99–108; Dokl. Math., 108:suppl. 2 (2023), S248–S256
Linking options:
https://www.mathnet.ru/eng/danma455 https://www.mathnet.ru/eng/danma/v514/i2/p99
|
Statistics & downloads: |
Abstract page: | 50 | References: | 16 |
|