The continuous stochastic gradient method: part II–application and numerics

Grieshammer M, Pflug L, Stingl M, Uihlein A (2024)


Publication Type: Journal article

Publication year: 2024

Journal

Book Volume: 87

Pages Range: 977-1008

Journal Issue: 3

DOI: 10.1007/s10589-023-00540-w

Abstract

In this contribution, we present a numerical analysis of the continuous stochastic gradient (CSG) method, including applications from topology optimization and convergence rates. In contrast to standard stochastic gradient optimization schemes, CSG does not discard old gradient samples from previous iterations. Instead, design dependent integration weights are calculated to form a convex combination as an approximation to the true gradient at the current design. As the approximation error vanishes in the course of the iterations, CSG represents a hybrid approach, starting off like a purely stochastic method and behaving like a full gradient scheme in the limit. In this work, the efficiency of CSG is demonstrated for practically relevant applications from topology optimization. These settings are characterized by both, a large number of optimization variables and an objective function, whose evaluation requires the numerical computation of multiple integrals concatenated in a nonlinear fashion. Such problems could not be solved by any existing optimization method before. Lastly, with regards to convergence rates, first estimates are provided and confirmed with the help of numerical experiments.

Authors with CRIS profile

How to cite

APA:

Grieshammer, M., Pflug, L., Stingl, M., & Uihlein, A. (2024). The continuous stochastic gradient method: part II–application and numerics. Computational Optimization and Applications, 87(3), 977-1008. https://doi.org/10.1007/s10589-023-00540-w

MLA:

Grieshammer, Max, et al. "The continuous stochastic gradient method: part II–application and numerics." Computational Optimization and Applications 87.3 (2024): 977-1008.

BibTeX: Download