Langevin dynamics based algorithm e-TH$\varepsilon$O POULA for stochastic optimization problems with discontinuous stochastic gradient
From MaRDI portal
Publication:6414917
arXiv2210.13193MaRDI QIDQ6414917
Author name not available (Why is that?)
Publication date: 24 October 2022
Abstract: We introduce a new Langevin dynamics based algorithm, called e-THO POULA, to solve optimization problems with discontinuous stochastic gradients which naturally appear in real-world applications such as quantile estimation, vector quantization, CVaR minimization, and regularized optimization problems involving ReLU neural networks. We demonstrate both theoretically and numerically the applicability of the e-THO POULA algorithm. More precisely, under the conditions that the stochastic gradient is locally Lipschitz in average and satisfies a certain convexity at infinity condition, we establish non-asymptotic error bounds for e-THO POULA in Wasserstein distances and provide a non-asymptotic estimate for the expected excess risk, which can be controlled to be arbitrarily small. Three key applications in finance and insurance are provided, namely, multi-period portfolio optimization, transfer learning in multi-period portfolio optimization, and insurance claim prediction, which involve neural networks with (Leaky)-ReLU activation functions. Numerical experiments conducted using real-world datasets illustrate the superior empirical performance of e-THO POULA compared to SGLD, ADAM, and AMSGrad in terms of model accuracy.
Has companion code repository: https://github.com/dongyounglim/etheopoula
This page was built for publication: Langevin dynamics based algorithm e-TH$\varepsilon$O POULA for stochastic optimization problems with discontinuous stochastic gradient
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6414917)