Deprecated: $wgMWOAuthSharedUserIDs=false is deprecated, set $wgMWOAuthSharedUserIDs=true, $wgMWOAuthSharedUserSource='local' instead [Called from MediaWiki\HookContainer\HookContainer::run in /var/www/html/w/includes/HookContainer/HookContainer.php at line 135] in /var/www/html/w/includes/Debug/MWDebug.php on line 372
Small errors in random zeroth-order optimization are imaginary - MaRDI portal

Small errors in random zeroth-order optimization are imaginary

From MaRDI portal
Publication:6507546

arXiv2103.05478MaRDI QIDQ6507546

Author name not available (Why is that?), Man-Chung Yue, Daniel Kuhn


Abstract: Most zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some noisy gradient estimator that can be computed from a small number of function evaluations. This estimator is constructed randomly, and its expectation matches the gradient of a smooth approximation of the objective function whose quality improves as the underlying smoothing parameter delta is reduced. Gradient estimators requiring a smaller number of function evaluations are preferable from a computational point of view. While estimators based on a single function evaluation can be obtained by a clever use of the divergence theorem from vector calculus, their variance explodes as delta tends to 0. Estimators based on multiple function evaluations, on the other hand, suffer from numerical cancellation when delta tends to 0. To combat both effects simultaneously, we extend the objective function to the complex domain and construct a gradient estimator that evaluates the objective at a complex point whose coordinates have small imaginary parts of the order delta. As this estimator requires only one function evaluation, it is immune to cancellation. In addition, its variance remains bounded as delta tends to 0. We prove that zeroth-order algorithms that use our estimator offer the same theoretical convergence guarantees as the state-of-the-art methods. Numerical experiments suggest, however, that they often converge faster in practice.












This page was built for publication: Small errors in random zeroth-order optimization are imaginary

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6507546)