no code implementations • 13 Nov 2023 • Wouter Jongeneel, Mengmeng Li, Daniel Kuhn
Motivated by policy gradient methods in the context of reinforcement learning, we derive the first large deviation rate function for the iterates generated by stochastic gradient descent for possibly non-convex objectives satisfying a Polyak-Lojasiewicz condition.
no code implementations • 9 Mar 2021 • Wouter Jongeneel, Man-Chung Yue, Daniel Kuhn
Most zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some gradient estimator that can be computed from a small number of function evaluations.
Optimization and Control 65D25, 65G50, 65K05, 65Y04, 65Y20, 90C56
no code implementations • 5 Mar 2021 • Wouter Jongeneel, Tobias Sutter, Daniel Kuhn
Two dynamical systems are topologically equivalent when their phase-portraits can be morphed into each other by a homeomorphic coordinate transformation on the state space.
Optimization and Control