no code implementations • 5 Apr 2019 • Soeren Laue
We show that reverse mode automatic differentiation and symbolic differentiation are equivalent in the sense that they both perform the same operations when computing derivatives.
no code implementations • NeurIPS 2018 • Soeren Laue, Matthias Mitterreiter, Joachim Giesen
Optimization is an integral part of most machine learning systems and most numerical optimization schemes rely on the computation of derivatives.
no code implementations • NeurIPS 2012 • Joachim Giesen, Jens Mueller, Soeren Laue, Sascha Swiercy
We consider an abstract class of optimization problems that are parameterized concavely in a single parameter, and show that the solution path along the parameter can always be approximated with accuracy $\varepsilon >0$ by a set of size $O(1/\sqrt{\varepsilon})$.