1 code implementation • 25 May 2023 • Benjamin Wright, Youngjae Min, Jeremy Bernstein, Navid Azizan
This paper proposes a memory-efficient solution to catastrophic forgetting, improving upon an established algorithm known as orthogonal gradient descent (OGD).
no code implementations • 6 Mar 2023 • Youngjae Min, Spencer M. Richards, Navid Azizan
Recent advances in learning-based control leverage deep function approximators, such as neural networks, to model the evolution of controlled dynamical systems over time.
no code implementations • 28 Jul 2022 • Youngjae Min, Kwangjun Ahn, Navid Azizan
While deep neural networks are capable of achieving state-of-the-art performance in various domains, their training typically requires iterating for many passes over the dataset.
no code implementations • 19 Apr 2019 • Youngjae Min, Hye Won Chung
This paper constructs shallow sigmoid-type neural networks that achieve 100% accuracy in classification for datasets following a linear separability condition.