no code implementations • 14 Jul 2023 • Ryan Pyle, Sebastian Musslick, Jonathan D. Cohen, Ankit B. Patel
A key property of neural networks (both biological and artificial) is how they learn to represent and manipulate input information in order to solve a task.
no code implementations • 4 Aug 2020 • Justin Sahs, Ryan Pyle, Aneel Damaraju, Josue Ortega Caro, Onur Tavaslioglu, Andy Lu, Ankit Patel
Our implicit regularization results are complementary to recent work arXiv:1906. 07842, done independently, which showed that initialization scale critically controls implicit regularization via a kernel-based argument.
no code implementations • 19 Jun 2020 • Josue Ortega Caro, Yilong Ju, Ryan Pyle, Sourav Dey, Wieland Brendel, Fabio Anselmi, Ankit Patel
Inspired by theoretical work on linear full-width convolutional models, we hypothesize that the local (i. e. bounded-width) convolutional operations commonly used in current neural networks are implicitly biased to learn high frequency features, and that this is one of the root causes of high frequency adversarial examples.
no code implementations • 25 Sep 2019 • Justin Sahs, Aneel Damaraju, Ryan Pyle, Onur Tavaslioglu, Josue Ortega Caro, Hao Yang Lu, Ankit Patel
Despite their popularity and successes, deep neural networks are poorly understood theoretically and treated as 'black box' systems.
no code implementations • 8 Mar 2018 • Ryan Pyle, Robert Rosenbaum
Many recent studies of the motor system are divided into two distinct approaches: Those that investigate how motor responses are encoded in cortical neurons' firing rate dynamics and those that study the learning rules by which mammals and songbirds develop reliable motor responses.