1 code implementation • 28 Dec 2024 • Alfredo Fernandez, Ankur Mali
We propose the Hyperbolic Tangent Exponential Linear Unit (TeLU), a neural network hidden activation function defined as TeLU(x)=xtanh(exp(x)).
no code implementations • 7 Oct 2024 • Ankur Mali, Tommaso Salvatori, Alexander Ororbia
Energy-based learning algorithms, such as predictive coding (PC), have garnered significant attention in the machine learning community due to their theoretical properties, such as local operations and biologically plausible mechanisms for error correction.
no code implementations • 4 Oct 2024 • Neisarg Dave, Daniel Kifer, Lee Giles, Ankur Mali
However, our research challenges this notion by demonstrating that RNNs primarily operate as state machines, where their linguistic capabilities are heavily influenced by the precision of their embeddings and the strategies used for sampling negative examples.
no code implementations • 4 Oct 2024 • Shrabon Das, Ankur Mali
This study explores the learnability of memory-less and memory-augmented RNNs, which are theoretically equivalent to Pushdown Automata.
no code implementations • 21 Aug 2024 • Romit Chatterjee, Vikram Chundawat, Ayush Tarun, Ankur Mali, Murari Mandal
Continual learning and machine unlearning are crucial challenges in machine learning, typically addressed separately.
no code implementations • 21 May 2024 • Neisarg Dave, Daniel Kifer, C. Lee Giles, Ankur Mali
However, most research has predominantly focused on language-based reasoning and word problems, often overlooking the potential of LLMs in handling symbol-based calculations and reasoning.
no code implementations • 19 Feb 2024 • Hitesh Vaidya, Travis Desell, Ankur Mali, Alexander Ororbia
The major challenge that makes crafting such a system difficult is known as catastrophic forgetting - an agent, such as one based on artificial neural networks (ANNs), struggles to retain previously acquired knowledge when learning from new samples.
no code implementations • 16 Feb 2024 • Alexander Ororbia, Ankur Mali, Adam Kohan, Beren Millidge, Tommaso Salvatori
As a result, it accommodates hardware and scientific modeling, e. g. learning with physical systems and non-differentiable behavior.
no code implementations • 5 Feb 2024 • Alfredo Fernandez, Ankur Mali
In this paper, we introduce the Hyperbolic Tangent Exponential Linear Unit (TeLU), a novel neural network activation function, represented as $f(x) = x{\cdot}tanh(e^x)$.
no code implementations • 4 Feb 2024 • Neisarg Dave, Daniel Kifer, C. Lee Giles, Ankur Mali
We sampled the datasets from $7$ Tomita and $4$ Dyck grammars and trained them on $4$ RNN cells: LSTM, GRU, O2RNN, and MIRNN.
no code implementations • 26 Sep 2023 • Ankur Mali, Alexander Ororbia, Daniel Kifer, Lee Giles
In this work, we extend the theoretical foundation for the $2^{nd}$-order recurrent network ($2^{nd}$ RNN) and prove there exists a class of a $2^{nd}$ RNN that is Turing-complete with bounded time.
no code implementations • 15 Aug 2023 • Tommaso Salvatori, Ankur Mali, Christopher L. Buckley, Thomas Lukasiewicz, Rajesh P. N. Rao, Karl Friston, Alexander Ororbia
Artificial intelligence (AI) is rapidly becoming one of the key technologies of this century.
1 code implementation • 4 Jan 2023 • Alexander Ororbia, Ankur Mali
We propose the predictive forward-forward (PFF) algorithm for conducting credit assignment in neural systems.
no code implementations • 22 Nov 2022 • Alexander Ororbia, Ankur Mali
In this work, we develop convolutional neural generative coding (Conv-NGC), a generalization of predictive coding to the case of convolution/deconvolution-based computation.
no code implementations • 11 Oct 2022 • Khai-Nguyen Nguyen, Zixin Tang, Ankur Mali, Alex Kelly
Unlike most neural language models, humans learn language in a rich, multi-sensory and, often, multi-lingual environment.
no code implementations • 19 Sep 2022 • Alexander Ororbia, Ankur Mali
In this article, we propose a backpropagation-free approach to robotic control through the neuro-cognitive computational framework of neural generative coding (NGC), designing an agent built completely from powerful predictive coding/processing circuits that facilitate dynamic, online learning from sparse rewards, embodying the principles of planning-as-inference.
1 code implementation • 3 Jun 2022 • Timothy Zee, Alexander G. Ororbia, Ankur Mali, Ifeoma Nwogu
While current deep learning algorithms have been successful for a wide variety of artificial intelligence (AI) tasks, including those involving structured image data, they present deep neurophysiological conceptual issues due to their reliance on the gradients that are computed by backpropagation of errors (backprop).
no code implementations • 27 Jan 2022 • Ankur Mali, Alexander Ororbia, Daniel Kifer, Lee Giles
In light of this, we propose a system that learns to improve the encoding performance by enhancing its internal neural representations on both the encoder and decoder ends, an approach we call Neural JPEG.
no code implementations • 27 Jan 2022 • Ankur Mali, Alexander Ororbia, Daniel Kifer, Lee Giles
Recent advances in deep learning have resulted in image compression algorithms that outperform JPEG and JPEG 2000 on the standard Kodak benchmark.
1 code implementation • 10 Jul 2021 • Alexander Ororbia, Ankur Mali
In humans, perceptual awareness facilitates the fast recognition and extraction of information from sensory input.
no code implementations • 19 Apr 2021 • Shivansh Rao, Vikas Kumar, Daniel Kifer, Lee Giles, Ankur Mali
A common approach has been to use standard convolutional networks to predict the corners and boundaries, followed by post-processing to generate the 3D layout.
no code implementations • 7 Apr 2021 • Ankur Mali, Alexander Ororbia, Daniel Kifer, C. Lee Giles
Two particular tasks that test this type of reasoning are (1) mathematical equation verification, which requires determining whether trigonometric and linear algebraic statements are valid identities or not, and (2) equation completion, which entails filling in a blank within an expression to make it true.
no code implementations • 5 Jun 2020 • John Stogin, Ankur Mali, C. Lee Giles
We introduce a neural stack architecture, including a differentiable parametrized stack operator that approximates stack push and pop operations for suitable choices of parameters that explicitly represents a stack.
no code implementations • 4 Apr 2020 • Ankur Mali, Alexander Ororbia, Daniel Kifer, Clyde Lee Giles
In this paper, we improve the memory-augmented RNN with important architectural and state updating mechanisms that ensure that the model learns to properly balance the use of its latent states with external memory.
no code implementations • 10 Feb 2020 • Alexander Ororbia, Ankur Mali, Daniel Kifer, C. Lee Giles
Training deep neural networks on large-scale datasets requires significant hardware resources whose costs (even on cloud platforms) put them out of reach of smaller organizations, groups, and individuals.
no code implementations • 20 Nov 2019 • Ankur Mali, Alexander G. Ororbia, Clyde Lee Giles
For lossy image compression, we develop a neural-based system which learns a nonlinear estimator for decoding from quantized representations.
no code implementations • 7 Sep 2019 • Ankur Mali, Alexander Ororbia, C. Lee Giles
The NSPDA is also compared to a classical analog stack neural network pushdown automaton (NNPDA) as well as a wide array of first and second-order RNNs with and without external memory, trained using different learning algorithms.
no code implementations • 25 May 2019 • Alexander Ororbia, Ankur Mali, Daniel Kifer, C. Lee Giles
In lifelong learning systems based on artificial neural networks, one of the biggest obstacles is the inability to retain old knowledge as new information is encountered.
2 code implementations • 17 Oct 2018 • Alexander Ororbia, Ankur Mali, C. Lee Giles, Daniel Kifer
We compare our model and learning procedure to other back-propagation through time alternatives (which also tend to be computationally expensive), including real-time recurrent learning, echo state networks, and unbiased online recurrent optimization.
1 code implementation • CVPR 2019 • Anand Gopalakrishnan, Ankur Mali, Dan Kifer, C. Lee Giles, Alexander G. Ororbia
We propose novel neural temporal models for predicting and synthesizing human motion, achieving state-of-the-art in modeling long-term motion trajectories while being competitive with prior work in short-term prediction and requiring significantly less computation.
no code implementations • ACL 2019 • Alexander G. Ororbia, Ankur Mali, Matthew A. Kelly, David Reitter
We examine the benefits of visual context in training neural language models to perform next-word prediction.
1 code implementation • 26 May 2018 • Alexander G. Ororbia, Ankur Mali
Finding biologically plausible alternatives to back-propagation of errors is a fundamentally important challenge in artificial neural network research.
no code implementations • 15 Mar 2018 • Alexander G. Ororbia, Ankur Mali, Jian Wu, Scott O'Connell, David Miller, C. Lee Giles
For lossy image compression systems, we develop an algorithm, iterative refinement, to improve the decoder's reconstruction compared to standard decoding techniques.
no code implementations • 5 Mar 2018 • Alexander G. Ororbia, Ankur Mali, Daniel Kifer, C. Lee Giles
Using back-propagation and its variants to train deep networks is often problematic for new users.