no code implementations • 19 Mar 2024 • Sara Abdali, Richard Anarfi, CJ Barberan, Jia He
Large language models (LLMs) have significantly transformed the landscape of Natural Language Processing (NLP).
no code implementations • 9 Mar 2024 • Sara Abdali, Richard Anarfi, CJ Barberan, Jia He
Large Language Models (LLMs) have revolutionized the field of Natural Language Generation (NLG) by demonstrating an impressive ability to generate human-like text.
no code implementations • 23 Feb 2022 • CJ Barberan, Sina AlEMohammad, Naiming Liu, Randall Balestriero, Richard G. Baraniuk
A key interpretability issue with RNNs is that it is not clear how each hidden state per time step contributes to the decision-making process in a quantitative manner.
no code implementations • 15 Oct 2021 • CJ Barberan, Randall Balestriero, Richard G. Baraniuk
Each member of the family is derived from a standard DN architecture by vector quantizing the unit output values and feeding them into a global linear classifier.
1 code implementation • 11 Oct 2021 • Sina AlEMohammad, Hossein Babaei, CJ Barberan, Naiming Liu, Lorenzo Luzi, Blake Mason, Richard G. Baraniuk
To further contribute interpretability with respect to classification and the layers, we develop a new network as a combination of multiple neural tangent kernels, one to model each layer of the deep neural network individually as opposed to past work which attempts to represent the entire network via a single neural tangent kernel.