no code implementations • 30 Jul 2024 • Sara Abdali, Jia He, CJ Barberan, Richard Anarfi
The advent of Large Language Models (LLMs) has garnered significant popularity and wielded immense power across various domains within Natural Language Processing (NLP).
no code implementations • 19 Mar 2024 • Sara Abdali, Richard Anarfi, CJ Barberan, Jia He
Large language models (LLMs) have significantly transformed the landscape of Natural Language Processing (NLP).
no code implementations • 9 Mar 2024 • Sara Abdali, Richard Anarfi, CJ Barberan, Jia He
Large Language Models (LLMs) have revolutionized the field of Natural Language Generation (NLG) by demonstrating an impressive ability to generate human-like text.
no code implementations • 23 Feb 2022 • CJ Barberan, Sina AlEMohammad, Naiming Liu, Randall Balestriero, Richard G. Baraniuk
A key interpretability issue with RNNs is that it is not clear how each hidden state per time step contributes to the decision-making process in a quantitative manner.
no code implementations • 15 Oct 2021 • CJ Barberan, Randall Balestriero, Richard G. Baraniuk
Each member of the family is derived from a standard DN architecture by vector quantizing the unit output values and feeding them into a global linear classifier.
1 code implementation • 11 Oct 2021 • Sina AlEMohammad, Hossein Babaei, CJ Barberan, Naiming Liu, Lorenzo Luzi, Blake Mason, Richard G. Baraniuk
To further contribute interpretability with respect to classification and the layers, we develop a new network as a combination of multiple neural tangent kernels, one to model each layer of the deep neural network individually as opposed to past work which attempts to represent the entire network via a single neural tangent kernel.