no code implementations • 16 Oct 2024 • Shashank Sonkar, Xinghe Chen, Naiming Liu, Richard G. Baraniuk, Mrinmaya Sachan
Our findings reveal that LLMs trained on misconception examples can efficiently learn to replicate errors.
1 code implementation • 1 Jul 2024 • Naiming Liu, Shashank Sonkar, MyCo Le, Richard Baraniuk
We propose the Malgorithm Identification task, where LLMs are assessed based on their ability to identify corresponding malgorithm given an incorrect answer choice.
no code implementations • 19 Jun 2024 • Naiming Liu, Zichao Wang, Richard Baraniuk
Despite rapid advancements in large language models (LLMs), QG remains a challenging problem due to its complicated process, open-ended nature, and the diverse settings in which question generation occurs.
no code implementations • 23 Apr 2024 • Shashank Sonkar, Naiming Liu, Richard G. Baraniuk
Our findings reveal significant declines in the models' performance across these diverse benchmarks, indicating a broad impact on their capabilities when trained to model student behavior.
no code implementations • 22 Apr 2024 • Shashank Sonkar, Naiming Liu, Debshila B. Mallick, Richard G. Baraniuk
We subsequently train language models to identify entailment, contradiction, and neutrality from student response, akin to NLI, and with the added dimension of identifying omissions from gold answers.
no code implementations • 3 Oct 2023 • Naiming Liu, Shashank Sonkar, Zichao Wang, Simon Woodhead, Richard G. Baraniuk
We propose novel evaluations for mathematical reasoning capabilities of Large Language Models (LLMs) based on mathematical misconceptions.
1 code implementation • 21 Sep 2023 • Shashank Sonkar, MyCo Le, Xinghe Chen, Naiming Liu, Debshila Basu Mallick, Richard G. Baraniuk
Our approach notably enhances the quality of synthetic conversation datasets, especially for subjects that are calculation-intensive.
1 code implementation • 22 May 2023 • Shashank Sonkar, Naiming Liu, Debshila Basu Mallick, Richard G. Baraniuk
We present a design framework called Conversational Learning with Analytical Step-by-Step Strategies (CLASS) for building advanced Intelligent Tutoring Systems (ITS) powered by high-performance Large Language Models (LLMs).
1 code implementation • 1 Nov 2022 • Lorenzo Luzi, Daniel LeJeune, Ali Siahkoohi, Sina AlEMohammad, Vishwanath Saragadam, Hossein Babaei, Naiming Liu, Zichao Wang, Richard G. Baraniuk
We study the interpolation capabilities of implicit neural representations (INRs) of images.
no code implementations • 22 Oct 2022 • Shashank Sonkar, Naiming Liu, Richard G. Baraniuk
Transformer models trained on massive text corpora have become the de facto models for a wide range of natural language processing tasks.
1 code implementation • 19 May 2022 • Nigel Fernandez, Aritra Ghosh, Naiming Liu, Zichao Wang, Benoît Choffin, Richard Baraniuk, Andrew Lan
Our approach, in-context BERT fine-tuning, produces a single shared scoring model for all items with a carefully-designed input structure to provide contextual information on each item.
no code implementations • 23 Feb 2022 • CJ Barberan, Sina AlEMohammad, Naiming Liu, Randall Balestriero, Richard G. Baraniuk
A key interpretability issue with RNNs is that it is not clear how each hidden state per time step contributes to the decision-making process in a quantitative manner.
1 code implementation • 21 Feb 2022 • Naiming Liu, Zichao Wang, Richard G. Baraniuk, Andrew Lan
In education applications, knowledge tracing refers to the problem of estimating students' time-varying concept/skill mastery level from their past responses to questions and predicting their future performance.
1 code implementation • 11 Oct 2021 • Sina AlEMohammad, Hossein Babaei, CJ Barberan, Naiming Liu, Lorenzo Luzi, Blake Mason, Richard G. Baraniuk
To further contribute interpretability with respect to classification and the layers, we develop a new network as a combination of multiple neural tangent kernels, one to model each layer of the deep neural network individually as opposed to past work which attempts to represent the entire network via a single neural tangent kernel.
1 code implementation • 27 Oct 2020 • Sina AlEMohammad, Hossein Babaei, Randall Balestriero, Matt Y. Cheung, Ahmed Imtiaz Humayun, Daniel LeJeune, Naiming Liu, Lorenzo Luzi, Jasper Tan, Zichao Wang, Richard G. Baraniuk
High dimensionality poses many challenges to the use of data, from visualization and interpretation, to prediction and storage for historical preservation.