no code implementations • 12 Jan 2024 • Gantavya Bhatt, Yifang Chen, Arnav M. Das, Jifan Zhang, Sang T. Truong, Stephen Mussmann, Yinglun Zhu, Jeffrey Bilmes, Simon S. Du, Kevin Jamieson, Jordan T. Ash, Robert D. Nowak
To mitigate the annotation cost of SFT and circumvent the computational bottlenecks of active learning, we propose using experimental design.
1 code implementation • 21 Dec 2023 • Pratyusha Sharma, Jordan T. Ash, Dipendra Misra
Transformer-based Large Language Models (LLMs) have become a fixture in modern machine learning.
2 code implementations • 5 Mar 2023 • Akanksha Saran, Safoora Yousefi, Akshay Krishnamurthy, John Langford, Jordan T. Ash
Active learning is perhaps most naturally posed as an online learning problem.
1 code implementation • 2 Nov 2022 • Savya Khosla, Chew Kin Whye, Jordan T. Ash, Cyril Zhang, Kenji Kawaguchi, Alex Lamb
To this end, we demonstrate the catastrophic failure of these active learning algorithms on heteroskedastic distributions and propose a fine-tuning-based approach to mitigate these failures.
1 code implementation • 25 Oct 2022 • Mark Rucker, Jordan T. Ash, John Langford, Paul Mineiro, Ida Momennejad
This work introduces the Eigen Memory Tree (EMT), a novel online memory model for sequential learning scenarios.
no code implementations • 19 Oct 2022 • Bingbin Liu, Jordan T. Ash, Surbhi Goel, Akshay Krishnamurthy, Cyril Zhang
Algorithmic reasoning requires capabilities which are most naturally understood through recurrent models of computation, like the Turing machine.
no code implementations • ICLR 2022 • Jordan T. Ash, Cyril Zhang, Surbhi Goel, Akshay Krishnamurthy, Sham Kakade
Intrinsic rewards play a central role in handling the exploration-exploitation trade-off when designing sequential decision-making algorithms, in both foundational theory and state-of-the-art deep reinforcement learning.
no code implementations • 18 Jun 2021 • Jordan T. Ash, Surbhi Goel, Akshay Krishnamurthy, Dipendra Misra
We focus on disambiguating the role of one of these parameters: the number of negative examples.
1 code implementation • NeurIPS 2021 • Jordan T. Ash, Surbhi Goel, Akshay Krishnamurthy, Sham Kakade
There is an increasing need for effective active learning algorithms that are compatible with deep neural networks.
no code implementations • NeurIPS 2020 • Alex Beatson, Jordan T. Ash, Geoffrey Roeder, Tianju Xue, Ryan P. Adams
We use a neural network to model the stored potential energy in a component given boundary conditions.
1 code implementation • NeurIPS 2020 • Jordan T. Ash, Ryan P. Adams
We would like each of these models in the sequence to be performant and take advantage of all the data that are available to that point.
no code implementations • 25 Sep 2019 • Jordan T. Ash, Ryan P. Adams
We would like each of these models in the sequence to be performant and take advantage of all the data that are available to that point.
4 code implementations • ICLR 2020 • Jordan T. Ash, Chicheng Zhang, Akshay Krishnamurthy, John Langford, Alekh Agarwal
We design a new algorithm for batch active learning with deep neural network models.
no code implementations • 16 Feb 2016 • Jordan T. Ash, Robert E. Schapire, Barbara E. Engelhardt
Domain adaptation addresses the problem created when training data is generated by a so-called source distribution, but test data is generated by a significantly different target distribution.