1 code implementation • 16 Jun 2023 • Jifan Zhang, Yifang Chen, Gregory Canal, Stephen Mussmann, Arnav M. Das, Gantavya Bhatt, Yinglun Zhu, Jeffrey Bilmes, Simon Shaolei Du, Kevin Jamieson, Robert D Nowak
Labeled data are critical to modern machine learning applications, but obtaining labels can be expensive.
no code implementations • 15 Jun 2023 • Haoyue Bai, Gregory Canal, Xuefeng Du, Jeongyeol Kwon, Robert Nowak, Yixuan Li
Modern machine learning models deployed in the wild can encounter both covariate and semantic shifts, giving rise to the problems of out-of-distribution (OOD) generalization and OOD detection respectively.
2 code implementations • 7 Jul 2022 • Gregory Canal, Blake Mason, Ramya Korlakai Vinayak, Robert Nowak
This paper investigates simultaneous preference and metric learning from a crowd of respondents.
no code implementations • 27 May 2022 • Gregory Canal, Yancy Diaz-Mercado, Magnus Egerstedt, Christopher Rozell
We construct a scalable dictionary of robotic behaviors that can be searched simply and efficiently by a BCI user, as we demonstrate through a large-scale user study testing the feasibility of our interaction algorithm, a user test of the full BCI system on (virtual and real) robot swarms, and simulations that verify our results against theoretical models.
1 code implementation • 28 Feb 2021 • Gregory Canal, Matthieu Bloch, Christopher Rozell
The iterative selection of examples for labeling in active machine learning is conceptually similar to feedback channel coding in information theory: in both tasks, the objective is to seek a minimal sequence of actions to encode information in the presence of noise.
2 code implementations • NeurIPS 2020 • Matthew O'Shaughnessy, Gregory Canal, Marissa Connor, Mark Davenport, Christopher Rozell
Our objective function encourages both the generative model to faithfully represent the data distribution and the latent factors to have a large causal influence on the classifier output.
1 code implementation • 9 Oct 2019 • Gregory Canal, Stefano Fenu, Christopher Rozell
Many machine learning tasks such as clustering, classification, and dataset search benefit from embedding data points in a space where distances reflect notions of relative similarity as perceived by humans.