1 code implementation • 16 Apr 2024 • Kyle Hsu, Jubayer Ibn Hamid, Kaylee Burns, Chelsea Finn, Jiajun Wu
Inductive biases are crucial in disentangled representation learning for narrowing down an underspecified solution set.
no code implementations • 9 Apr 2024 • Kaylee Burns, Ajinkya Jain, Keegan Go, Fei Xia, Michael Stark, Stefan Schaal, Karol Hausman
Large Language Models (LLMs) have been successful at generating robot policy code, but so far these results have been limited to high-level tasks that do not require precise movement.
no code implementations • 3 Nov 2023 • Kaylee Burns, Zach Witzel, Jubayer Ibn Hamid, Tianhe Yu, Chelsea Finn, Karol Hausman
Inspired by the success of transfer learning in computer vision, roboticists have investigated visual pre-training as a means to improve the learning efficiency and generalization ability of policies learned from pixels.
no code implementations • 26 Jul 2022 • Kaylee Burns, Tianhe Yu, Chelsea Finn, Karol Hausman
In this paper, we focus on one particular aspect of heterogeneity: learning from offline data collected at different control frequencies.
no code implementations • 20 Jul 2021 • Kaylee Burns, Christopher D. Manning, Li Fei-Fei
Although virtual agents are increasingly situated in environments where natural language is the most effective mode of interaction with humans, these exchanges are rarely used as an opportunity for learning.
no code implementations • 3 Dec 2018 • Eric Tzeng, Kaylee Burns, Kate Saenko, Trevor Darrell
Without dense labels, as is the case when only detection labels are available in the source, transformations are learned using CycleGAN alignment.
no code implementations • WS 2018 • Kaylee Burns, Aida Nematzadeh, Erin Grant, Alison Gopnik, Tom Griffiths
The decision making processes of deep networks are difficult to understand and while their accuracy often improves with increased architectural complexity, so too does their opacity.
1 code implementation • EMNLP 2018 • Anna Rohrbach, Lisa Anne Hendricks, Kaylee Burns, Trevor Darrell, Kate Saenko
Despite continuously improving performance, contemporary image captioning models are prone to "hallucinating" objects that are not actually in a scene.
2 code implementations • EMNLP 2018 • Aida Nematzadeh, Kaylee Burns, Erin Grant, Alison Gopnik, Thomas L. Griffiths
We propose a new dataset for evaluating question answering models with respect to their capacity to reason about beliefs.
no code implementations • 2 Jul 2018 • Lisa Anne Hendricks, Kaylee Burns, Kate Saenko, Trevor Darrell, Anna Rohrbach
Most machine learning methods are known to capture and exploit biases of the training data.
2 code implementations • ECCV 2018 • Kaylee Burns, Lisa Anne Hendricks, Kate Saenko, Trevor Darrell, Anna Rohrbach
We introduce a new Equalizer model that ensures equal gender probability when gender evidence is occluded in a scene and confident predictions when gender evidence is present.