1 code implementation • 15 Feb 2023 • Sehoon Kim, Karttikeya Mangalam, Suhong Moon, John Canny, Jitendra Malik, Michael W. Mahoney, Amir Gholami, Kurt Keutzer
To address this, we propose Big Little Decoder (BiLD), a framework that can improve inference efficiency and latency for a wide range of text generation applications.
1 code implementation • 2 Feb 2023 • David M. Chan, Austin Myers, Sudheendra Vijayanarasimhan, David A. Ross, John Canny
If you ask a human to describe an image, they might do so in a thousand different ways.
no code implementations • 15 Sep 2022 • David M Chan, Yiming Ni, David A Ross, Sudheendra Vijayanarasimhan, Austin Myers, John Canny
In this work we argue that existing metrics are not appropriate for domains such as visual description or summarization where ground truths are semantically diverse, and where the diversity in those captions captures useful additional information about the context.
no code implementations • 16 Jul 2022 • Sumanth Gurram, Andy Fang, David Chan, John Canny
Generating representations of video data is of key importance in advancing the field of machine perception.
no code implementations • 7 Jul 2022 • Suhong Moon, Domas Buracas, Seunghyun Park, Jinkyu Kim, John Canny
It also uses a purely-dynamic local dispersive force (Brownian motion) that shows improved performance over other methods and does not require knowledge of other particle coordinates.
1 code implementation • 16 Jun 2022 • Eliza Kosoy, David M. Chan, Adrian Liu, Jasmine Collins, Bryanna Kaufmann, Sandy Han Huang, Jessica B. Hamrick, John Canny, Nan Rosemary Ke, Alison Gopnik
Recent work in machine learning and cognitive science has suggested that understanding causal information is essential to the development of intelligence.
no code implementations • 21 Feb 2022 • Eliza Kosoy, Adrian Liu, Jasmine Collins, David M Chan, Jessica B Hamrick, Nan Rosemary Ke, Sandy H Huang, Bryanna Kaufmann, John Canny, Alison Gopnik
Despite recent progress in reinforcement learning (RL), RL algorithms for exploration still remain an active area of research.
no code implementations • 15 Feb 2022 • Kehan Wang, David Chan, Seth Z. Zhao, John Canny, Avideh Zakhor
With the growing adoption of short-form video by social media platforms, reducing the spread of misinformation through video posts has become a critical challenge for social media providers.
no code implementations • 15 Feb 2022 • Philippe Laban, Elicia Ye, Srujay Korlakunta, John Canny, Marti A. Hearst
News podcasts are a popular medium to stay informed and dive deep into news topics.
no code implementations • 19 Nov 2021 • Forrest Huang, Eldon Schoop, David Ha, Jeffrey Nichols, John Canny
Sketching is a natural and effective visual communication medium commonly used in creative processes.
1 code implementation • NeurIPS 2021 • Kuang-Huei Lee, Anurag Arnab, Sergio Guadarrama, John Canny, Ian Fischer
We verify this by developing SimCLR and BYOL formulations compatible with the Conditional Entropy Bottleneck (CEB) objective, allowing us to both measure and control the amount of compression in the learned representation, and observe their impact on downstream tasks.
Ranked #31 on
Self-Supervised Image Classification
on ImageNet
1 code implementation • 15 Sep 2021 • Daniel Seita, Abhinav Gopal, Zhao Mandi, John Canny
Then, students learn by running either offline RL or by using teacher data in combination with a small amount of self-generated data.
no code implementations • ACL 2020 • Philippe Laban, John Canny, Marti A. Hearst
This work describes an automatic news chatbot that draws content from a diverse set of news articles and creates conversations with a user about the news.
1 code implementation • ACL 2020 • Philippe Laban, Andrew Hsi, John Canny, Marti A. Hearst
This work presents a new approach to unsupervised abstractive summarization based on maximizing a combination of coverage and fluency for a given length constraint.
Ranked #48 on
Abstractive Text Summarization
on CNN / Daily Mail
no code implementations • 27 Jul 2020 • David M. Chan, Sudheendra Vijayanarasimhan, David A. Ross, John Canny
Automatic video captioning aims to train models to generate text descriptions for all segments in a video, however, the most effective approaches require large amounts of manual annotation which is slow and expensive.
1 code implementation • NeurIPS 2020 • Kuang-Huei Lee, Ian Fischer, Anthony Liu, Yijie Guo, Honglak Lee, John Canny, Sergio Guadarrama
The Predictive Information is the mutual information between the past and the future, I(X_past; X_future).
no code implementations • 5 Jun 2020 • Bofan Xue, David Chan, John Canny
We present a new publicly available dataset with the goal of advancing multi-modality learning by offering vision and language data within the same context.
no code implementations • 12 May 2020 • Forrest Huang, Eldon Schoop, David Ha, John Canny
Iteratively refining and critiquing sketches are crucial steps to developing effective designs.
1 code implementation • 6 May 2020 • Eliza Kosoy, Jasmine Collins, David M. Chan, Sandy Huang, Deepak Pathak, Pulkit Agrawal, John Canny, Alison Gopnik, Jessica B. Hamrick
Research in developmental psychology consistently shows that children explore the world thoroughly and efficiently and that this exploration allows them to learn.
1 code implementation • ICLR 2020 • Stephanie C. Y. Chan, Samuel Fishman, John Canny, Anoop Korattikara, Sergio Guadarrama
To aid RL researchers and production users with the evaluation and improvement of reliability, we propose a set of metrics that quantitatively measure different aspects of reliability.
no code implementations • CVPR 2019 • Jinkyu Kim, Teruhisa Misu, Yi-Ting Chen, Ashish Tawari, John Canny
We show that taking advice improves the performance of the end-to-end network, while the network cues on a variety of visual features that are provided by advice.
2 code implementations • 26 Oct 2019 • Daniel Seita, David Chan, Roshan Rao, Chen Tang, Mandi Zhao, John Canny
Learning from demonstrations is a popular tool for accelerating and reducing the exploration requirements of reinforcement learning.
1 code implementation • 23 Sep 2019 • Daniel Seita, Aditya Ganapathi, Ryan Hoque, Minho Hwang, Edward Cen, Ajay Kumar Tanwani, Ashwin Balakrishna, Brijen Thananjeyan, Jeffrey Ichnowski, Nawid Jamali, Katsu Yamane, Soshi Iba, John Canny, Ken Goldberg
In 180 physical experiments with the da Vinci Research Kit (dVRK) surgical robot, RGBD policies trained in simulation attain coverage of 83% to 95% depending on difficulty tier, suggesting that effective fabric smoothing policies can be learned from an algorithmic supervisor and that depth sensing is a valuable addition to color alone.
6 code implementations • NeurIPS 2019 • Roshan Rao, Nicholas Bhattacharya, Neil Thomas, Yan Duan, Xi Chen, John Canny, Pieter Abbeel, Yun S. Song
Semi-supervised learning has emerged as an important paradigm in protein modeling due to the high cost of acquiring supervised protein labels, but the current literature is fragmented when it comes to datasets and standardized evaluation techniques.
no code implementations • 31 Mar 2019 • Xinlei Pan, Daniel Seita, Yang Gao, John Canny
In this paper we introduce risk-averse robust adversarial reinforcement learning (RARARL), using a risk-averse protagonist and a risk-seeking adversary.
1 code implementation • 24 Mar 2019 • Ye Xia, Jinkyu Kim, John Canny, Karl Zipser, David Whitney
Inspired by human vision, we propose a new periphery-fovea multi-resolution driving model that predicts vehicle speed from dash camera videos.
2 code implementations • 26 Sep 2018 • Daniel Seita, Nawid Jamali, Michael Laskey, Ajay Kumar Tanwani, Ron Berenstein, Prakash Baskaran, Soshi Iba, John Canny, Ken Goldberg
We compare coverage results from (1) human supervision, (2) a baseline of picking at the uppermost blanket point, and (3) learned pick points.
no code implementations • 26 Aug 2018 • Xinlei Pan, Sung-Li Chiang, John Canny
First, we note that an optimal subset (relative to all the objects encountered and labeled) of labeled objects in images can be obtained by importance sampling using gradients of the recognition network.
2 code implementations • ECCV 2018 • Jinkyu Kim, Anna Rohrbach, Trevor Darrell, John Canny, Zeynep Akata
Finally, we explore a version of our model that generates rationalizations, and compare with introspective explanations on the same video segments.
1 code implementation • 19 Sep 2017 • Daniel Seita, Sanjay Krishnan, Roy Fox, Stephen McKinley, John Canny, Ken Goldberg
In Phase II (fine), the bias from Phase I is applied to move the end-effector toward a small set of specific target points on a printed sheet.
Robotics
no code implementations • ICCV 2017 • Jinkyu Kim, John Canny
The attention model highlights image regions that potentially influence the network's output.
no code implementations • CVPR 2017 • Matthew Trager, Bernd Sturmfels, John Canny, Martial Hebert, Jean Ponce
The rational camera model recently introduced in [19] provides a general methodology for studying abstract nonlinear imaging systems and their multi-view geometry.
no code implementations • 19 Oct 2016 • Daniel Seita, Xinlei Pan, Haoyu Chen, John Canny
We present a novel Metropolis-Hastings method for large datasets that uses small expected-size minibatches of data.
no code implementations • 19 Nov 2015 • Daniel Seita, Haoyu Chen, John Canny
A fundamental task in machine learning and related fields is to perform inference on Bayesian networks.
2 code implementations • 18 Sep 2014 • Huasha Zhao, Biye Jiang, John Canny
SAME (State Augmentation for Marginal Estimation) \cite{Doucet99, Doucet02} is an approach to MAP parameter estimation which gives improved parameter estimates over direct Gibbs sampling.
no code implementations • 11 Dec 2013 • Huasha Zhao, John Canny
Many large datasets exhibit power-law statistics: The web graph, social networks, text data, click through data etc.
no code implementations • NeurIPS 2009 • Ye Chen, Michael Kapralov, John Canny, Dmitry Y. Pavlov
We adapt a probabilistic latent variable model, namely GaP (Gamma-Poisson), to ad targeting in the contexts of sponsored search (SS) and behaviorally targeted (BT) display advertising.