no code implementations • 28 Nov 2023 • YiXuan Luo, Mengye Ren, Sai Qian Zhang
This approach significantly reduces computational costs in comparison with training each DNN backbone individually.
no code implementations • CVPR 2023 • Lunjun Zhang, Anqi Joyce Yang, Yuwen Xiong, Sergio Casas, Bin Yang, Mengye Ren, Raquel Urtasun
In this paper, we study the problem of unsupervised object detection from 3D point clouds in self-driving scenes.
no code implementations • 27 Jun 2023 • Chris Zhang, Runsheng Guo, Wenyuan Zeng, Yuwen Xiong, Binbin Dai, Rui Hu, Mengye Ren, Raquel Urtasun
Recent advances in high-fidelity simulators have enabled closed-loop training of autonomous driving agents, potentially solving the distribution shift in training v. s.
1 code implementation • 19 Oct 2022 • Renjie Liao, Simon Kornblith, Mengye Ren, David J. Fleet, Geoffrey Hinton
We revisit the challenging problem of training Gaussian-Bernoulli restricted Boltzmann machines (GRBMs), introducing two innovations.
1 code implementation • 7 Oct 2022 • Mengye Ren, Simon Kornblith, Renjie Liao, Geoffrey Hinton
Forward gradient learning computes a noisy directional gradient and is a biologically plausible alternative to backprop for learning deep neural networks.
no code implementations • 6 Oct 2022 • Andrew J. Nam, Mengye Ren, Chelsea Finn, James L. McClelland
Large language models have recently shown promising progress in mathematical reasoning when fine-tuned with human-generated sequences walking through a sequence of solution steps.
1 code implementation • 13 Sep 2021 • Mengye Ren, Tyler R. Scott, Michael L. Iuzzolino, Michael C. Mozer, Richard Zemel
Real world learning scenarios involve a nonstationary distribution of classes with sequential dependencies among the samples, in contrast to the standard machine learning formulation of drawing samples independently from a fixed, typically uniform distribution.
no code implementations • 8 Apr 2021 • Sean Segal, Nishanth Kumar, Sergio Casas, Wenyuan Zeng, Mengye Ren, Jingkang Wang, Raquel Urtasun
As data collection is often significantly cheaper than labeling in this domain, the decision of which subset of examples to label can have a profound impact on model performance.
no code implementations • 17 Jan 2021 • James Tu, Huichen Li, Xinchen Yan, Mengye Ren, Yun Chen, Ming Liang, Eilyan Bitar, Ersin Yumer, Raquel Urtasun
Yet, there have been limited studies on the adversarial robustness of multi-modal models that fuse LiDAR features with image features.
no code implementations • 17 Jan 2021 • Jingkang Wang, Mengye Ren, Ilija Bogunovic, Yuwen Xiong, Raquel Urtasun
Recent work on hyperparameters optimization (HPO) has shown the possibility of training certain hyperparameters together with regular parameters.
no code implementations • ICCV 2021 • James Tu, TsunHsuan Wang, Jingkang Wang, Sivabalan Manivasagam, Mengye Ren, Raquel Urtasun
Growing at a fast pace, modern autonomous systems will soon be deployed at scale, opening up the possibility for cooperative multi-agent systems.
no code implementations • ICCV 2021 • Yuwen Xiong, Mengye Ren, Wenyuan Zeng, Raquel Urtasun
Motivated by this ability, we present a new self-supervised learning representation framework that can be directly deployed on a video stream of complex scenes with many moving objects.
no code implementations • CVPR 2021 • Shuhan Tan, Kelvin Wong, Shenlong Wang, Sivabalan Manivasagam, Mengye Ren, Raquel Urtasun
Existing methods typically insert actors into the scene according to a set of hand-crafted heuristics and are limited in their ability to model the true complexity and diversity of real traffic scenes, thus inducing a content gap between synthesized traffic scenes versus real ones.
no code implementations • CVPR 2021 • Jingkang Wang, Ava Pun, James Tu, Sivabalan Manivasagam, Abbas Sadat, Sergio Casas, Mengye Ren, Raquel Urtasun
Importantly, by simulating directly from sensor data, we obtain adversarial scenarios that are safety-critical for the full autonomy stack.
no code implementations • 1 Jan 2021 • Mengye Ren, Eleni Triantafillou, Kuan-Chieh Wang, James Lucas, Jake Snell, Xaq Pitkow, Andreas S. Tolias, Richard Zemel
In this work, we consider a realistic setting where the relationship between examples can change from episode to episode depending on the task context, which is not given to the learner.
no code implementations • 10 Dec 2020 • Mengye Ren, Eleni Triantafillou, Kuan-Chieh Wang, James Lucas, Jake Snell, Xaq Pitkow, Andreas S. Tolias, Richard Zemel
Despite impressive progress in deep learning, generalizing far beyond the training distribution is an important open challenge.
no code implementations • 10 Nov 2020 • Nicholas Vadivelu, Mengye Ren, James Tu, Jingkang Wang, Raquel Urtasun
Learned communication makes multi-agent systems more effective by aggregating distributed information.
no code implementations • 2 Nov 2020 • Bob Wei, Mengye Ren, Wenyuan Zeng, Ming Liang, Bin Yang, Raquel Urtasun
In this paper, we propose an end-to-end self-driving network featuring a sparse attention module that learns to automatically attend to important regions of the input.
no code implementations • ICLR 2021 • James Lucas, Mengye Ren, Irene Kameni, Toniann Pitassi, Richard Zemel
Machine learning models have traditionally been developed under the assumption that the training and test distributions match exactly.
1 code implementation • 27 Aug 2020 • Alexander Wang, Mengye Ren, Richard S. Zemel
Sketch drawings capture the salient information of visual concepts.
no code implementations • 13 Aug 2020 • Lingyun Luke Li, Bin Yang, Ming Liang, Wenyuan Zeng, Mengye Ren, Sean Segal, Raquel Urtasun
We show that our approach can outperform the state-of-the-art on both datasets.
no code implementations • ECCV 2020 • Abbas Sadat, Sergio Casas, Mengye Ren, Xinyu Wu, Pranaab Dhawan, Raquel Urtasun
In this paper we propose a novel end-to-end learnable network that performs joint perception, prediction and motion planning for self-driving vehicles and produces interpretable intermediate representations.
no code implementations • NeurIPS 2020 • Yuwen Xiong, Mengye Ren, Raquel Urtasun
Deep neural nets typically perform end-to-end backpropagation to learn the weights, a procedure that creates synchronization constraints in the weight update step across layers and is not biologically plausible.
1 code implementation • ICLR 2021 • Mengye Ren, Michael L. Iuzzolino, Michael C. Mozer, Richard S. Zemel
We aim to bridge the gap between typical human and machine-learning environments by extending the standard framework of few-shot learning to an online, continual setting.
1 code implementation • ICML 2020 • Quinlan Sykora, Mengye Ren, Raquel Urtasun
In this paper we tackle the problem of routing multiple agents in a coordinated manner.
no code implementations • CVPR 2020 • James Tu, Mengye Ren, Siva Manivasagam, Ming Liang, Bin Yang, Richard Du, Frank Cheng, Raquel Urtasun
Modern autonomous driving systems rely heavily on deep learning models to process point cloud sensory data; meanwhile, deep models have been shown to be susceptible to adversarial attacks with visually imperceptible perturbations.
no code implementations • 24 Oct 2019 • Kelvin Wong, Shenlong Wang, Mengye Ren, Ming Liang, Raquel Urtasun
In the past few years, we have seen great progress in perception algorithms, particular through the use of deep learning.
no code implementations • 10 Oct 2019 • Yuwen Xiong, Mengye Ren, Raquel Urtasun
Recent studies on catastrophic forgetting during sequential learning typically focus on fixing the accuracy of the predictions for a previously learned task.
no code implementations • 10 Oct 2019 • Abbas Sadat, Mengye Ren, Andrei Pokrovsky, Yen-Chen Lin, Ersin Yumer, Raquel Urtasun
The motion planners used in self-driving vehicles need to generate trajectories that are safe, comfortable, and obey the traffic rules.
no code implementations • 30 Jul 2019 • Yuwen Xiong, Mengye Ren, Renjie Liao, Kelvin Wong, Raquel Urtasun
Point clouds are the native output of many real-world 3D sensors.
1 code implementation • NeurIPS 2019 • Mengye Ren, Renjie Liao, Ethan Fetaya, Richard S. Zemel
This paper addresses this problem, incremental few-shot learning, where a regular classification network has already been trained to recognize a set of base classes, and several extra novel classes are being considered, each with only a few labeled examples.
1 code implementation • ICLR 2019 • Chris Zhang, Mengye Ren, Raquel Urtasun
Neural architecture search (NAS) automatically finds the best task-specific neural network topology, outperforming many manual architecture designs.
9 code implementations • ICML 2018 • Mengye Ren, Wenyuan Zeng, Bin Yang, Raquel Urtasun
Deep neural networks have been shown to be very powerful modeling tools for many supervised learning tasks involving complex input patterns.
1 code implementation • ICLR 2018 • Yuhuai Wu, Mengye Ren, Renjie Liao, Roger Grosse
Careful tuning of the learning rate, or even schedules thereof, can be crucial to effective neural net training.
8 code implementations • ICLR 2018 • Mengye Ren, Eleni Triantafillou, Sachin Ravi, Jake Snell, Kevin Swersky, Joshua B. Tenenbaum, Hugo Larochelle, Richard S. Zemel
To address this paradigm, we propose novel extensions of Prototypical Networks (Snell et al., 2017) that are augmented with the ability to use unlabeled examples when producing prototypes.
2 code implementations • CVPR 2018 • Mengye Ren, Andrei Pokrovsky, Bin Yang, Raquel Urtasun
Conventional deep convolutional neural networks (CNNs) apply convolution operators uniformly in space across all feature maps for hundreds of layers - this incurs a high computational cost for real-time applications.
9 code implementations • NeurIPS 2017 • Aidan N. Gomez, Mengye Ren, Raquel Urtasun, Roger B. Grosse
Deep residual networks (ResNets) have significantly pushed forward the state-of-the-art on image classification, increasing in performance as networks grow both deeper and wider.
no code implementations • 14 Nov 2016 • Mengye Ren, Renjie Liao, Raquel Urtasun, Fabian H. Sinz, Richard S. Zemel
On the other hand, layer normalization normalizes the activations across all activities within a layer.
1 code implementation • CVPR 2017 • Mengye Ren, Richard S. Zemel
While convolutional neural networks have gained impressive success recently in solving structured prediction problems such as semantic segmentation, it remains a challenge to differentiate individual object instances in the scene.
3 code implementations • NeurIPS 2015 • Mengye Ren, Ryan Kiros, Richard Zemel
A suite of baseline results on this new dataset are also presented.
Ranked #5 on
Video Question Answering
on SUTD-TrafficQA