no code implementations • 31 Oct 2024 • Nikita Durasov, Rafid Mahmood, Jiwoong Choi, Marc T. Law, James Lucas, Pascal Fua, Jose M. Alvarez
We introduce a framework for quantifying uncertainty in 3D object detection by leveraging an evidential learning loss on Bird's Eye View representations in the 3D detector.
no code implementations • 25 Aug 2024 • Haitz Sáez de Ocáriz Borde, Anastasis Kratsios, Marc T. Law, Xiaowen Dong, Michael Bronstein
NSTs are implemented as three neural networks trained in an end-to-end manner: an embedding network, which learns to optimize the location of nodes as events in the spacetime manifold, and two other networks that optimize the space and time geometries in parallel, which we call a neural (quasi-)metric and a neural partial order, respectively.
no code implementations • 5 Feb 2024 • Anastasis Kratsios, Haitz Sáez de Ocáriz Borde, Takashi Furuya, Marc T. Law
Mixture-of-Experts (MoEs) can scale up beyond traditional deep learning models by employing a routing strategy in which each input is processed by a single "expert" deep learning model.
no code implementations • 7 Dec 2023 • Derek Lim, Haggai Maron, Marc T. Law, Jonathan Lorraine, James Lucas
However, those works developed architectures tailored to specific networks such as MLPs and CNNs without normalization layers, and generalizing such architectures to other types of networks can be challenging.
no code implementations • 9 Feb 2023 • Viraj Prabhu, David Acuna, Andrew Liao, Rafid Mahmood, Marc T. Law, Judy Hoffman, Sanja Fidler, James Lucas
Sim2Real domain adaptation (DA) research focuses on the constrained setting of adapting from a labeled synthetic source domain to an unlabeled or sparsely labeled real target domain.
no code implementations • 3 Oct 2022 • Rafid Mahmood, James Lucas, Jose M. Alvarez, Sanja Fidler, Marc T. Law
Modern deep learning systems require huge data sets to achieve impressive performance, but there is little guidance on how much or what kind of data to collect.
no code implementations • CVPR 2022 • Rafid Mahmood, James Lucas, David Acuna, Daiqing Li, Jonah Philion, Jose M. Alvarez, Zhiding Yu, Sanja Fidler, Marc T. Law
Given a small training data set and a learning algorithm, how much more data is necessary to reach a target validation or test performance?
1 code implementation • 10 Nov 2021 • Rahul Vigneswaran, Marc T. Law, Vineeth N. Balasubramanian, Makarand Tapaswi
Oversampling instances of the tail classes attempts to solve this imbalance.
Ranked #1 on Long-tail Learning on mini-ImageNet-LT
1 code implementation • 21 Jun 2021 • David Acuna, Guojun Zhang, Marc T. Law, Sanja Fidler
Unsupervised domain adaptation is used in many machine learning applications where, during training, a model has access to unlabeled data in the target domain, and a related labeled dataset.
no code implementations • 5 Jun 2021 • Rafid Mahmood, Sanja Fidler, Marc T. Law
Active learning is the process of training a model with limited labeled data by selecting a core subset of an unlabeled data pool to label.
no code implementations • ICCV 2021 • Aayush Prakash, Shoubhik Debnath, Jean-Francois Lafleche, Eric Cameracci, Gavriel State, Stan Birchfield, Marc T. Law
Synthetic data is emerging as a promising solution to the scalability issue of supervised deep learning, especially when real data are difficult to acquire or hard to annotate.
1 code implementation • NeurIPS 2020 • Marc T. Law, Jos Stam
In machine learning, data is usually represented in a (flat) Euclidean space where distances between points are along straight lines.
1 code implementation • ICCV 2019 • Makarand Tapaswi, Marc T. Law, Sanja Fidler
Understanding videos such as TV series and movies requires analyzing who the characters are and what they are doing.
no code implementations • ICLR 2019 • Marc T. Law, Jake Snell, Amir-Massoud Farahmand, Raquel Urtasun, Richard S. Zemel
Most deep learning models rely on expressive high-dimensional representations to achieve good performance on tasks such as classification.
no code implementations • ICML 2017 • Marc T. Law, Raquel Urtasun, Richard S. Zemel
We derive a closed-form expression for the gradient that is efficient to compute: the complexity to compute the gradient is linear in the size of the training mini-batch and quadratic in the representation dimensionality.
no code implementations • CVPR 2017 • Marc T. Law, Yao-Liang Yu, Raquel Urtasun, Richard S. Zemel, Eric P. Xing
Classic approaches alternate the optimization over the learned metric and the assignment of similar instances.
no code implementations • CVPR 2016 • Marc T. Law, Yao-Liang Yu, Matthieu Cord, Eric P. Xing
Clustering is the task of grouping a set of objects so that objects in the same cluster are more similar to each other than to those in other clusters.
no code implementations • CVPR 2014 • Marc T. Law, Nicolas Thome, Matthieu Cord
This paper introduces a regularization method to explicitly control the rank of a learned symmetric positive semidefinite distance matrix in distance metric learning.