no code implementations • 14 Mar 2025 • Chonghao Sima, Kashyap Chitta, Zhiding Yu, Shiyi Lan, Ping Luo, Andreas Geiger, Hongyang Li, Jose M. Alvarez
In this work, we propose Centaur (Cluster Entropy for Test-time trAining using Uncertainty) which updates a planner's behavior via test-time training, without relying on hand-engineered rules or cost functions.
no code implementations • 5 Mar 2025 • Zi Wang, Shiyi Lan, Xinglong Sun, Nadine Chang, Zhenxin Li, Zhiding Yu, Jose M. Alvarez
In this paper, we propose SafeFusion, a training framework to learn from collision data.
1 code implementation • 3 Feb 2025 • Jose M. Alvarez, Salvatore Ruggieri
ST finds for each complainant similar protected and non-protected instances in the dataset; constructs, respectively, a control and test group; and compares the groups such that a difference in outcomes implies a potential case of individual discrimination.
1 code implementation • 20 Jan 2025 • Zhiqi Li, Guo Chen, Shilong Liu, Shihao Wang, Vibashan VS, Yishen Ji, Shiyi Lan, Hao Zhang, Yilin Zhao, Subhashree Radhakrishnan, Nadine Chang, Karan Sapra, Amala Sanjay Deshmukh, Tuomas Rintamaki, Matthieu Le, Ilia Karmanov, Lukas Voegtle, Philipp Fischer, De-An Huang, Timo Roman, Tong Lu, Jose M. Alvarez, Bryan Catanzaro, Jan Kautz, Andrew Tao, Guilin Liu, Zhiding Yu
Recently, promising progress has been made by open-source vision-language models (VLMs) in bringing their capabilities closer to those of proprietary frontier models.
no code implementations • 2 Dec 2024 • Alberto Gonzalo Rodriguez Salgado, Maying Shen, Philipp Harzig, Peter Mayer, Jose M. Alvarez
Robustness to out-of-distribution data is crucial for deploying modern neural networks.
no code implementations • 31 Oct 2024 • Nikita Durasov, Rafid Mahmood, Jiwoong Choi, Marc T. Law, James Lucas, Pascal Fua, Jose M. Alvarez
We introduce a framework for quantifying uncertainty in 3D object detection by leveraging an evidential learning loss on Bird's Eye View representations in the 3D detector.
no code implementations • 20 Sep 2024 • Maying Shen, Nadine Chang, Sifei Liu, Jose M. Alvarez
We quantitatively show that our Semantic Selection and Enrichment framework (SSE) can a) successfully maintain model performance with a smaller training dataset and b) improve model performance by enriching the smaller dataset without exceeding the original dataset size.
no code implementations • 9 Jul 2024 • Barath Lakshmanan, Joshua Chen, Shiyi Lan, Maying Shen, Zhiding Yu, Jose M. Alvarez
The cornerstone of autonomous vehicles (AV) is a solid perception system, where camera encoders play a crucial role.
2 code implementations • 11 Jun 2024 • Zhenxin Li, Kailin Li, Shihao Wang, Shiyi Lan, Zhiding Yu, Yishen Ji, Zhiqi Li, Ziyue Zhu, Jan Kautz, Zuxuan Wu, Yu-Gang Jiang, Jose M. Alvarez
We propose Hydra-MDP, a novel paradigm employing multiple teachers in a teacher-student model.
no code implementations • 6 Jun 2024 • Maying Shen, Hongxu Yin, Pavlo Molchanov, Lei Mao, Jose M. Alvarez
First, a novel feature regularization (FeatReg) to retain and refine knowledge from existing checkpoints; Second, we propose adaptive knowledge distillation (AdaKD), a novel approach to forget mitigation and knowledge transfer.
1 code implementation • 27 May 2024 • Yiming Li, Zehong Wang, Yue Wang, Zhiding Yu, Zan Gojcic, Marco Pavone, Chen Feng, Jose M. Alvarez
Humans naturally retain memories of permanent elements, while ephemeral moments often slip through the cracks of memory.
1 code implementation • 22 May 2024 • Jose M. Alvarez, Salvatore Ruggieri
The CP comparator is the standard one among discrimination testing, representing an idealized comparison as it aims for having a complainant-comparator pair that only differs on membership to the protected attribute.
1 code implementation • 2 May 2024 • Shihao Wang, Zhiding Yu, Xiaohui Jiang, Shiyi Lan, Min Shi, Nadine Chang, Jan Kautz, Ying Li, Jose M. Alvarez
We further propose OmniDrive-nuScenes, a new visual question-answering dataset challenging the true 3D situational awareness of a model with comprehensive visual question-answering (VQA) tasks, including scene description, traffic regulation, 3D grounding, counterfactual reasoning, decision making and planning.
1 code implementation • CVPR 2024 • Hoang Chuong Nguyen, Tianyu Wang, Jose M. Alvarez, Miaomiao Liu
In the next stage, we use an object network to estimate the depth of those moving objects assuming rigid motions.
no code implementations • 1 Apr 2024 • Shuaiyi Huang, De-An Huang, Zhiding Yu, Shiyi Lan, Subhashree Radhakrishnan, Jose M. Alvarez, Abhinav Shrivastava, Anima Anandkumar
Video instance segmentation (VIS) is a challenging vision task that aims to detect, segment, and track objects in videos.
no code implementations • CVPR 2024 • Zetong Yang, Zhiding Yu, Chris Choy, Renhao Wang, Anima Anandkumar, Jose M. Alvarez
This mapping allows the depth estimation of distant objects conditioned on their 2D boxes, making long-range 3D detection with 2D supervision feasible.
no code implementations • 24 Jan 2024 • Jose M. Alvarez, Salvatore Ruggieri
We present two kinds of causal perception, unfaithful and inconsistent, based on the SCM properties of faithfulness and consistency.
1 code implementation • ICCV 2023 • Bingyin Zhao, Zhiding Yu, Shiyi Lan, Yutao Cheng, Anima Anandkumar, Yingjie Lao, Jose M. Alvarez
With the proposed STL framework, our best model based on FAN-L-Hybrid (77. 3M parameters) achieves 84. 8% Top-1 accuracy and 42. 1% mCE on ImageNet-1K and ImageNet-C, and sets a new state-of-the-art for ImageNet-A (46. 1%) and ImageNet-R (56. 6%) without using extra data, outperforming the original FAN counterpart by significant margins.
Ranked #16 on
Domain Generalization
on ImageNet-C
1 code implementation • CVPR 2024 • Zhiqi Li, Zhiding Yu, Shiyi Lan, Jiahan Li, Jan Kautz, Tong Lu, Jose M. Alvarez
We initially observed that the nuScenes dataset, characterized by relatively simple driving scenarios, leads to an under-utilization of perception information in end-to-end models incorporating ego status, such as the ego vehicle's velocity.
1 code implementation • CVPR 2024 • Zhenxin Li, Shiyi Lan, Jose M. Alvarez, Zuxuan Wu
Recently, the rise of query-based Transformer decoders is reshaping camera-based 3D object detection.
1 code implementation • 24 Nov 2023 • Lingchen Meng, Shiyi Lan, Hengduo Li, Jose M. Alvarez, Zuxuan Wu, Yu-Gang Jiang
In-context segmentation aims at segmenting novel images using a few labeled example images, termed as "in-context examples", exploring content similarities between examples and the target.
1 code implementation • 30 Oct 2023 • Ali Hatamizadeh, Michael Ranzinger, Shiyi Lan, Jose M. Alvarez, Sanja Fidler, Jan Kautz
Inspired by this trend, we propose a new class of computer vision models, dubbed Vision Retention Networks (ViR), with dual parallel and recurrent formulations, which strike an optimal balance between fast inference and parallel training with competitive performance.
no code implementations • ICCV 2023 • Tzofi Klinghoffer, Jonah Philion, Wenzheng Chen, Or Litany, Zan Gojcic, Jungseock Joo, Ramesh Raskar, Sanja Fidler, Jose M. Alvarez
We introduce a technique for novel view synthesis and use it to transform collected data to the viewpoint of target rigs, allowing us to train BEV segmentation models for diverse target rigs without any additional data collection or labeling cost.
1 code implementation • ICCV 2023 • Zhiqi Li, Zhiding Yu, Wenhai Wang, Anima Anandkumar, Tong Lu, Jose M. Alvarez
Currently, the two most prominent VTM paradigms are forward projection and backward projection.
1 code implementation • 28 Jul 2023 • Jose M. Alvarez, Antonio Mastropietro, Salvatore Ruggieri
To study the impact of the ISO, we introduce a human-like screener and compare it to its algorithmic counterpart, where the human-like screener is conceived to be inconsistent over time.
1 code implementation • 9 Jul 2023 • Jiayu Yang, Enze Xie, Miaomiao Liu, Jose M. Alvarez
In contrast, we propose to use parametric depth distribution modeling for feature transformation.
1 code implementation • 4 Jul 2023 • Zhiqi Li, Zhiding Yu, David Austin, Mingsheng Fang, Shiyi Lan, Jan Kautz, Jose M. Alvarez
This technical report summarizes the winning solution for the 3D Occupancy Prediction Challenge, which is held in conjunction with the CVPR 2023 Workshop on End-to-End Autonomous Driving and CVPR 23 Workshop on Vision-Centric Autonomous Driving Workshop.
Ranked #2 on
Prediction Of Occupancy Grid Maps
on Occ3D-nuScenes
2 code implementations • 9 Jun 2023 • Ali Hatamizadeh, Greg Heinrich, Hongxu Yin, Andrew Tao, Jose M. Alvarez, Jan Kautz, Pavlo Molchanov
At a high level, global self-attentions enable the efficient cross-window communication at lower costs.
Ranked #188 on
Image Classification
on ImageNet
1 code implementation • 27 Feb 2023 • Jose M. Alvarez, Kristen M. Scott, Salvatore Ruggieri, Bettina Berendt
In uses of pre-trained machine learning models, it is a known issue that the target population in which the model is being deployed may not have been reflected in the source population with which the model was trained.
1 code implementation • CVPR 2023 • Yiming Li, Zhiding Yu, Christopher Choy, Chaowei Xiao, Jose M. Alvarez, Sanja Fidler, Chen Feng, Anima Anandkumar
To enable such capability in AI systems, we propose VoxFormer, a Transformer-based semantic scene completion framework that can output complete 3D volumetric semantics from only 2D images.
3D geometry
3D Semantic Scene Completion from a single RGB image
+1
1 code implementation • 23 Feb 2023 • Jose M. Alvarez, Salvatore Ruggieri
For any complainant, we find and compare similar protected and non-protected instances in the dataset used by the classifier to construct a control and test group, where a difference between the decision outcomes of the two groups implies potential individual discrimination.
no code implementations • CVPR 2023 • Shiyi Lan, Xitong Yang, Zhiding Yu, Zuxuan Wu, Jose M. Alvarez, Anima Anandkumar
We propose Mask Auto-Labeler (MAL), a high-quality Transformer-based mask auto-labeling framework for instance segmentation using only box annotations.
1 code implementation • ICCV 2023 • Yilun Chen, Zhiding Yu, Yukang Chen, Shiyi Lan, Anima Anandkumar, Jiaya Jia, Jose M. Alvarez
For 3D object detection, we instantiate this method as FocalFormer3D, a simple yet effective detector that excels at excavating difficult objects and improving prediction recall.
1 code implementation • ICCV 2023 • Jiayu Yang, Enze Xie, Miaomiao Liu, Jose M. Alvarez
In contrast, we propose to use parametric depth distribution modeling for feature transformation.
1 code implementation • 4 Nov 2022 • Ryan Humble, Maying Shen, Jorge Albericio Latorre, Eric Darve1, Jose M. Alvarez
Structured channel pruning has been shown to significantly accelerate inference time for convolution neural networks (CNNs) on modern hardware, with a relatively minor loss of network accuracy.
1 code implementation • 13 Oct 2022 • Maying Shen, Hongxu Yin, Pavlo Molchanov, Lei Mao, Jianna Liu, Jose M. Alvarez
We propose Hardware-Aware Latency Pruning (HALP) that formulates structural pruning as a global resource allocation optimization problem, aiming at maximizing the accuracy while constraining latency under a predefined budget on targeting device.
no code implementations • 3 Oct 2022 • Rafid Mahmood, James Lucas, Jose M. Alvarez, Sanja Fidler, Marc T. Law
Modern deep learning systems require huge data sets to achieve impressive performance, but there is little guidance on how much or what kind of data to collect.
no code implementations • 5 Jul 2022 • Suraj Kothawade, Donna Roy, Michele Fenzi, Elmar Haussmann, Jose M. Alvarez, Christoph Angerer
Existing semantic image retrieval methods often focus on mining for larger sized geographical landmarks, and/or require extra labeled data, such as images/image-pairs with similar objects, for mining images with generic objects.
no code implementations • CVPR 2022 • Rafid Mahmood, James Lucas, David Acuna, Daiqing Li, Jonah Philion, Jose M. Alvarez, Zhiding Yu, Sanja Fidler, Marc T. Law
Given a small training data set and a learning algorithm, how much more data is necessary to reach a target validation or test performance?
no code implementations • CVPR 2023 • Shuxuan Guo, Yinlin Hu, Jose M. Alvarez, Mathieu Salzmann
Knowledge distillation facilitates the training of a compact student network by using a deep teacher one.
1 code implementation • CVPR 2022 • Jiayu Yang, Jose M. Alvarez, Miaomiao Liu
Boundary pixels usually follow a multi-modal distribution as they represent different depths; Therefore, the assumption results in an erroneous depth prediction at the coarser level of the cost volume pyramid and can not be corrected in the refinement levels leading to wrong depth predictions.
2 code implementations • 26 Apr 2022 • Daquan Zhou, Zhiding Yu, Enze Xie, Chaowei Xiao, Anima Anandkumar, Jiashi Feng, Jose M. Alvarez
Our study is motivated by the intriguing properties of the emerging visual grouping in Vision Transformers, which indicates that self-attention may promote robustness through improved mid-level representations.
Ranked #4 on
Domain Generalization
on ImageNet-R
(using extra training data)
no code implementations • 11 Apr 2022 • Enze Xie, Zhiding Yu, Daquan Zhou, Jonah Philion, Anima Anandkumar, Sanja Fidler, Ping Luo, Jose M. Alvarez
In this paper, we propose M$^2$BEV, a unified framework that jointly performs 3D object detection and map segmentation in the Birds Eye View~(BEV) space with multi-camera image inputs.
1 code implementation • CVPR 2022 • Xinlong Wang, Zhiding Yu, Shalini De Mello, Jan Kautz, Anima Anandkumar, Chunhua Shen, Jose M. Alvarez
FreeSOLO further demonstrates superiority as a strong pre-training method, outperforming state-of-the-art self-supervised pre-training methods by +9. 8% AP when fine-tuning instance segmentation with only 5% COCO masks.
2 code implementations • 27 Jan 2022 • Carlos Mougan, Jose M. Alvarez, Salvatore Ruggieri, Steffen Staab
We investigate the interaction between categorical encodings and target encoding regularization methods that reduce unfairness.
no code implementations • 18 Nov 2021 • Xinnan Du, William Zhang, Jose M. Alvarez
In this paper, we propose a new light-weight self-supervised learning framework that could boost supervised learning performance with minimum additional computation cost.
no code implementations • CVPR 2022 • Maying Shen, Pavlo Molchanov, Hongxu Yin, Jose M. Alvarez
Through extensive experiments on ImageNet, we show that EPI empowers a quick tracking of early training epochs suitable for pruning, offering same efficacy as an otherwise ``oracle'' grid-search that scans through epochs and requires orders of magnitude more compute.
1 code implementation • 20 Oct 2021 • Maying Shen, Hongxu Yin, Pavlo Molchanov, Lei Mao, Jianna Liu, Jose M. Alvarez
We propose Hardware-Aware Latency Pruning (HALP) that formulates structural pruning as a global resource allocation optimization problem, aiming at maximizing the accuracy while constraining latency under a predefined budget.
3 code implementations • CVPR 2022 • Zhiqi Li, Wenhai Wang, Enze Xie, Zhiding Yu, Anima Anandkumar, Jose M. Alvarez, Ping Luo, Tong Lu
Specifically, we supervise the attention modules in the mask decoder in a layer-wise manner.
Ranked #4 on
Panoptic Segmentation
on COCO test-dev
no code implementations • 13 Jul 2021 • Xin Dong, Hongxu Yin, Jose M. Alvarez, Jan Kautz, Pavlo Molchanov, H. T. Kung
Prior works usually assume that SC offers privacy benefits as only intermediate features, instead of private data, are shared from devices to the cloud.
no code implementations • CVPR 2022 • Ismail Elezi, Zhiding Yu, Anima Anandkumar, Laura Leal-Taixe, Jose M. Alvarez
Deep neural networks have reached high accuracy on object detection but their success hinges on large amounts of labeled data.
no code implementations • CVPR 2021 • Yerlan Idelbayev, Pavlo Molchanov, Maying Shen, Hongxu Yin, Miguel A. Carreira-Perpinan, Jose M. Alvarez
We study the problem of quantizing N sorted, scalar datapoints with a fixed codebook containing K entries that are allowed to be rescaled.
1 code implementation • NeurIPS 2021 • Shuxuan Guo, Jose M. Alvarez, Mathieu Salzmann
Knowledge distillation constitutes a simple yet effective way to improve the performance of a compact student network by exploiting the knowledge of a more powerful teacher.
26 code implementations • NeurIPS 2021 • Enze Xie, Wenhai Wang, Zhiding Yu, Anima Anandkumar, Jose M. Alvarez, Ping Luo
We present SegFormer, a simple, efficient yet powerful semantic segmentation framework which unifies Transformers with lightweight multilayer perception (MLP) decoders.
Ranked #1 on
Semantic Segmentation
on COCO-Stuff full
1 code implementation • CVPR 2021 • Hongxu Yin, Arun Mallya, Arash Vahdat, Jose M. Alvarez, Jan Kautz, Pavlo Molchanov
In this work, we introduce GradInversion, using which input images from a larger batch (8 - 48 images) can also be recovered for large networks such as ResNets (50 layers), on complex datasets such as ImageNet (1000 classes, 224x224 px).
1 code implementation • 12 Apr 2021 • Nadine Chang, Zhiding Yu, Yu-Xiong Wang, Anima Anandkumar, Sanja Fidler, Jose M. Alvarez
As a result, image resampling alone is not enough to yield a sufficiently balanced distribution at the object level.
1 code implementation • CVPR 2021 • Jiayu Yang, Jose M. Alvarez, Miaomiao Liu
Here, we propose a self-supervised learning framework for multi-view stereo that exploit pseudo labels from the input data.
2 code implementations • ICLR 2021 • Wuyang Chen, Zhiding Yu, Shalini De Mello, Sifei Liu, Jose M. Alvarez, Zhangyang Wang, Anima Anandkumar
Training on synthetic data can be beneficial for label or data-scarce scenarios.
1 code implementation • ICCV 2021 • Jiwoong Choi, Ismail Elezi, Hyuk-Jae Lee, Clement Farabet, Jose M. Alvarez
Most of these methods are based on multiple models or are straightforward extensions of classification methods, hence estimate an image's informativeness using only the classification head.
no code implementations • 1 Jan 2021 • Jiwoong Choi, Ismail Elezi, Hyuk-Jae Lee, Clement Farabet, Jose M. Alvarez
For active learning, we propose a scoring function that aggregates uncertainties from both the classification and the localization outputs of the network.
3 code implementations • ICLR 2021 • Michael Zhang, Karan Sapra, Sanja Fidler, Serena Yeung, Jose M. Alvarez
While federated learning traditionally aims to train a single global model across decentralized local datasets, one model may not always be ideal for all participating clients.
no code implementations • 9 Apr 2020 • Elmar Haussmann, Michele Fenzi, Kashyap Chitta, Jan Ivanecky, Hanson Xu, Donna Roy, Akshita Mittel, Nicolas Koumchatzky, Clement Farabet, Jose M. Alvarez
We have built a scalable production system for active learning in the domain of autonomous driving.
3 code implementations • 30 Mar 2020 • Ronak Kosti, Jose M. Alvarez, Adria Recasens, Agata Lapedriza
In this paper we present EMOTIC, a dataset of images of people in a diverse set of natural situations, annotated with their apparent emotion.
Ranked #5 on
Emotion Recognition in Context
on EMOTIC
(using extra training data)
2 code implementations • CVPR 2020 • Hongxu Yin, Pavlo Molchanov, Zhizhong Li, Jose M. Alvarez, Arun Mallya, Derek Hoiem, Niraj K. Jha, Jan Kautz
We introduce DeepInversion, a new method for synthesizing images from the image distribution used to train a deep neural network.
1 code implementation • CVPR 2020 • Jiayu Yang, Wei Mao, Jose M. Alvarez, Miaomiao Liu
We propose a cost volume-based neural network for depth inference from multi-view images.
Ranked #15 on
3D Reconstruction
on DTU
no code implementations • 25 Sep 2019 • Kashyap Chitta, Jose M. Alvarez, Elmar Haussmann, Clement Farabet
In this paper, we propose to scale up ensemble Active Learning methods to perform acquisition at a large scale (10k to 500k samples at a time).
no code implementations • 10 Sep 2019 • Shuang Gao, Xin Liu, Lung-Sheng Chien, William Zhang, Jose M. Alvarez
Our approach can effectively improve the structural sparsity of residual models.
1 code implementation • 27 Jul 2019 • Kashyap Chitta, Jose M. Alvarez, Martial Hebert
Semantic segmentation with Convolutional Neural Networks is a memory-intensive task due to the high spatial resolution of feature maps and output predictions.
no code implementations • 29 May 2019 • Kashyap Chitta, Jose M. Alvarez, Elmar Haussmann, Clement Farabet
In this paper, we propose to scale up ensemble Active Learning (AL) methods to perform acquisition at a large scale (10k to 500k samples at a time).
no code implementations • 29 Nov 2018 • Jiaming Zeng, Adam Lesnikowski, Jose M. Alvarez
One of the main challenges of deep learning tools is their inability to capture model uncertainty.
no code implementations • NeurIPS 2020 • Shuxuan Guo, Jose M. Alvarez, Mathieu Salzmann
As evidenced by our experiments, our approach outperforms both training the compact network from scratch and performing knowledge distillation from a teacher.
no code implementations • 8 Nov 2018 • Kashyap Chitta, Jose M. Alvarez, Adam Lesnikowski
Annotating the right data for training deep neural networks is an important challenge.
no code implementations • 6 Nov 2018 • Kashyap Chitta, Jose M. Alvarez, Adam Lesnikowski
In this paper, we introduce Deep Probabilistic Ensembles (DPEs), a scalable technique that uses a regularized ensemble to approximate a deep Bayesian Neural Network (BNN).
no code implementations • ECCV 2018 • Fatemeh Sadat Saleh, Mohammad Sadegh Aliakbarian, Mathieu Salzmann, Lars Petersson, Jose M. Alvarez
Our approach builds on the observation that foreground and background classes are not affected in the same manner by the domain shift, and thus should be treated differently.
no code implementations • NeurIPS 2017 • Jose M. Alvarez, Mathieu Salzmann
In recent years, great progress has been made in a variety of application domains thanks to the development of increasingly deeper neural networks.
2 code implementations • ICCV 2017 • Marc Masana, Joost Van de Weijer, Luis Herranz, Andrew D. Bagdanov, Jose M. Alvarez
We show that domain transfer leads to large shifts in network activations and that it is desirable to take this into account when compressing.
no code implementations • ICCV 2017 • Fatemeh Sadat Saleh, Mohammad Sadegh Aliakbarian, Mathieu Salzmann, Lars Petersson, Jose M. Alvarez
Our experiments demonstrate the benefits of our classifier heatmaps and of our two-stream architecture on challenging urban scene datasets and on the YouTube-Objects benchmark, where we obtain state-of-the-art results.
2 code implementations • 9 Jul 2017 • Albert Jimenez, Jose M. Alvarez, Xavier Giro-i-Nieto
In this paper, we go beyond this spatial information and propose a local-aware encoding of convolutional features based on semantic information predicted in the target image.
no code implementations • CVPR 2017 • Ronak Kosti, Jose M. Alvarez, Adria Recasens, Agata Lapedriza
In this paper we present the Emotions in Context Database (EMCO), a dataset of images containing people in context in non-controlled environments.
no code implementations • 6 Jun 2017 • Fatemeh Sadat Saleh, Mohammad Sadegh Aliakbarian, Mathieu Salzmann, Lars Petersson, Jose M. Alvarez, Stephen Gould
We then show how to obtain multi-class masks by the fusion of foreground/background ones with information extracted from a weakly-supervised localization network.
no code implementations • NeurIPS 2016 • Jose M. Alvarez, Mathieu Salzmann
In this paper, we introduce an approach to automatically determining the number of neurons in each layer of a deep network during learning.
no code implementations • 2 Sep 2016 • Fatemehsadat Saleh, Mohammad Sadegh Ali Akbarian, Mathieu Salzmann, Lars Petersson, Stephen Gould, Jose M. Alvarez
Hence, weak supervision using only image tags could have a significant impact in semantic segmentation.
no code implementations • 19 Mar 2016 • Gucan Long, Laurent Kneip, Jose M. Alvarez, Hongdong Li
This work presents an unsupervised learning based approach to the ubiquitous computer vision problem of image matching.
no code implementations • 11 Dec 2014 • Jose M. Alvarez, Theo Gevers, Antonio M. Lopez
These algorithms reduce the effect of lighting variations and weather conditions by exploiting the discriminant/invariant properties of different color representations.