no code implementations • 20 Apr 2023 • Tang Tao, Longfei Gao, Guangrun Wang, Peng Chen, Dayang Hao, Xiaodan Liang, Mathieu Salzmann, Kaicheng Yu
To evaluate the effectiveness of our approach, we establish an object-centric multi-view LiDAR dataset, dubbed NeRF-MVL.
1 code implementation • CVPR 2023 • Haobo Jiang, Zheng Dang, Zhen Wei, Jin Xie, Jian Yang, Mathieu Salzmann
Embedded with the inlier/outlier label, the posterior feature distribution is label-dependent and discriminative.
no code implementations • 29 Mar 2023 • Congpei Qiu, Tong Zhang, Wei Ke, Mathieu Salzmann, Sabine Süsstrunk
Dense Self-Supervised Learning (SSL) methods address the limitations of using image-level feature representations when handling images with multiple objects.
1 code implementation • CVPR 2023 • Yanhao Wu, Tong Zhang, Wei Ke, Sabine Süsstrunk, Mathieu Salzmann
In this paper, we introduce an SSL strategy that leverages positive pairs in both the spatial and temporal domain.
1 code implementation • 23 Mar 2023 • Van Nguyen Nguyen, Thibault Groueix, Yinlin Hu, Mathieu Salzmann, Vincent Lepetit
The practicality of 3D object pose estimation remains limited for many applications due to the need for prior knowledge of a 3D model and a training period for new objects.
1 code implementation • CVPR 2023 • Yang Hai, Rui Song, Jiaojiao Li, Mathieu Salzmann, Yinlin Hu
To address this, we propose a rigidity-aware detection method exploiting the fact that, in 6D pose estimation, the target objects are rigid.
no code implementations • 21 Mar 2023 • Fulin Liu, Yinlin Hu, Mathieu Salzmann
Here, we argue that this conflicts with the averaging nature of the PnP problem, leading to gradients that may encourage the network to degrade the accuracy of individual correspondences.
no code implementations • 16 Mar 2023 • Qiao Wu, Jiaqi Yang, Kun Sun, Chu'ai Zhang, Yanning Zhang, Mathieu Salzmann
Specifically, we introduce two cycle-consistency strategies for supervision: 1) Self tracking cycles, which leverage labels to help the model converge better in the early stages of training; 2) forward-backward cycles, which strengthen the tracker's robustness to motion variations and the template noise caused by the template update strategy.
no code implementations • 12 Mar 2023 • Saqib Javed, Andrew Price, Yinlin Hu, Mathieu Salzmann
Many edge applications, such as collaborative robotics and spacecraft rendezvous, can benefit from 6D object pose estimation, but must do so on embedded platforms.
no code implementations • CVPR 2023 • Vidit Vidit, Martin Engilberge, Mathieu Salzmann
Single Domain Generalization (SDG) tackles the problem of training a model on a single source domain so that it generalizes to any unseen target domain.
no code implementations • CVPR 2023 • Vidit Vidit, Martin Engilberge, Mathieu Salzmann
The performance of modern object detectors drops when the test distribution differs from the training one.
no code implementations • 5 Jan 2023 • Bahar Aydemir, Ludo Hoffstetter, Tong Zhang, Mathieu Salzmann, Sabine Süsstrunk
Deep saliency prediction algorithms complement the object recognition features, they typically rely on additional information, such as scene context, semantic relationships, gaze direction, and object dissimilarity.
no code implementations • CVPR 2023 • Bahar Aydemir, Ludo Hoffstetter, Tong Zhang, Mathieu Salzmann, Sabine Süsstrunk
Deep saliency prediction algorithms complement the object recognition features, they typically rely on additional information such as scene context, semantic relationships, gaze direction, and object dissimilarity.
no code implementations • 29 Dec 2022 • Krzysztof Lis, Matthias Rottmann, Sina Honari, Pascal Fua, Mathieu Salzmann
In other words, vision transformers trained to segment a fixed set of object classes generalize to objects well beyond this set.
no code implementations • 26 Dec 2022 • Baran Ozaydin, Tong Zhang, Sabine Süsstrunk, Mathieu Salzmann
Unpaired exemplar-based image-to-image (UEI2I) translation aims to translate a source image to a target image domain with the style of a target image exemplar, without ground-truth input-translation pairs.
no code implementations • 29 Nov 2022 • Chen Zhao, Yinlin Hu, Mathieu Salzmann
Object location priors have been shown to be critical for the standard 6D object pose estimation setting, where the training and testing objects are the same.
1 code implementation • CVPR 2023 • Luca De Luigi, Ren Li, Benoît Guillard, Mathieu Salzmann, Pascal Fua
Recent approaches to drape garments quickly over arbitrary human bodies leverage self-supervision to eliminate the need for large training sets.
1 code implementation • 8 Oct 2022 • Wei Mao, Miaomiao Liu, Richard Hartley, Mathieu Salzmann
In this paper, we tackle the task of scene-aware 3D human motion forecasting, which consists of predicting future human poses given a 3D scene and a past human motion.
no code implementations • 4 Oct 2022 • Krzysztof Lis, Sina Honari, Pascal Fua, Mathieu Salzmann
While road obstacle detection techniques have become increasingly effective, they typically ignore the fact that, in practice, the apparent size of the obstacles decreases as their distance to the vehicle increases.
1 code implementation • 5 Aug 2022 • Ziyi Zhao, Sena Kiciroglu, Hugues Vinzant, Yuan Cheng, Isinsu Katircioglu, Mathieu Salzmann, Pascal Fua
To evaluate our approach, we introduce a dataset with 3 different physical exercises.
no code implementations • 6 Jun 2022 • Zhichao Huang, Yanbo Fan, Chen Liu, Weizhong Zhang, Yong Zhang, Mathieu Salzmann, Sabine Süsstrunk, Jue Wang
While adversarial training and its variants have shown to be the most effective algorithms to defend against adversarial attacks, their extremely slow training process makes it hard to scale to large datasets like ImageNet.
1 code implementation • CVPR 2022 • Wei Mao, Miaomiao Liu, Mathieu Salzmann
We introduce the task of action-driven stochastic human motion prediction, which aims to predict multiple plausible future motions given a sequence of action labels and a short motion history.
no code implementations • CVPR 2023 • Shuxuan Guo, Yinlin Hu, Jose M. Alvarez, Mathieu Salzmann
Knowledge distillation facilitates the training of a compact student network by using a deep teacher one.
no code implementations • CVPR 2022 • Deblina Bhattacharjee, Tong Zhang, Sabine Süsstrunk, Mathieu Salzmann
At the heart of our approach is a shared attention mechanism modeling the dependencies across the tasks.
no code implementations • CVPR 2022 • Tong Zhang, Congpei Qiu, Wei Ke, Sabine Süsstrunk, Mathieu Salzmann
In essence, this strategy ignores the fact that two crops may truly contain different image information, e. g., background and small objects, and thus tends to restrain the diversity of the learned representations.
2 code implementations • CVPR 2022 • Van Nguyen Nguyen, Yinlin Hu, Yang Xiao, Mathieu Salzmann, Vincent Lepetit
It relies on a small set of training objects to learn local object representations, which allow us to locally match the input image to a set of "templates", rendered images of the CAD models for the new objects.
no code implementations • 29 Mar 2022 • Zheng Dang, Lizhou Wang, Yu Guo, Mathieu Salzmann
Our two contributions are general and can be applied to many existing learning-based 3D object registration frameworks, which we illustrate by implementing them in two of them, DCP and IDAM.
1 code implementation • 18 Mar 2022 • Yinlin Hu, Pascal Fua, Mathieu Salzmann
Given a rough pose estimate obtained from a first network, it uses a second network to predict a dense 2D correspondence field between the image rendered using the rough pose and the real image and infers the required pose correction.
no code implementations • 16 Mar 2022 • Chen Zhao, Yinlin Hu, Mathieu Salzmann
In this paper, we tackle the task of estimating the 3D orientation of previously-unseen objects from monocular images.
1 code implementation • 3 Feb 2022 • Chen Liu, Ziqi Zhao, Sabine Süsstrunk, Mathieu Salzmann
In this paper, we introduce an approach to obtain robust yet compact models by pruning randomly-initialized binary networks.
no code implementations • 14 Dec 2021 • Chen Liu, Zhichao Huang, Mathieu Salzmann, Tong Zhang, Sabine Süsstrunk
This lets us show that the decay in generalization performance of adversarial training is a result of the model's attempt to fit hard adversarial instances.
no code implementations • CVPR 2022 • Andrey Davydov, Anastasia Remizova, Victor Constantin, Sina Honari, Mathieu Salzmann, Pascal Fua
The Skinned Multi-Person Linear (SMPL) model can represent a human body by mapping pose and shape parameters to body meshes.
1 code implementation • NeurIPS 2021 • Krishna Kanth Nakka, Mathieu Salzmann
While effective, deep neural networks (DNNs) are vulnerable to adversarial attacks.
no code implementations • 1 Dec 2021 • Isinsu Katircioglu, Costa Georgantas, Mathieu Salzmann, Pascal Fua
To evaluate this, and because no existing motion prediction datasets depict two closely-interacting subjects, we introduce the LindyHop600K dance dataset.
no code implementations • 19 Nov 2021 • Zheng Dang, Lizhou Wang, Junning Qiu, Minglei Lu, Mathieu Salzmann
We summarise our findings into a set of guidelines and demonstrate their effectiveness by applying them to different baseline methods, DCP and IDAM.
1 code implementation • 12 Nov 2021 • Jan Bednarik, Noam Aigerman, Vladimir G. Kim, Siddhartha Chaudhuri, Shaifali Parashar, Mathieu Salzmann, Pascal Fua
The key to making these correspondences semantically meaningful is to guarantee that the metric tensors computed at corresponding points are as similar as possible.
Ranked #1 on
Surface Reconstruction
on ANIM
1 code implementation • 7 Oct 2021 • Deblina Bhattacharjee, Martin Everaert, Mathieu Salzmann, Sabine Süsstrunk
Estimating the depth of comics images is challenging as such images a) are monocular; b) lack ground-truth depth annotations; c) differ across different artistic styles; d) are sparse and noisy.
Ranked #1 on
Depth Estimation
on eBDtheque
no code implementations • 4 Oct 2021 • Kaicheng Yu, René Ranftl, Mathieu Salzmann
Weight sharing promises to make neural architecture search (NAS) tractable even on commodity hardware.
no code implementations • 29 Sep 2021 • Zhichao Huang, Chen Liu, Mathieu Salzmann, Sabine Süsstrunk, Tong Zhang
Although adversarial training and its variants currently constitute the most effective way to achieve robustness against adversarial attacks, their poor generalization limits their performance on the test samples.
1 code implementation • ICCV 2021 • Wei Mao, Miaomiao Liu, Mathieu Salzmann
Recent progress in stochastic motion prediction, i. e., predicting multiple possible future human motions given a single past pose sequence, has led to producing truly diverse future motions and even providing control over the motion of some body parts.
Ranked #2 on
Human Pose Forecasting
on AMASS
(ADE metric)
1 code implementation • 17 Jun 2021 • Wei Mao, Miaomiao Liu, Mathieu Salzmann, Hongdong Li
Whether based on recurrent or feed-forward neural networks, existing learning based methods fail to model the observation that human motion tends to repeat itself, even for complex sports actions and cooking activities.
no code implementations • 14 Jun 2021 • Vidit Vidit, Mathieu Salzmann
While domain adaptation has been used to improve the performance of object detectors when the training and test data follow different distributions, previous work has mostly focused on two-stage detectors.
1 code implementation • NeurIPS 2021 • Shuxuan Guo, Jose M. Alvarez, Mathieu Salzmann
Knowledge distillation constitutes a simple yet effective way to improve the performance of a compact student network by exploiting the knowledge of a more powerful teacher.
1 code implementation • 30 May 2021 • Samuel von Baußnern, Johannes Otterbach, Adrian Loy, Mathieu Salzmann, Thomas Wollmann
We demonstrate the effectiveness of our approach using an ESPNet trained on the Cityscapes dataset as segmentation model, an affine Normalizing Flow as density estimator and use blue noise to ensure homogeneous sampling.
3 code implementations • 30 Apr 2021 • Robin Chan, Krzysztof Lis, Svenja Uhlemeyer, Hermann Blum, Sina Honari, Roland Siegwart, Pascal Fua, Mathieu Salzmann, Matthias Rottmann
State-of-the-art semantic or instance segmentation deep neural networks (DNNs) are usually trained on a closed set of semantic classes.
1 code implementation • ICCV 2021 • Jan Bednarik, Vladimir G. Kim, Siddhartha Chaudhuri, Shaifali Parashar, Mathieu Salzmann, Pascal Fua, Noam Aigerman
We propose a method for the unsupervised reconstruction of a temporally-coherent sequence of surfaces from a sequence of time-evolving point clouds, yielding dense, semantically meaningful correspondences between all keyframes.
1 code implementation • CVPR 2021 • Kaicheng Yu, Rene Ranftl, Mathieu Salzmann
Weight sharing has become a de facto standard in neural architecture search because it enables the search to be done on commodity hardware.
2 code implementations • 8 Apr 2021 • Wei Wang, Zheng Dang, Yinlin Hu, Pascal Fua, Mathieu Salzmann
Eigendecomposition of symmetric matrices is at the heart of many computer vision algorithms.
1 code implementation • 8 Apr 2021 • Bahar Aydemir, Deblina Bhattacharjee, Tong Zhang, Seungryong Kim, Mathieu Salzmann, Sabine Süsstrunk
Saliency prediction has made great strides over the past two decades, with current techniques modeling low-level information, such as color, intensity and size contrasts, and high-level ones, such as attention and gaze direction for entire objects.
1 code implementation • CVPR 2021 • Yinlin Hu, Sebastien Speierer, Wenzel Jakob, Pascal Fua, Mathieu Salzmann
6D pose estimation in space poses unique challenges that are not commonly encountered in the terrestrial setting.
no code implementations • ICCV 2021 • Chen Zhao, Yixiao Ge, Feng Zhu, Rui Zhao, Hongsheng Li, Mathieu Salzmann
Correspondence selection aims to correctly select the consistent matches (inliers) from an initial set of putative correspondences.
no code implementations • 1 Jan 2021 • Mahsa Baktashmotlagh, Tianle Chen, Mathieu Salzmann
In this setting, existing techniques focus on the challenging task of isolating the unknown target samples, so as to avoid the negative transfer resulting from aligning the source feature distributions with the broader target one that encompasses the additional unknown classes.
no code implementations • 30 Dec 2020 • Krishna Kanth Nakka, Mathieu Salzmann
While these methods were shown to be vulnerable to adversarial attacks, as most deep networks for visual recognition tasks, the existing attacks for VOT trackers all require perturbing the search region of every input frame to be effective, which comes at a non-negligible cost, considering that VOT is a real-time task.
no code implementations • 25 Dec 2020 • Krzysztof Lis, Sina Honari, Pascal Fua, Mathieu Salzmann
Vehicles can encounter a myriad of obstacles on the road, and it is impossible to record them all beforehand to train a detector.
no code implementations • 21 Dec 2020 • Mengshi Qi, Edoardo Remelli, Mathieu Salzmann, Pascal Fua
Deep learning-solutions for hand-object 3D pose and shape estimation are now very effective when an annotated dataset is available to train them to handle the scenarios and lighting conditions they will encounter at test time.
1 code implementation • ICCV 2021 • Isinsu Katircioglu, Helge Rhodin, Jörg Spörri, Mathieu Salzmann, Pascal Fua
Self-supervised detection and segmentation of foreground objects aims for accuracy without annotated training data.
1 code implementation • 8 Dec 2020 • Sena Kiciroglu, Wei Wang, Mathieu Salzmann, Pascal Fua
Long term human motion prediction is essential in safety-critical applications such as human-robot interaction and autonomous driving.
no code implementations • CVPR 2021 • Fatemeh Saleh, Sadegh Aliakbarian, Hamid Rezatofighi, Mathieu Salzmann, Stephen Gould
Despite the recent advances in multiple object tracking (MOT), achieved by joint detection and tracking, dealing with long occlusions remains a challenge.
no code implementations • 2 Dec 2020 • Sina Honari, Victor Constantin, Helge Rhodin, Mathieu Salzmann, Pascal Fua
In this paper we propose an unsupervised feature extraction method to capture temporal information on monocular videos, where we detect and encode subject of interest in each frame and leverage contrastive self-supervised (CSS) learning to extract rich latent vectors.
1 code implementation • 1 Dec 2020 • Weizhe Liu, Mathieu Salzmann, Pascal Fua
Modern methods for counting people in crowded scenes rely on deep networks to estimate people densities in individual images.
1 code implementation • CVPR 2021 • Frank Yu, Mathieu Salzmann, Pascal Fua, Helge Rhodin
Our conclusion is that it is important to utilize camera calibration information when available, for classical and deep-learning-based computer vision alike.
no code implementations • 23 Nov 2020 • Shaifali Parashar, Yuxuan Long, Mathieu Salzmann, Pascal Fua
A recent trend in Non-Rigid Structure-from-Motion (NRSfM) is to express local, differential constraints between pairs of images, from which the surface normal at any point can be obtained by solving a system of polynomial equations.
no code implementations • 23 Nov 2020 • Zheng Dang, Fei Wang, Mathieu Salzmann
While much progress has been made on the task of 3D point cloud registration, there still exists no learning-based method able to estimate the 6D pose of an object observed by a 2. 5D sensor in a scene.
no code implementations • 11 Nov 2020 • Isinsu Katircioglu, Helge Rhodin, Victor Constantin, Jörg Spörri, Mathieu Salzmann, Pascal Fua
While supervised object detection and segmentation methods achieve impressive accuracy, they generalize poorly to images whose appearance significantly differs from the data they have been trained on.
1 code implementation • 14 Oct 2020 • Zhantao Deng, Jan Bednařík, Mathieu Salzmann, Pascal Fua
We introduce an approach that explicitly encourages global consistency of the local mappings.
1 code implementation • 6 Oct 2020 • Tim Lebailly, Sena Kiciroglu, Mathieu Salzmann, Pascal Fua, Wei Wang
We argue that the diverse temporal scales are important as they allow us to look at the past frames with different receptive fields, which can lead to better predictions.
no code implementations • 19 Aug 2020 • Zhigang Li, Yinlin Hu, Mathieu Salzmann, Xiangyang Ji
We achieve state of the art performance on LINEMOD, and OccludedLINEMOD in without real-pose setting, even outperforming methods that rely on real annotations during training on Occluded-LINEMOD.
3 code implementations • ECCV 2020 • Wei Mao, Miaomiao Liu, Mathieu Salzmann
Human motion prediction aims to forecast future human poses given a past motion.
no code implementations • 20 Jul 2020 • Erhan Gundogdu, Victor Constantin, Shaifali Parashar, Amrollah Seifoddini, Minh Dang, Mathieu Salzmann, Pascal Fua
We introduce a two-stream deep network model that produces a visually plausible draping of a template cloth on virtual 3D bodies by extracting features from both the body and garment shapes.
no code implementations • ECCV 2020 • Seungryong Kim, Sabine Süsstrunk, Mathieu Salzmann
We design our VTN as an encoder-decoder network, with modules dedicated to letting the information flow across the feature channels, to account for the dependencies between the semantic parts.
1 code implementation • NeurIPS 2020 • Chen Liu, Mathieu Salzmann, Tao Lin, Ryota Tomioka, Sabine Süsstrunk
We analyze the influence of adversarial training on the loss landscape of machine learning models.
no code implementations • 10 Jun 2020 • Krishna Kanth Nakka, Mathieu Salzmann
In this paper, we identify the proximity of the latent representations of different classes in fine-grained recognition networks as a key factor to the success of adversarial attacks.
no code implementations • 8 Jun 2020 • Zheng Dang, Fei Wang, Mathieu Salzmann
While 3D-3D registration is traditionally tacked by optimization-based methods, recent work has shown that learning-based techniques could achieve faster and more robust results.
no code implementations • ICLR 2020 • Róger Bermúdez-Chacón, Mathieu Salzmann, Pascal Fua
We tackle unsupervised domain adaptation by accounting for the fact that different domains may need to be processed differently to arrive to a common feature representation effective for recognition.
no code implementations • 16 Apr 2020 • Fatemeh Saleh, Sadegh Aliakbarian, Mathieu Salzmann, Stephen Gould
One of the core components in online multiple object tracking (MOT) frameworks is associating new detections with existing tracklets, typically done via a scoring function.
no code implementations • 15 Apr 2020 • Zheng Dang, Kwang Moo Yi, Yinlin Hu, Fei Wang, Pascal Fua, Mathieu Salzmann
In this paper, we introduce an eigendecomposition-free approach to training a deep network whose loss depends on the eigenvector corresponding to a zero eigenvalue of a matrix predicted by the network.
no code implementations • 9 Mar 2020 • Kaicheng Yu, Rene Ranftl, Mathieu Salzmann
Weight sharing promises to make neural architecture search (NAS) tractable even on commodity hardware.
1 code implementation • CVPR 2020 • Sena Kiciroglu, Helge Rhodin, Sudipta N. Sinha, Mathieu Salzmann, Pascal Fua
The accuracy of monocular 3D human pose estimation depends on the viewpoint from which the image is captured.
no code implementations • ICCV 2021 • Sadegh Aliakbarian, Fatemeh Sadat Saleh, Lars Petersson, Stephen Gould, Mathieu Salzmann
We tackle the task of diverse 3D human motion prediction, that is, forecasting multiple plausible future 3D poses given a sequence of observed 3D poses.
1 code implementation • 10 Dec 2019 • Chen Liu, Mathieu Salzmann, Sabine Süsstrunk
Training certifiable neural networks enables one to obtain models with robustness guarantees against adversarial attacks.
1 code implementation • ECCV 2020 • Krishna Kanth Nakka, Mathieu Salzmann
Recently, deep networks have achieved impressive semantic segmentation performance, in particular thanks to their use of larger contextual information.
no code implementations • 26 Nov 2019 • Weizhe Liu, Mathieu Salzmann, Pascal Fua
State-of-the-art methods for counting people in crowded scenes rely on deep networks to estimate crowd density.
1 code implementation • CVPR 2020 • Jan Bednarik, Shaifali Parashar, Erhan Gundogdu, Mathieu Salzmann, Pascal Fua
Generative models that produce point clouds have emerged as a powerful tool to represent 3D surfaces, and the best current ones rely on learning an ensemble of parametric representations.
1 code implementation • ECCV 2020 • Weizhe Liu, Mathieu Salzmann, Pascal Fua
In this paper, we advocate estimating people flows across image locations between consecutive images and inferring the people densities from these flows instead of directly regressing.
1 code implementation • CVPR 2020 • Yinlin Hu, Pascal Fua, Wei Wang, Mathieu Salzmann
Second, training the deep network relies on a surrogate loss that does not directly reflect the final 6D pose estimation task.
1 code implementation • 23 Sep 2019 • Ciprian Tomoiaga, Paul Feng, Mathieu Salzmann, Patrick Jayet
Offline handwriting recognition has undergone continuous progress over the past decades.
4 code implementations • ICCV 2019 • Wei Mao, Miaomiao Liu, Mathieu Salzmann, Hongdong Li
In this paper, we propose a simple feed-forward deep network for motion prediction, which takes into account both temporal smoothness and spatial dependencies among human body joints.
Ranked #6 on
Human Pose Forecasting
on Human3.6M
no code implementations • 2 Aug 2019 • Mohammad Sadegh Aliakbarian, Fatemeh Sadat Saleh, Mathieu Salzmann, Lars Petersson, Stephen Gould, Amirhossein Habibian
In this paper, we introduce an approach to stochastically combine the root of variations with previous pose information, which forces the model to take the noise into account.
no code implementations • 18 Jul 2019 • Isinsu Katircioglu, Helge Rhodin, Victor Constantin, Jörg Spörri, Mathieu Salzmann, Pascal Fua
While supervised object detection methods achieve impressive accuracy, they generalize poorly to images whose appearance significantly differs from the data they have been trained on.
2 code implementations • NeurIPS 2019 • Wei Wang, Zheng Dang, Yinlin Hu, Pascal Fua, Mathieu Salzmann
Eigendecomposition (ED) is widely used in deep networks.
no code implementations • ICCV 2019 • Wei Wang, Kaicheng Yu, Joachim Hugonot, Pascal Fua, Mathieu Salzmann
State-of-the-art segmentation methods rely on very deep networks that are not always easy to train without very large training datasets and tend to be relatively slow to run on standard GPUs.
3 code implementations • ICCV 2019 • Krzysztof Lis, Krishna Nakka, Pascal Fua, Mathieu Salzmann
In this paper, we tackle the more realistic scenario where unexpected objects of unknown classes can appear at test time.
1 code implementation • CVPR 2019 • Helge Rhodin, Victor Constantin, Isinsu Katircioglu, Mathieu Salzmann, Pascal Fua
To this end, we introduce a self-supervised approach to learning what we call a neural scene decomposition (NSD) that can be exploited for 3D pose estimation.
no code implementations • ICLR 2019 • Yassine Benyahia, Kaicheng Yu, Kamil Bennani-Smires, Martin Jaggi, Anthony Davison, Mathieu Salzmann, Claudiu Musat
We identify a phenomenon, which we refer to as multi-model forgetting, that occurs when sequentially training multiple deep networks with partially-shared parameters; the performance of previously-trained models degrades as one optimizes a subsequent one, due to the overwriting of shared parameters.
1 code implementation • ICLR 2020 • Kaicheng Yu, Christian Sciuto, Martin Jaggi, Claudiu Musat, Mathieu Salzmann
Neural Architecture Search (NAS) aims to facilitate the design of deep networks for new tasks.
no code implementations • 8 Jan 2019 • Krishna Kanth Nakka, Mathieu Salzmann
The reason behind the prediction for a new sample can then be interpreted by looking at the visual representation of the most highly activated codeword.
5 code implementations • CVPR 2019 • Yinlin Hu, Joachim Hugonot, Pascal Fua, Mathieu Salzmann
The most recent trend in estimating the 6D pose of rigid objects has been to train deep networks to either directly regress the pose from the image or to predict the 2D locations of 3D keypoints, from which the pose can be obtained using a PnP algorithm.
Ranked #4 on
6D Pose Estimation using RGB
on YCB-Video
no code implementations • ICCV 2019 • Erhan Gundogdu, Victor Constantin, Amrollah Seifoddini, Minh Dang, Mathieu Salzmann, Pascal Fua
We fuse these features with those extracted in parallel from the 3D body, so as to model the cloth-body interactions.
no code implementations • 27 Nov 2018 • Wei Wang, Kaicheng Yu, Joachim Hugonot, Pascal Fua, Mathieu Salzmann
As evidenced by our results on standard hand segmentation benchmarks and on our own dataset, our approach outperforms these other, simpler recurrent segmentation techniques, as well as the state-of-the-art hand segmentation one.
3 code implementations • CVPR 2019 • Weizhe Liu, Mathieu Salzmann, Pascal Fua
State-of-the-art methods for counting people in crowded scenes rely on deep networks to estimate crowd density.
Ranked #1 on
Crowd Counting
on Venice
no code implementations • NeurIPS 2020 • Shuxuan Guo, Jose M. Alvarez, Mathieu Salzmann
As evidenced by our experiments, our approach outperforms both training the compact network from scratch and performing knowledge distillation from a teacher.
1 code implementation • 26 Nov 2018 • Mateusz Koziński, Agata Mosinska, Mathieu Salzmann, Pascal Fua
The difficulty of obtaining annotations to build training databases still slows down the adoption of recent deep learning approaches for biomedical image analysis.
no code implementations • 22 Oct 2018 • Mohammad Sadegh Aliakbarian, Fatemeh Sadat Saleh, Mathieu Salzmann, Basura Fernando, Lars Petersson, Lars Andersson
Action anticipation is critical in scenarios where one needs to react before the action is finalized.
no code implementations • ECCV 2018 • Fatemeh Sadat Saleh, Mohammad Sadegh Aliakbarian, Mathieu Salzmann, Lars Petersson, Jose M. Alvarez
Our approach builds on the observation that foreground and background classes are not affected in the same manner by the domain shift, and thus should be treated differently.
no code implementations • ICLR 2019 • Mahsa Baktashmotlagh, Masoud Faraki, Tom Drummond, Mathieu Salzmann
To this end, we rely on the intuition that the source and target samples depicting the known classes can be generated by a shared subspace, whereas the target samples from unknown classes come from a different, private subspace.
no code implementations • 23 May 2018 • Thomas Joy, Alban Desmaison, Thalaiyasingam Ajanthan, Rudy Bunel, Mathieu Salzmann, Pushmeet Kohli, Philip H. S. Torr, M. Pawan Kumar
The presented algorithms can be applied to any labelling problem using a dense CRF with sparse higher-order potentials.
no code implementations • 14 May 2018 • Krishna Kanth Nakka, Mathieu Salzmann
Structured representations, such as Bags of Words, VLAD and Fisher Vectors, have proven highly effective to tackle complex visual recognition tasks.
no code implementations • CVPR 2018 • Miaomiao Liu, Xuming He, Mathieu Salzmann
By contrast, in this paper, we propose to exploit the 3D geometry of the scene to synthesize a novel view.
2 code implementations • ECCV 2018 • Helge Rhodin, Mathieu Salzmann, Pascal Fua
In this paper, we propose to overcome this problem by learning a geometry-aware body representation from multi-view images without annotations.
Ranked #27 on
Weakly-supervised 3D Human Pose Estimation
on Human3.6M
no code implementations • 23 Mar 2018 • Weizhe Liu, Krzysztof Lis, Mathieu Salzmann, Pascal Fua
In this paper, we explicitly model the scale changes and reason in terms of people per square-meter.
1 code implementation • 23 Mar 2018 • Jan Bednařík, Pascal Fua, Mathieu Salzmann
Recent years have seen the development of mature solutions for reconstructing deformable surfaces from a single image, provided that they are relatively well-textured.
no code implementations • ECCV 2018 • Zheng Dang, Kwang Moo Yi, Yinlin Hu, Fei Wang, Pascal Fua, Mathieu Salzmann
Many classical Computer Vision problems, such as essential matrix computation and pose estimation from 3D to 2D correspondences, can be solved by finding the eigenvector corresponding to the smallest, or zero, eigenvalue of a matrix representing a linear system.
no code implementations • CVPR 2018 • Helge Rhodin, Jörg Spörri, Isinsu Katircioglu, Victor Constantin, Frédéric Meyer, Erich Müller, Mathieu Salzmann, Pascal Fua
Accurate 3D human pose estimation from single images is possible with sophisticated deep-net architectures that have been trained on very large datasets.
1 code implementation • 23 Jan 2018 • Kaicheng Yu, Mathieu Salzmann
Our approach is motivated by a statistical analysis of the network's activations, relying on operations that lead to a Gaussian-distributed final representation, as inherently used by first-order deep networks.
no code implementations • CVPR 2018 • Artem Rozantsev, Mathieu Salzmann, Pascal Fua
The goal of Deep Domain Adaptation is to make it possible to use Deep Nets trained in one domain where there is enough annotated training data in another where there is little or none.
3 code implementations • CVPR 2018 • Kwang Moo Yi, Eduard Trulls, Yuki Ono, Vincent Lepetit, Mathieu Salzmann, Pascal Fua
We develop a deep architecture to learn to find good correspondences for wide-baseline stereo.
no code implementations • NeurIPS 2017 • Jose M. Alvarez, Mathieu Salzmann
In recent years, great progress has been made in a variety of application domains thanks to the development of increasingly deeper neural networks.
no code implementations • 28 Sep 2017 • Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson
We propose to address this issue, by formulating multimodal semantic labeling as inference in a CRF and introducing latent nodes to explicitly model inconsistencies between two modalities.
2 code implementations • NeurIPS 2017 • Pan Ji, Tong Zhang, Hongdong Li, Mathieu Salzmann, Ian Reid
We present a novel deep neural network architecture for unsupervised subspace clustering.
Ranked #3 on
Image Clustering
on Extended Yale-B
no code implementations • ICCV 2017 • Fatemeh Sadat Saleh, Mohammad Sadegh Aliakbarian, Mathieu Salzmann, Lars Petersson, Jose M. Alvarez
Our experiments demonstrate the benefits of our classifier heatmaps and of our two-stream architecture on challenging urban scene datasets and on the YouTube-Objects benchmark, where we obtain state-of-the-art results.
no code implementations • ICML 2017 • Mehrtash Harandi, Mathieu Salzmann, Richard Hartley
To be tractable and robust to data noise, existing metric learning algorithms commonly rely on PCA as a pre-processing step.
1 code implementation • 17 Jul 2017 • Pan Ji, Ian Reid, Ravi Garg, Hongdong Li, Mathieu Salzmann
In this paper, we present a kernel subspace clustering method that can handle non-linear models.
no code implementations • CVPR 2017 • Wei Zhuo, Mathieu Salzmann, Xuming He, Miaomiao Liu
In particular, while some of them aim at segmenting the image into regions, such as object or surface instances, others aim at inferring the semantic labels of given regions, or their support relationships.
no code implementations • 7 Jun 2017 • Pablo Márquez-Neila, Mathieu Salzmann, Pascal Fua
Imposing constraints on the output of a Deep Neural Net is one way to improve the quality of its predictions while loosening the requirements for labeled training data.
no code implementations • 6 Jun 2017 • Fatemeh Sadat Saleh, Mohammad Sadegh Aliakbarian, Mathieu Salzmann, Lars Petersson, Jose M. Alvarez, Stephen Gould
We then show how to obtain multi-class masks by the fusion of foreground/background ones with information extracted from a weakly-supervised localization network.
1 code implementation • ICCV 2017 • Mohammad Sadegh Aliakbarian, Fatemeh Sadat Saleh, Mathieu Salzmann, Basura Fernando, Lars Petersson, Lars Andersson
In contrast to the widely studied problem of recognizing an action given a complete sequence, action anticipation aims to identify the action from only partially available videos.
no code implementations • 20 Mar 2017 • Kaicheng Yu, Mathieu Salzmann
By performing linear combinations and element-wise nonlinear operations, these networks can be thought of as extracting solely first-order information from an input image.
1 code implementation • CVPR 2016 • Thalaiyasingam Ajanthan, Richard Hartley, Mathieu Salzmann
Multi-label submodular Markov Random Fields (MRFs) have been shown to be solvable using max-flow based on an encoding of the labels proposed by Ishikawa, in which each variable $X_i$ is represented by $\ell$ nodes (where $\ell$ is the number of labels) arranged in a column.
no code implementations • CVPR 2017 • Zeeshan Hayder, Xuming He, Mathieu Salzmann
In this context, existing methods typically propose candidate objects, usually as bounding boxes, and directly predict a binary mask within each such proposal.
no code implementations • CVPR 2017 • Thalaiyasingam Ajanthan, Alban Desmaison, Rudy Bunel, Mathieu Salzmann, Philip H. S. Torr, M. Pawan Kumar
To this end, we develop a proximal minimization framework, where the dual of each proximal problem is optimized via block coordinate descent.
no code implementations • NeurIPS 2016 • Jose M. Alvarez, Mathieu Salzmann
In this paper, we introduce an approach to automatically determining the number of neurons in each layer of a deep network during learning.
1 code implementation • ICCV 2017 • Bugra Tekin, Pablo Márquez-Neila, Mathieu Salzmann, Pascal Fua
Most recent approaches to monocular 3D human pose estimation rely on Deep Learning.
Ranked #238 on
3D Human Pose Estimation
on Human3.6M
no code implementations • 17 Nov 2016 • Mohammad Sadegh Aliakbarian, Fatemehsadat Saleh, Basura Fernando, Mathieu Salzmann, Lars Petersson, Lars Andersson
We outperform the state-of-the-art methods that, as us, rely only on RGB frames as input for both action recognition and anticipation.
no code implementations • 2 Sep 2016 • Fatemehsadat Saleh, Mohammad Sadegh Ali Akbarian, Mathieu Salzmann, Lars Petersson, Stephen Gould, Jose M. Alvarez
Hence, weak supervision using only image tags could have a significant impact in semantic segmentation.
no code implementations • CVPR 2016 • Zeeshan Hayder, Xuming He, Mathieu Salzmann
In particular, we introduce a deep structured network that jointly predicts the objectness scores and the bounding box locations of multiple object candidates.
no code implementations • 31 May 2016 • Miaomiao Liu, Mathieu Salzmann, Xuming He
Despite much progress, state-of-the-art techniques suffer from two drawbacks: (i) they rely on the assumption that intensity edges coincide with depth discontinuities, which, unfortunately, is only true in controlled environments; and (ii) they typically exploit the availability of high-resolution training depth maps, which can often not be acquired in practice due to the sensors' limitations.
no code implementations • 20 May 2016 • Mehrtash Harandi, Mathieu Salzmann, Richard Hartley
This lets us formulate dimensionality reduction as the problem of finding a projection that yields a low-dimensional manifold either with maximum discriminative power in the supervised scenario, or with maximum variance of the data in the unsupervised one.
no code implementations • 17 May 2016 • Bugra Tekin, Isinsu Katircioglu, Mathieu Salzmann, Vincent Lepetit, Pascal Fua
Most recent approaches to monocular 3D pose estimation rely on Deep Learning.
Ranked #272 on
3D Human Pose Estimation
on Human3.6M
no code implementations • 21 Mar 2016 • Artem Rozantsev, Mathieu Salzmann, Pascal Fua
To this end, we introduce a two-stream architecture, where one operates in the source domain and the other in the target domain.
no code implementations • CVPR 2016 • Pan Ji, Hongdong Li, Mathieu Salzmann, Yiran Zhong
Feature tracking is a fundamental problem in computer vision, with applications in many computer vision tasks, such as visual SLAM and action recognition.
no code implementations • ICCV 2015 • Weipeng Xu, Mathieu Salzmann, Yongtian Wang, Yue Liu
Capturing the 3D motion of dynamic, non-rigid objects has attracted significant attention in computer vision.
no code implementations • ICCV 2015 • Zeeshan Hayder, Xuming He, Mathieu Salzmann
To exploit the correlations between objects, we build a fully-connected CRF on the candidates, which explicitly incorporates both geometric layout relations across object classes and similarity relations across multiple images.
no code implementations • ICCV 2015 • Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson
In this paper, we address the problem of data misalignment and label inconsistencies, e. g., due to moving objects, in semantic labeling, which violate the assumption of existing techniques.
no code implementations • 19 Nov 2015 • Miaomiao Liu, Mathieu Salzmann, Xuming He
To this end, we first study the problem of depth estimation from a single image.
no code implementations • CVPR 2016 • Mohammad Najafi, Sarah Taghavi Namin, Mathieu Salzmann, Lars Petersson
By contrast, nonparametric approaches, which bypass any learning phase and directly transfer the labels from the training data to the query images, can readily exploit new labeled samples as they become available.
1 code implementation • ICCV 2015 • Pan Ji, Mathieu Salzmann, Hongdong Li
The Shape Interaction Matrix (SIM) is one of the earliest approaches to performing subspace clustering (i. e., separating points drawn from a union of subspaces).
Ranked #2 on
Motion Segmentation
on Hopkins155
no code implementations • ICCV 2015 • Mehrtash Harandi, Mathieu Salzmann, Mahsa Baktashmotlagh
State-of-the-art image-set matching techniques typically implicitly model each image-set with a Gaussian distribution.
no code implementations • CVPR 2016 • Mehrtash Harandi, Mathieu Salzmann, Fatih Porikli
Vectors of Locally Aggregated Descriptors (VLAD) have emerged as powerful image/video representations that compete with or even outperform state-of-the-art approaches on many challenging visual recognition tasks.
no code implementations • CVPR 2015 • Mehrtash Harandi, Mathieu Salzmann
While sparse coding on non-flat Riemannian manifolds has recently become increasingly popular, existing solutions either are dedicated to specific manifolds, or rely on optimization problems that are difficult to solve, especially when it comes to dictionary learning.
no code implementations • CVPR 2015 • Wei Zhuo, Mathieu Salzmann, Xuming He, Miaomiao Liu
We tackle the problem of single image depth estimation, which, without additional knowledge, suffers from many ambiguities.
no code implementations • CVPR 2013 • Sadeep Jayasumana, Richard Hartley, Mathieu Salzmann, Hongdong Li, Mehrtash Harandi
To encode the geometry of the manifold in the mapping, we introduce a family of provably positive definite kernels on the Riemannian manifold of SPD matrices.
no code implementations • CVPR 2014 • Sadeep Jayasumana, Richard Hartley, Mathieu Salzmann, Hongdong Li, Mehrtash Harandi
We tackle the problem of optimizing over all possible positive definite radial kernels on Riemannian manifolds for classification.
no code implementations • 13 Dec 2014 • Sadeep Jayasumana, Mathieu Salzmann, Hongdong Li, Mehrtash Harandi
We propose a framework for 2D shape analysis using positive definite kernels defined on Kendall's shape manifold.
no code implementations • 30 Nov 2014 • Sadeep Jayasumana, Richard Hartley, Mathieu Salzmann, Hongdong Li, Mehrtash Harandi
We then use the proposed framework to identify positive definite kernels on two specific manifolds commonly encountered in computer vision: the Riemannian manifold of symmetric positive definite matrices and the Grassmann manifold, i. e., the Riemannian manifold of linear subspaces of a Euclidean space.
no code implementations • CVPR 2015 • Thalaiyasingam Ajanthan, Richard Hartley, Mathieu Salzmann, Hongdong Li
While widely acknowledged as highly effective in computer vision, multi-label MRFs with non-convex priors are difficult to optimize.
no code implementations • 30 Aug 2014 • Mehrtash Harandi, Mathieu Salzmann
In contrast, here, we study the problem of performing coding in a high-dimensional Hilbert space, where the classes are expected to be more easily separable.
no code implementations • 4 Jul 2014 • Mehrtash T. Harandi, Mathieu Salzmann, Richard Hartley
In particular, we search for a projection that yields a low-dimensional manifold with maximum discriminative power encoded via an affinity-weighted similarity measure based on metrics on the manifold.
no code implementations • 4 Jul 2014 • Mehrtash T. Harandi, Mathieu Salzmann, Sadeep Jayasumana, Richard Hartley, Hongdong Li
Modeling videos and image-sets as linear subspaces has proven beneficial for many visual recognition tasks.
no code implementations • CVPR 2014 • Mahsa Baktashmotlagh, Mehrtash T. Harandi, Brian C. Lovell, Mathieu Salzmann
Here, we propose to make better use of the structure of this manifold and rely on the distance on the manifold to compare the source and target distributions.
no code implementations • CVPR 2014 • Miaomiao Liu, Mathieu Salzmann, Xuming He
The unary potentials in this graphical model are computed by making use of the images with known depth.
no code implementations • CVPR 2014 • Mehrtash Harandi, Mathieu Salzmann, Fatih Porikli
We introduce an approach to computing and comparing Covariance Descriptors (CovDs) in infinite-dimensional spaces.
no code implementations • CVPR 2013 • Miaomiao Liu, Richard Hartley, Mathieu Salzmann
In such conditions, our differential geometry analysis provides a theoretical proof that the shape of the mirror surface can be uniquely recovered if the pose of the reference target is known.
no code implementations • CVPR 2013 • Mathieu Salzmann
In this paper, we tackle the problem of performing inference in graphical models whose energy is a polynomial function of continuous variables.
no code implementations • NeurIPS 2010 • Mathieu Salzmann, Raquel Urtasun
Estimating 3D pose from monocular images is a highly ambiguous problem.
no code implementations • NeurIPS 2010 • Yangqing Jia, Mathieu Salzmann, Trevor Darrell
Recent approaches to multi-view learning have shown that factorizing the information into parts that are shared across all views and parts that are private to each view could effectively account for the dependencies and independencies between the different input modalities.