1 code implementation • ECCV 2020 • Luca Cavalli, Viktor Larsson, Martin Ralf Oswald, Torsten Sattler, Marc Pollefeys
As a result, outlier detection is a fundamental problem in computer vision and a wide range of approaches, from simple checks based on descriptor similarity to geometric verification, have been proposed over the last decades.
no code implementations • ECCV 2020 • Viktor Larsson, Nicolas Zobernig, Kasim Taskin, Marc Pollefeys
In this paper we consider the problem of Structure-from-Motion from images with unknown intrinsic calibration.
no code implementations • ECCV 2020 • Marcel Geppert, Viktor Larsson, Pablo Speciale, Johannes L. Schönberger, Marc Pollefeys
The recent trend towards cloud-based localization and mapping systems has raised significant privacy concerns.
1 code implementation • 10 Sep 2024 • Alexander Veicht, Paul-Edouard Sarlin, Philipp Lindenberger, Marc Pollefeys
This single-image calibration can benefit various downstream applications like image editing and 3D mapping.
no code implementations • 3 Sep 2024 • Yiming Zhao, Taein Kwon, Paul Streli, Marc Pollefeys, Christian Holz
Estimating touch contact and pressure in egocentric vision is a central task for downstream applications in Augmented Reality, Virtual Reality, as well as many robotic applications, because it provides precise physical insights into hand-object interaction and object manipulation.
no code implementations • 29 Aug 2024 • Emilia Szymanska, Mihai Dusmanu, Jan-Willem Buurlage, Mahdi Rad, Marc Pollefeys
Answering questions about the spatial properties of the environment poses challenges for existing language and vision foundation models due to a lack of understanding of the 3D world notably in terms of relationships between objects.
no code implementations • 29 Aug 2024 • Mathias Vogel, Keisuke Tateno, Marc Pollefeys, Federico Tombari, Marie-Julie Rakotosaona, Francis Engelmann
In this work, we tackle the task of point cloud denoising through a novel framework that adapts Diffusion Schr\"odinger bridges to points clouds.
no code implementations • 29 Aug 2024 • Linyan Yang, Lukas Hoyer, Mark Weber, Tobias Fischer, Dengxin Dai, Laura Leal-Taixé, Marc Pollefeys, Daniel Cremers, Luc van Gool
Unsupervised Domain Adaptation (UDA) is the task of bridging the domain gap between a labeled source domain, e. g., synthetic data, and an unlabeled target domain.
no code implementations • 27 Aug 2024 • Fangjinhua Wang, Qingtian Zhu, Di Chang, Quankai Gao, Junlin Han, Tong Zhang, Richard Hartley, Marc Pollefeys
3D reconstruction aims to recover the dense 3D structure of a scene.
1 code implementation • 29 Jul 2024 • Linfei Pan, Dániel Baráth, Marc Pollefeys, Johannes L. Schönberger
Recovering 3D structure and camera motion from images has been a long-standing focus of computer vision research and is known as Structure-from-Motion (SfM).
1 code implementation • 22 Jul 2024 • Luca Di Giammarino, Boyang Sun, Giorgio Grisetti, Marc Pollefeys, Hermann Blum, Daniel Barath
Our contributions involve using a data-driven approach with a simple architecture designed for real-time operation, a self-supervised data training method, and the capability to consistently integrate our map into a planning framework tailored for real-world robotics applications.
1 code implementation • 18 Jul 2024 • Lukas Bösiger, Mihai Dusmanu, Marc Pollefeys, Zuria Bauer
Rendering realistic images from 3D reconstruction is an essential task of many Computer Vision and Robotics pipelines, notably for mixed-reality applications as well as training autonomous agents in simulated environments.
1 code implementation • 16 Jul 2024 • Shinjeong Kim, Marc Pollefeys, Daniel Barath
This work addresses the challenge of sub-pixel accuracy in detecting 2D local features, a cornerstone problem in computer vision.
1 code implementation • 11 Jul 2024 • Jonas Kulhanek, Songyou Peng, Zuzana Kukelova, Marc Pollefeys, Torsten Sattler
While the field of 3D scene reconstruction is dominated by NeRFs due to their photorealistic quality, 3D Gaussian Splatting (3DGS) has recently emerged, offering similar quality with real-time rendering speeds.
no code implementations • 28 Jun 2024 • Daiwei Zhang, Gengyan Li, Jiajie Li, Mickaël Bressieux, Otmar Hilliges, Marc Pollefeys, Luc van Gool, Xi Wang
Human activities are inherently complex, and even simple household tasks involve numerous object interactions.
1 code implementation • CVPR 2024 • Fangjinhua Wang, Xudong Jiang, Silvano Galliani, Christoph Vogel, Marc Pollefeys
We propose GLACE, which integrates pre-trained global and local encodings and enables SCR to scale to large scenes with only a single small-sized network.
no code implementations • 5 Jun 2024 • Tobias Fischer, Jonas Kulhanek, Samuel Rota Bulò, Lorenzo Porzi, Marc Pollefeys, Peter Kontschieder
We present an efficient neural 3D scene representation for novel-view synthesis (NVS) in large-scale, dynamic urban areas.
no code implementations • 30 May 2024 • Gonca Yilmaz, Songyou Peng, Francis Engelmann, Marc Pollefeys, Hermann Blum
We, therefore, introduce a new task domain adaptation for open-vocabulary segmentation, enhancing VLMs with domain-specific priors while preserving their open-vocabulary nature.
3D Open-Vocabulary Instance Segmentation Domain Adaptation +2
1 code implementation • CVPR 2024 • Weining Ren, Zihan Zhu, Boyang Sun, Jiaqi Chen, Marc Pollefeys, Songyou Peng
Neural Radiance Fields (NeRFs) have shown remarkable success in synthesizing photorealistic views from multi-view images of static scenes, but face challenges in dynamic, real-world environments with distractors like moving objects, shadows, and lighting changes.
no code implementations • CVPR 2024 • Lei LI, Songyou Peng, Zehao Yu, Shaohui Liu, Rémi Pautrat, Xiaochuan Yin, Marc Pollefeys
Real-world objects and environments are predominantly composed of edge features, including straight lines and curves.
1 code implementation • 16 May 2024 • Xianzheng Ma, Yash Bhalgat, Brandon Smart, Shuai Chen, Xinghui Li, Jian Ding, Jindong Gu, Dave Zhenyu Chen, Songyou Peng, Jia-Wang Bian, Philip H Torr, Marc Pollefeys, Matthias Nießner, Ian D Reid, Angel X. Chang, Iro Laina, Victor Adrian Prisacariu
Hence, with this paper, we aim to chart a course for future research that explores and expands the capabilities of 3D-LLMs in understanding and interacting with the complex 3D world.
no code implementations • 2 May 2024 • Guangming Wang, Lei Pan, Songyou Peng, Shaohui Liu, Chenfeng Xu, Yanzi Miao, Wei Zhan, Masayoshi Tomizuka, Marc Pollefeys, Hesheng Wang
Meticulous 3D environment representations have been a longstanding goal in computer vision and robotics fields.
1 code implementation • 27 Apr 2024 • Rong Zou, Marc Pollefeys, Denys Rozumnyi
We propose a method for object retrieval in images that are affected by motion blur.
1 code implementation • CVPR 2024 • Petr Hruby, Timothy Duff, Marc Pollefeys
We revisit certain problems of pose estimation based on 3D--2D correspondences between features which may be points or lines.
no code implementations • 22 Apr 2024 • Jiaqi Chen, Daniel Barath, Iro Armeni, Marc Pollefeys, Hermann Blum
As such, we need methods that interface between natural language and map representations of the environment.
no code implementations • 18 Apr 2024 • Oliver Lemke, Zuria Bauer, René Zurbrügg, Marc Pollefeys, Francis Engelmann, Hermann Blum
This allows for accurate detection directly in 3D scenes, object- and environment-aware grasp prediction, as well as robust and repeatable robotic manipulation.
no code implementations • 4 Apr 2024 • Francis Engelmann, Fabian Manhardt, Michael Niemeyer, Keisuke Tateno, Marc Pollefeys, Federico Tombari
Our OpenNeRF further leverages NeRF's ability to render novel views and extract open-set VLM features from areas that are not well observed in the initial posed images.
1 code implementation • CVPR 2024 • Rui Li, Tobias Fischer, Mattia Segu, Marc Pollefeys, Luc van Gool, Federico Tombari
We propose KYN, a novel method for single-view scene reconstruction that reasons about semantic and spatial context to predict each point's density.
no code implementations • CVPR 2024 • Chong Bao, yinda zhang, Yuan Li, Xiyu Zhang, Bangbang Yang, Hujun Bao, Marc Pollefeys, Guofeng Zhang, Zhaopeng Cui
Recently, we have witnessed the explosive growth of various volumetric representations in modeling animatable head avatars.
no code implementations • 30 Mar 2024 • Yang Miao, Francis Engelmann, Olga Vysotska, Federico Tombari, Marc Pollefeys, Dániel Béla Baráth
We introduce a novel problem, i. e., the localization of an input image within a multi-modal reference map represented by a database of 3D scene graphs.
no code implementations • 25 Mar 2024 • Jonas Hein, Frédéric Giraud, Lilian Calvet, Alexander Schwarz, Nicola Alessandro Cavalcanti, Sergey Prokudin, Mazda Farshad, Siyu Tang, Marc Pollefeys, Fabio Carrillo, Philipp Fürnstahl
Surgery digitalization is the process of creating a virtual replica of real-world surgery, also referred to as a surgical digital twin (SDT).
1 code implementation • 22 Mar 2024 • Nicolas Baumann, Michael Baumgartner, Edoardo Ghignone, Jonas Kühne, Tobias Fischer, Yung-Hsu Yang, Marc Pollefeys, Michele Magno
To enable self-driving vehicles accurate detection and tracking of surrounding objects is essential.
1 code implementation • 21 Mar 2024 • Yuedong Chen, Haofei Xu, Chuanxia Zheng, Bohan Zhuang, Marc Pollefeys, Andreas Geiger, Tat-Jen Cham, Jianfei Cai
We introduce MVSplat, an efficient model that, given sparse multi-view images as input, predicts clean feed-forward 3D Gaussians.
Ranked #1 on Generalizable Novel View Synthesis on ACID
no code implementations • 5 Mar 2024 • Changan Chen, Rui Wang, Christoph Vogel, Marc Pollefeys
In this paper we propose an efficient data-driven solution to self-localization within a floorplan.
no code implementations • 23 Feb 2024 • Francis Engelmann, Ayca Takmaz, Jonas Schult, Elisabetta Fedele, Johanna Wald, Songyou Peng, Xi Wang, Or Litany, Siyu Tang, Federico Tombari, Marc Pollefeys, Leonidas Guibas, Hongbo Tian, Chunjie Wang, Xiaosheng Yan, Bingwen Wang, Xuanyang Zhang, Xiao Liu, Phuc Nguyen, Khoi Nguyen, Anh Tran, Cuong Pham, Zhening Huang, Xiaoyang Wu, Xi Chen, Hengshuang Zhao, Lei Zhu, Joan Lasenby
This report provides an overview of the challenge hosted at the OpenSUN3D Workshop on Open-Vocabulary 3D Scene Understanding held in conjunction with ICCV 2023.
no code implementations • CVPR 2024 • Zuoyue Li, Zhenqiang Li, Zhaopeng Cui, Marc Pollefeys, Martin R. Oswald
Directly generating scenes from satellite imagery offers exciting possibilities for integration into applications like games and map services.
no code implementations • CVPR 2024 • Gen Li, Kaifeng Zhao, Siwei Zhang, Xiaozhong Lyu, Mihai Dusmanu, Yan Zhang, Marc Pollefeys, Siyu Tang
To address this challenge, we introduce EgoGen, a new synthetic data generator that can produce accurate and rich ground-truth training data for egocentric perception tasks.
no code implementations • 8 Jan 2024 • Casimir Feldmann, Niall Siegenheim, Nikolas Hars, Lovro Rabuzin, Mert Ertugrul, Luca Wolfart, Marc Pollefeys, Zuria Bauer, Martin R. Oswald
In the case of MDE models for autonomous driving, this issue is exacerbated by the linearity of the captured data trajectories.
no code implementations • CVPR 2024 • Weirong Chen, Le Chen, Rui Wang, Marc Pollefeys
Visual odometry estimates the motion of a moving camera based on visual input.
no code implementations • CVPR 2024 • Alexandros Delitzas, Ayca Takmaz, Federico Tombari, Robert Sumner, Marc Pollefeys, Francis Engelmann
Existing 3D scene understanding methods are heavily focused on 3D semantic and instance segmentation.
no code implementations • CVPR 2024 • Changan Chen, Rui Wang, Christoph Vogel, Marc Pollefeys
In this paper we propose an efficient data-driven solution to self-localization within a floorplan.
no code implementations • CVPR 2024 • Tobias Fischer, Lorenzo Porzi, Samuel Rota Bulo, Marc Pollefeys, Peter Kontschieder
To this end we present a novel decomposable radiance field approach for dynamic urban environments.
1 code implementation • CVPR 2024 • Shengze Jin, Iro Armeni, Marc Pollefeys, Daniel Barath
We introduce a novel framework for multiway point cloud mosaicking (named Wednesday) designed to co-align sets of partially overlapping point clouds -- typically obtained from 3D scanners or moving RGB-D cameras -- into a unified coordinate system.
no code implementations • 28 Dec 2023 • Rui Huang, Songyou Peng, Ayca Takmaz, Federico Tombari, Marc Pollefeys, Shiji Song, Gao Huang, Francis Engelmann
Therefore, we explore the use of image segmentation foundation models to automatically generate training labels for 3D segmentation.
no code implementations • 20 Dec 2023 • Fangjinhua Wang, Marie-Julie Rakotosaona, Michael Niemeyer, Richard Szeliski, Marc Pollefeys, Federico Tombari
In this work, we propose UniSDF, a general purpose 3D reconstruction method that can reconstruct large complex scenes with reflections.
1 code implementation • CVPR 2024 • Haofei Xu, Anpei Chen, Yuedong Chen, Christos Sakaridis, Yulun Zhang, Marc Pollefeys, Andreas Geiger, Fisher Yu
We present Multi-Baseline Radiance Fields (MuRF), a general feed-forward approach to solving sparse view synthesis under multiple different baseline settings (small and large baselines, and different number of input views).
no code implementations • 29 Nov 2023 • Silvan Weder, Francis Engelmann, Johannes L. Schönberger, Akihito Seki, Marc Pollefeys, Martin R. Oswald
Using these main contributions, our method can enable scenarios with real-time constraints and can scale to arbitrary scene sizes by processing and updating the scene only in a local region defined by the new measurement.
no code implementations • 29 Nov 2023 • Yu Zheng, Guangming Wang, Jiuming Liu, Marc Pollefeys, Hesheng Wang
Through the hash-based representation, we propose the Spherical Frustum sparse Convolution (SFC) and Frustum Fast Point Sampling (F2PS) to convolve and sample the points stored in spherical frustums respectively.
no code implementations • 20 Nov 2023 • Silvan Weder, Hermann Blum, Francis Engelmann, Marc Pollefeys
Semantic annotations are indispensable to train or evaluate perception models, yet very costly to acquire.
Ranked #1 on Semantic Segmentation on Replica
no code implementations • 15 Nov 2023 • Tao Sun, Yan Hao, Shengyu Huang, Silvio Savarese, Konrad Schindler, Marc Pollefeys, Iro Armeni
To this end, we introduce the Nothing Stands Still (NSS) benchmark, which focuses on the spatiotemporal registration of 3D scenes undergoing large spatial and temporal change, ultimately creating one coherent spatiotemporal map.
no code implementations • 6 Nov 2023 • Zador Pataki, Mohammad Altillawi, Menelaos Kanakis, Rémi Pautrat, Fengyi Shen, Ziyuan Liu, Luc van Gool, Marc Pollefeys
Our proposed method enhances cross-domain localization performance, significantly reducing the performance gap.
no code implementations • 10 Oct 2023 • Le Chen, Weirong Chen, Rui Wang, Marc Pollefeys
As a promising fashion for visual localization, scene coordinate regression (SCR) has seen tremendous progress in the past decade.
no code implementations • 8 Oct 2023 • Dominik Hollidt, Clinton Wang, Polina Golland, Marc Pollefeys
We present a novel approach to perform 3D semantic segmentation solely from 2D supervision by leveraging Neural Radiance Fields (NeRFs).
no code implementations • 4 Oct 2023 • Matthew Hanlon, Boyang Sun, Marc Pollefeys, Hermann Blum
However, localizing e. g. a ground robot in the map of a drone or head-mounted MR headset presents unique challenges due to viewpoint changes.
no code implementations • ICCV 2023 • Xin Wang, Taein Kwon, Mahdi Rad, Bowen Pan, Ishani Chakraborty, Sean Andrist, Dan Bohus, Ashley Feniello, Bugra Tekin, Felipe Vieira Frujeri, Neel Joshi, Marc Pollefeys
Building an interactive AI assistant that can perceive, reason, and collaborate with humans in the real world has been a long-standing pursuit in the AI community.
no code implementations • 27 Sep 2023 • Shengze Jin, Daniel Barath, Marc Pollefeys, Iro Armeni
Point cloud registration has seen recent success with several learning-based methods that focus on correspondence matching and, as such, optimize only for this objective.
no code implementations • 27 Sep 2023 • Petr Hruby, Shaohui Liu, Rémi Pautrat, Marc Pollefeys, Daniel Barath
We propose an approach for estimating the relative pose between calibrated image pairs by jointly exploiting points, lines, and their coincidences in a hybrid manner.
1 code implementation • 26 Sep 2023 • Yang Miao, Iro Armeni, Marc Pollefeys, Daniel Barath
We introduce an online 2D-to-3D semantic instance mapping algorithm aimed at generating comprehensive, accurate, and efficient semantic 3D maps suitable for autonomous agents in unstructured environments.
no code implementations • 17 Sep 2023 • Junan Lin, Zhichao Sun, Enjie Cao, Taein Kwon, Mahdi Rad, Marc Pollefeys
Skeletal Action recognition from an egocentric view is important for applications such as interfaces in AR/VR glasses and human-robot interaction, where the device has limited resources.
no code implementations • 12 Sep 2023 • Yao Feng, Weiyang Liu, Timo Bolkart, Jinlong Yang, Marc Pollefeys, Michael J. Black
Towards this end, both explicit and implicit 3D representations are heavily studied for a holistic modeling and capture of the whole human (e. g., body, clothing, face and hair), but neither representation is an optimal choice in terms of representation efficacy since different parts of the human avatar have different modeling desiderata.
1 code implementation • 6 Sep 2023 • Marko Mihajlovic, Sergey Prokudin, Marc Pollefeys, Siyu Tang
Neural fields, a category of neural networks trained to represent high-frequency signals, have gained significant attention in recent years due to their impressive performance in modeling complex 3D data, such as signed distance (SDFs) or radiance fields (NeRFs), via a single multi-layer perceptron (MLP).
no code implementations • ICCV 2023 • Aron Schmied, Tobias Fischer, Martin Danelljan, Marc Pollefeys, Fisher Yu
We propose R3D3, a multi-camera system for dense 3D reconstruction and ego-motion estimation.
1 code implementation • ICCV 2023 • Rémi Pautrat, Shaohui Liu, Petr Hruby, Marc Pollefeys, Daniel Barath
We tackle the problem of estimating a Manhattan frame, i. e. three orthogonal vanishing points, and the unknown focal length of the camera, leveraging a prior vertical direction.
1 code implementation • ICCV 2023 • Chang Nie, Guangming Wang, Zhe Liu, Luca Cavalli, Marc Pollefeys, Hesheng Wang
Therefore, RLSAC can avoid differentiating to learn the features and the feedback of downstream tasks for end-to-end robust estimation.
no code implementations • 5 Aug 2023 • Florentin Liebmann, Marco von Atzigen, Dominik Stütz, Julian Wolf, Lukas Zingg, Daniel Suter, Laura Leoty, Hooman Esfandiari, Jess G. Snedeker, Martin R. Oswald, Marc Pollefeys, Mazda Farshad, Philipp Fürnstahl
An intuitive surgical guidance is provided thanks to the integration into an augmented reality based navigation system.
no code implementations • 3 Aug 2023 • Katarína Tóthová, Ľubor Ladický, Daniel Thul, Marc Pollefeys, Ender Konukoglu
Predictive variability due to data ambiguities has typically been addressed via construction of dedicated models with built-in probabilistic capabilities that are trained to predict uncertainty estimates as variables of interest.
no code implementations • 28 Jul 2023 • Daniel Barath, Dmytro Mishkin, Luca Cavalli, Paul-Edouard Sarlin, Petr Hruby, Marc Pollefeys
Moreover, we derive a new minimal solver for homography estimation, requiring only a single affine correspondence (AC) and a gravity prior.
1 code implementation • 26 Jul 2023 • Luca Cavalli, Daniel Barath, Marc Pollefeys, Viktor Larsson
The proposed attention mechanism and one-step transformer provide an adaptive behavior that enhances the performance of RANSAC, making it a more effective tool for robust estimation.
no code implementations • 19 Jul 2023 • Siyan Dong, Shaohui Liu, Hengkai Guo, Baoquan Chen, Marc Pollefeys
Visual (re)localization is critical for various applications in computer vision and robotics.
1 code implementation • 29 Jun 2023 • David Recasens, Martin R. Oswald, Marc Pollefeys, Javier Civera
Estimating camera motion in deformable scenes poses a complex and open research challenge.
2 code implementations • ICCV 2023 • Philipp Lindenberger, Paul-Edouard Sarlin, Marc Pollefeys
We introduce LightGlue, a deep neural network that learns to match local features across images.
1 code implementation • NeurIPS 2023 • Paul-Edouard Sarlin, Eduard Trulls, Marc Pollefeys, Jan Hosang, Simon Lynen
Semantic 2D maps are commonly used by humans and machines for navigation purposes, whether it's walking or driving.
no code implementations • 5 May 2023 • Jonas Hein, Nicola Cavalcanti, Daniel Suter, Lukas Zingg, Fabio Carrillo, Lilian Calvet, Mazda Farshad, Marc Pollefeys, Nassir Navab, Philipp Fürnstahl
Third, we evaluate three state-of-the-art single-view and multi-view methods for the task of 6DoF pose estimation of surgical instruments and analyze the influence of camera configurations, training data, and occlusions on the pose accuracy and generalization ability.
no code implementations • 3 May 2023 • Cathrin Elich, Iro Armeni, Martin R. Oswald, Marc Pollefeys, Joerg Stueckler
Our approach compares favorably to previous state-of-the-art object-level matching approaches and achieves improved performance over a pure keypoint-based approach for large view-point changes.
1 code implementation • 28 Apr 2023 • Sayan Deb Sarkar, Ondrej Miksik, Marc Pollefeys, Daniel Barath, Iro Armeni
We propose SGAligner, the first method for aligning pairs of 3D scene graphs that is robust to in-the-wild scenarios (ie, unknown overlap -- if any -- and changes in the environment).
Ranked #1 on Point Cloud Registration on 3RScan
no code implementations • ICCV 2023 • Denys Rozumnyi, Jiri Matas, Marc Pollefeys, Vittorio Ferrari, Martin R. Oswald
We argue that this representation is limited and instead propose to guide and improve 2D tracking with an explicit object representation, namely the textured 3D shape and 6DoF pose in each video frame.
2 code implementations • ICCV 2023 • Rémi Pautrat, Iago Suárez, Yifan Yu, Marc Pollefeys, Viktor Larsson
Line segments are powerful features complementary to points.
no code implementations • ICCV 2023 • Yiming Zhao, Denys Rozumnyi, Jie Song, Otmar Hilliges, Marc Pollefeys, Martin R. Oswald
The key idea is to tackle the inverse problem of image deblurring by modeling the forward problem with a 3D human model, a texture map, and a sequence of poses to describe human motion.
1 code implementation • CVPR 2023 • Shaohui Liu, Yifan Yu, Rémi Pautrat, Marc Pollefeys, Viktor Larsson
In contrast to sparse keypoints, a handful of line segments can concisely encode the high-level scene layout, as they often delineate the main structural elements.
1 code implementation • ICCV 2023 • Jiuming Liu, Guangming Wang, Zhe Liu, Chaokang Jiang, Marc Pollefeys, Hesheng Wang
Specifically, a projection-aware hierarchical transformer is proposed to capture long-range dependencies and filter outliers by extracting point features globally.
no code implementations • 7 Feb 2023 • Zihan Zhu, Songyou Peng, Viktor Larsson, Zhaopeng Cui, Martin R. Oswald, Andreas Geiger, Marc Pollefeys
Neural implicit representations have recently become popular in simultaneous localization and mapping (SLAM), especially in dense visual SLAM.
no code implementations • ICCV 2023 • Linfei Pan, Johannes L. Schönberger, Viktor Larsson, Marc Pollefeys
Recent methods on privacy-preserving image-based localization use a random line parameterization to protect the privacy of query images and database maps.
no code implementations • ICCV 2023 • Sayan Deb Sarkar, Ondrej Miksik, Marc Pollefeys, Daniel Barath, Iro Armeni
We propose SGAligner, the first method for aligning pairs of 3D scene graphs that is robust to in-the-wild scenarios (i. e., unknown overlap - if any - and changes in the environment).
no code implementations • CVPR 2023 • Petr Hruby, Viktor Korotynskiy, Timothy Duff, Luke Oeding, Marc Pollefeys, Tomas Pajdla, Viktor Larsson
The minimal case for reconstruction requires 13 points in 4 views for both the calibrated and uncalibrated cameras.
no code implementations • ICCV 2023 • Shuzhe Wang, Juho Kannala, Marc Pollefeys, Daniel Barath
We propose a new method, named curvature similarity extractor (CSE), for improving local feature matching across images.
no code implementations • CVPR 2023 • Silvan Weder, Guillermo Garcia-Hernando, Aron Monszpart, Marc Pollefeys, Gabriel Brostow, Michael Firman, Sara Vicente
We validate our approach using a new and still-challenging dataset for the task of NeRF inpainting.
1 code implementation • CVPR 2023 • Rémi Pautrat, Daniel Barath, Viktor Larsson, Martin R. Oswald, Marc Pollefeys
Their learned counterparts are more repeatable and can handle challenging images, but at the cost of a lower accuracy and a bias towards wireframe lines.
1 code implementation • CVPR 2023 • Yufan Ren, Fangjinhua Wang, Tong Zhang, Marc Pollefeys, Sabine Süsstrunk
The success of the Neural Radiance Fields (NeRF) in novel view synthesis has inspired researchers to propose neural implicit scene reconstruction.
1 code implementation • CVPR 2023 • Songyou Peng, Kyle Genova, Chiyu "Max" Jiang, Andrea Tagliasacchi, Marc Pollefeys, Thomas Funkhouser
Traditional 3D scene understanding approaches rely on labeled 3D datasets to train a model for a single task with supervision.
Ranked #7 on 3D Open-Vocabulary Instance Segmentation on Replica
3D Open-Vocabulary Instance Segmentation 3D Semantic Segmentation +1
1 code implementation • 19 Oct 2022 • Paul-Edouard Sarlin, Mihai Dusmanu, Johannes L. Schönberger, Pablo Speciale, Lukas Gruber, Viktor Larsson, Ondrej Miksik, Marc Pollefeys
To close this gap, we introduce LaMAR, a new benchmark with a comprehensive capture and GT pipeline that co-registers realistic trajectories and sensor streams captured by heterogeneous AR devices in large, unconstrained scenes.
no code implementations • 5 Oct 2022 • Mathias Vetsch, Sandro Lombardi, Marc Pollefeys, Martin R. Oswald
The generation of triangle meshes from point clouds, i. e. meshing, is a core task in computer graphics and computer vision.
1 code implementation • 4 Oct 2022 • Yao Feng, Jinlong Yang, Marc Pollefeys, Michael J. Black, Timo Bolkart
Building on this insight, we propose SCARF (Segmented Clothed Avatar Radiance Field), a hybrid model combining a mesh-based body with a neural radiance field.
1 code implementation • ICCV 2023 • Weicai Ye, Shuo Chen, Chong Bao, Hujun Bao, Marc Pollefeys, Zhaopeng Cui, Guofeng Zhang
Existing inverse rendering combined with neural rendering methods can only perform editable novel view synthesis on object-specific scenes, while we present intrinsic neural radiance fields, dubbed IntrinsicNeRF, which introduce intrinsic decomposition into the NeRF-based neural rendering method and can extend its application to room-scale scenes.
1 code implementation • 27 Sep 2022 • Hao Dong, Xieyuanli Chen, Mihai Dusmanu, Viktor Larsson, Marc Pollefeys, Cyrill Stachniss
A distinctive representation of image patches in form of features is a key component of many computer vision and robotics tasks, such as image matching, image retrieval, and visual localization.
1 code implementation • 7 Sep 2022 • Lei LI, Zhizheng Liu, Weining Ren, Liudi Yang, Fangjinhua Wang, Marc Pollefeys, Songyou Peng
3D textured shape recovery from partial scans is crucial for many real-world applications.
1 code implementation • 14 Aug 2022 • Siyan Dong, Shuzhe Wang, Yixin Zhuang, Juho Kannala, Marc Pollefeys, Baoquan Chen
Visual (re)localization addresses the problem of estimating the 6-DoF (Degree of Freedom) camera pose of a query image captured in a known scene, which is a key building block of many computer vision and robotics applications.
no code implementations • 23 Jul 2022 • Zuoyue Li, Tianxing Fan, Zhenqiang Li, Zhaopeng Cui, Yoichi Sato, Marc Pollefeys, Martin R. Oswald
We introduce a scalable framework for novel view synthesis from RGB-D images with largely incomplete scene coverage.
1 code implementation • 16 Jul 2022 • Luca Cavalli, Marc Pollefeys, Daniel Barath
We tested NeFSAC on more than 100k image pairs from three publicly available real-world datasets and found that it leads to one order of magnitude speed-up, while often finding more accurate results than USAC alone.
1 code implementation • CVPR 2022 • Taein Kwon, Bugra Tekin, Siyu Tang, Marc Pollefeys
Temporal alignment of fine-grained human actions in videos is important for numerous applications in computer vision, robotics, and mixed reality.
no code implementations • 3 Feb 2022 • Jeffrey Delmerico, Roi Poranne, Federica Bogo, Helen Oleynikova, Eric Vollenweider, Stelian Coros, Juan Nieto, Marc Pollefeys
Spatial computing -- the ability of devices to be aware of their surroundings and to represent this digitally -- offers novel capabilities in human-robot interaction.
no code implementations • CVPR 2022 • Daniel Barath, Luca Cavalli, Marc Pollefeys
We propose the Model Quality Network, MQ-Net in short, for predicting the quality, e. g. the pose error of essential matrices, of models generated inside RANSAC.
no code implementations • CVPR 2022 • Marcel Geppert, Viktor Larsson, Johannes L. Schönberger, Marc Pollefeys
We propose a principled approach overcoming these limitations, based on two observations.
no code implementations • CVPR 2022 • Linfei Pan, Marc Pollefeys, Viktor Larsson
Low-dimensional parametric models are the de-facto standard in computer vision for intrinsic camera calibration.
1 code implementation • CVPR 2022 • Zihan Zhu, Songyou Peng, Viktor Larsson, Weiwei Xu, Hujun Bao, Zhaopeng Cui, Martin R. Oswald, Marc Pollefeys
Neural implicit representations have recently shown encouraging results in various domains, including promising progress in simultaneous localization and mapping (SLAM).
no code implementations • 22 Dec 2021 • Zuria Bauer, Zuoyue Li, Sergio Orts-Escolano, Miguel Cazorla, Marc Pollefeys, Martin R. Oswald
Building upon the recent progress in novel view synthesis, we propose its application to improve monocular depth estimation.
Ranked #29 on Monocular Depth Estimation on KITTI Eigen split
1 code implementation • 14 Dec 2021 • Siwei Zhang, Qianli Ma, Yan Zhang, Zhiyin Qian, Taein Kwon, Marc Pollefeys, Federica Bogo, Siyu Tang
Key to reasoning about interactions is to understand the body pose and motion of the interaction partner from the egocentric view.
1 code implementation • CVPR 2022 • Fangjinhua Wang, Silvano Galliani, Christoph Vogel, Marc Pollefeys
We present IterMVS, a new data-driven method for high-resolution multi-view stereo.
no code implementations • 30 Nov 2021 • Sandro Lombardi, Bangbang Yang, Tianxing Fan, Hujun Bao, Guofeng Zhang, Marc Pollefeys, Zhaopeng Cui
In this work, we propose a novel neural implicit representation for the human body, which is fully differentiable and optimizable with disentangled shape and pose latent spaces.
1 code implementation • CVPR 2022 • Denys Rozumnyi, Martin R. Oswald, Vittorio Ferrari, Marc Pollefeys
We propose a method for jointly estimating the 3D motion, 3D shape, and appearance of highly motion-blurred objects from a video.
no code implementations • CVPR 2022 • Weizhe Liu, Bugra Tekin, Huseyin Coskun, Vibhav Vineet, Pascal Fua, Marc Pollefeys
To this end, we propose an approach to enforce temporal priors on the optimal transport matrix, which leverages temporal consistency, while allowing for variations in the order of actions.
1 code implementation • 26 Oct 2021 • George Chogovadze, Rémi Pautrat, Marc Pollefeys
At the heart of the success of deep learning is the quality of the data.
no code implementations • 13 Oct 2021 • Qingshan Xu, Martin R. Oswald, Wenbing Tao, Marc Pollefeys, Zhaopeng Cui
However, existing recurrent methods only model the local dependencies in the depth domain, which greatly limits the capability of capturing the global scene context along the depth dimension.
no code implementations • 9 Sep 2021 • Dimitri Zhukov, Ignacio Rocco, Ivan Laptev, Josef Sivic, Johannes L. Schönberger, Bugra Tekin, Marc Pollefeys
Contrary to the standard scenario of instance-level 3D reconstruction, where identical objects or scenes are present in all views, objects in different instructional videos may have large appearance variations given varying conditions and versions of the same product.
1 code implementation • ICCV 2021 • Siwei Zhang, Yan Zhang, Federica Bogo, Marc Pollefeys, Siyu Tang
To prove the effectiveness of the proposed motion priors, we combine them into a novel pipeline for 4D human body capture in 3D scenes.
2 code implementations • ICCV 2021 • Philipp Lindenberger, Paul-Edouard Sarlin, Viktor Larsson, Marc Pollefeys
Finding local features that are repeatable across multiple views is a cornerstone of sparse 3D reconstruction.
no code implementations • CVPR 2021 • Marcel Geppert, Viktor Larsson, Pablo Speciale, Johannes L. Schonberger, Marc Pollefeys
In this paper, we propose a solution to the uncalibrated privacy preserving localization and mapping problem.
1 code implementation • NeurIPS 2021 • Denys Rozumnyi, Martin R. Oswald, Vittorio Ferrari, Marc Pollefeys
We address the novel task of jointly reconstructing the 3D shape, texture, and motion of an object from a single motion-blurred image.
2 code implementations • NeurIPS 2021 • Songyou Peng, Chiyu "Max" Jiang, Yiyi Liao, Michael Niemeyer, Marc Pollefeys, Andreas Geiger
However, the implicit nature of neural implicit representations results in slow inference time and requires careful initialization.
no code implementations • ICCV 2021 • Taein Kwon, Bugra Tekin, Jan Stuhmer, Federica Bogo, Marc Pollefeys
To this end, we propose a method to create a unified dataset for egocentric 3D interaction recognition.
Ranked #8 on Action Recognition on H2O (2 Hands and Objects)
no code implementations • ICCV 2021 • Yawei Li, He Chen, Zhaopeng Cui, Radu Timofte, Marc Pollefeys, Gregory Chirikjian, Luc van Gool
In this paper, we aim at improving the computational efficiency of graph convolutional networks (GCNs) for learning on point clouds.
1 code implementation • CVPR 2021 • Rémi Pautrat, Juan-Ting Lin, Viktor Larsson, Martin R. Oswald, Marc Pollefeys
We thus hereby introduce the first joint detection and description of line segments in a single deep network.
no code implementations • ICCV 2021 • Peidong Liu, Xingxing Zuo, Viktor Larsson, Marc Pollefeys
Motion blur is one of the major challenges remaining for visual odometry methods.
2 code implementations • CVPR 2021 • Paul-Edouard Sarlin, Ajaykumar Unagar, Måns Larsson, Hugo Germain, Carl Toft, Viktor Larsson, Marc Pollefeys, Vincent Lepetit, Lars Hammarstrand, Fredrik Kahl, Torsten Sattler
In this paper, we go Back to the Feature: we argue that deep networks should focus on learning robust and invariant visual features, while the geometric estimation should be left to principled algorithms.
1 code implementation • CVPR 2021 • Cheng Zhang, Zhaopeng Cui, yinda zhang, Bing Zeng, Marc Pollefeys, Shuaicheng Liu
We not only propose an image-based local structured implicit network to improve the object shape estimation, but also refine the 3D object pose and scene layout via a novel implicit scene graph neural network that exploits the implicit local object features.
Ranked #1 on Monocular 3D Object Detection on SUN RGB-D (using extra training data)
1 code implementation • 6 Feb 2021 • Dalia El Badawy, Viktor Larsson, Marc Pollefeys, Ivan Dokmanić
We look at the general case where neither the emission times of the sources nor the reference time frames of the receivers are known.
no code implementations • ICCV 2021 • Viktor Larsson, Marc Pollefeys, Magnus Oskarsson
In this paper we consider the epipolar geometry between orthographic and perspective cameras.
no code implementations • 1 Jan 2021 • Yawei Li, He Chen, Zhaopeng Cui, Radu Timofte, Marc Pollefeys, Gregory Chirikjian, Luc van Gool
State-of-the-art GCNs adopt $K$-nearest neighbor (KNN) searches for local feature aggregation and feature extraction operations from layer to layer.
1 code implementation • CVPR 2021 • Marko Mihajlovic, Silvan Weder, Marc Pollefeys, Martin R. Oswald
We present DeepSurfels, a novel hybrid scene representation for geometry and appearance information.
no code implementations • 18 Dec 2020 • Xingxing Zuo, Nathaniel Merrill, Wei Li, Yong liu, Marc Pollefeys, Guoquan Huang
In this work, we present a lightweight, tightly-coupled deep depth network and visual-inertial odometry (VIO) system, which can provide accurate state estimates and dense depth maps of the immediate surroundings.
1 code implementation • ICCV 2021 • Denys Rozumnyi, Jiri Matas, Filip Sroubek, Marc Pollefeys, Martin R. Oswald
Compared to other methods, such as deblatting, the inference is of several orders of magnitude faster and allows applications such as real-time fast moving object detection and retrieval in large video collections.
no code implementations • ICCV 2021 • Zuoyue Li, Zhenqiang Li, Zhaopeng Cui, Rongjun Qin, Marc Pollefeys, Martin R. Oswald
For geometrical and temporal consistency, our approach explicitly creates a 3D point cloud representation of the scene and maintains dense 3D-2D correspondences across frames that reflect the geometric scene configuration inferred from the satellite view.
1 code implementation • CVPR 2021 • Arda Düzçeker, Silvano Galliani, Christoph Vogel, Pablo Speciale, Mihai Dusmanu, Marc Pollefeys
We propose an online multi-view depth prediction approach on posed video streams, where the scene geometry information computed in the previous time steps is propagated to the current time step in an efficient and geometrically plausible way.
1 code implementation • CVPR 2021 • Fangjinhua Wang, Silvano Galliani, Christoph Vogel, Pablo Speciale, Marc Pollefeys
We present PatchmatchNet, a novel and learnable cascade formulation of Patchmatch for high-resolution multi-view stereo.
Ranked #10 on Point Clouds on Tanks and Temples
1 code implementation • ICCV 2021 • Mihai Dusmanu, Ondrej Miksik, Johannes L. Schönberger, Marc Pollefeys
Visual localization and mapping is the key technology underlying the majority of mixed reality and robotics systems.
5 code implementations • CVPR 2021 • Denys Rozumnyi, Martin R. Oswald, Vittorio Ferrari, Jiri Matas, Marc Pollefeys
We propose a method that, given a single image with its estimated background, outputs the object's appearance and position in a series of sub-frames as if captured by a high-speed camera (i. e. temporal super-resolution).
Ranked #1 on Video Super-Resolution on Falling Objects
1 code implementation • CVPR 2021 • Silvan Weder, Johannes L. Schönberger, Marc Pollefeys, Martin R. Oswald
We present a novel online depth map fusion approach that learns depth map aggregation in a latent feature space.
no code implementations • 19 Oct 2020 • Alexander Millane, Helen Oleynikova, Christian Lanegger, Jeff Delmerico, Juan Nieto, Roland Siegwart, Marc Pollefeys, Cesar Cadena
Localization of a robotic system within a previously mapped environment is important for reducing estimation drift and for reusing previously built maps.
Robotics
no code implementations • 8 Oct 2020 • Cathrin Elich, Martin R. Oswald, Marc Pollefeys, Joerg Stueckler
Our approach learns to decompose images of synthetic scenes with multiple objects on a planar surface into its constituent scene objects and to infer their 3D properties from a single view.
no code implementations • 5 Oct 2020 • Katarína Tóthová, Sarah Parisot, Matthew Lee, Esther Puyol-Antón, Andrew King, Marc Pollefeys, Ender Konukoglu
Surface reconstruction from magnetic resonance (MR) imaging data is indispensable in medical image analysis and clinical research.
no code implementations • 28 Sep 2020 • Cathrin Elich, Martin R. Oswald, Marc Pollefeys, Joerg Stueckler
By differentiable rendering, we train our model to decompose scenes self-supervised from RGB-D images.
no code implementations • 22 Sep 2020 • Ivan Tishchenko, Sandro Lombardi, Martin R. Oswald, Marc Pollefeys
Most of the current scene flow methods choose to model scene flow as a per point translation vector without differentiating between static and dynamic components of 3D motion.
2 code implementations • 25 Aug 2020 • Dorin Ungureanu, Federica Bogo, Silvano Galliani, Pooja Sama, Xin Duan, Casey Meekhof, Jan Stühmer, Thomas J. Cashman, Bugra Tekin, Johannes L. Schönberger, Pawel Olszta, Marc Pollefeys
Mixed reality headsets, such as the Microsoft HoloLens 2, are powerful sensing devices with integrated compute capabilities, which makes it an ideal platform for computer vision research.
no code implementations • 17 Aug 2020 • Xingxing Zuo, Yulin Yang, Patrick Geneva, Jiajun Lv, Yong liu, Guoquan Huang, Marc Pollefeys
Only the tracked planar points belonging to the same plane will be used for plane initialization, which makes the plane extraction efficient and robust.
Robotics
no code implementations • 31 Jul 2020 • Audrey Richard, Ian Cherabier, Martin R. Oswald, Marc Pollefeys, Konrad Schindler
We present a novel 3D shape completion method that operates directly on unstructured point clouds, thus avoiding resource-intensive data structures like voxel grids.
1 code implementation • ECCV 2020 • Yukai Lin, Viktor Larsson, Marcel Geppert, Zuzana Kukelova, Marc Pollefeys, Torsten Sattler
In particular, our approach is more robust than the naive approach of first estimating intrinsic parameters and pose per camera before refining the extrinsic parameters of the system.
1 code implementation • ECCV 2020 • Rémi Pautrat, Viktor Larsson, Martin R. Oswald, Marc Pollefeys
To be invariant, or not to be invariant: that is the question formulated in this work about local descriptors.
no code implementations • CVPR 2021 • Mihai Dusmanu, Johannes L. Schönberger, Sudipta N. Sinha, Marc Pollefeys
Many computer vision systems require users to upload image features to the cloud for processing and storage.
3 code implementations • 7 Jun 2020 • Luca Cavalli, Viktor Larsson, Martin Ralf Oswald, Torsten Sattler, Marc Pollefeys
Local feature matching is a critical component of many computer vision pipelines, including among others Structure-from-Motion, SLAM, and Visual Localization.
1 code implementation • CVPR 2020 • Feitong Tan, Hao Zhu, Zhaopeng Cui, Siyu Zhu, Marc Pollefeys, Ping Tan
Previous methods on estimating detailed human depth often require supervised training with `ground truth' depth data.
no code implementations • CVPR 2020 • Yana Hasson, Bugra Tekin, Federica Bogo, Ivan Laptev, Marc Pollefeys, Cordelia Schmid
Modeling hand-object manipulations is essential for understanding how humans interact with their environment.
Ranked #9 on hand-object pose on HO-3D v2
no code implementations • 18 Mar 2020 • Changhee Won, Hochang Seok, Zhaopeng Cui, Marc Pollefeys, Jongwoo Lim
In this paper, we present an omnidirectional localization and dense mapping system for a wide-baseline multiview stereo setup with ultra-wide field-of-view (FOV) fisheye cameras, which has a 360 degrees coverage of stereo observations of the environment.
1 code implementation • ECCV 2020 • Mihai Dusmanu, Johannes L. Schönberger, Marc Pollefeys
In this work, we address the problem of refining the geometry of local image features from multiple views without known scene or camera geometry.
6 code implementations • ECCV 2020 • Songyou Peng, Michael Niemeyer, Lars Mescheder, Marc Pollefeys, Andreas Geiger
Recently, implicit neural representations have gained popularity for learning-based 3D reconstruction.
1 code implementation • 10 Feb 2020 • Peidong Liu, Joel Janai, Marc Pollefeys, Torsten Sattler, Andreas Geiger
Motion blurry images challenge many computer vision algorithms, e. g, feature detection, motion estimation, or object recognition.
2 code implementations • 17 Jan 2020 • Lucas Teixeira, Martin R. Oswald, Marc Pollefeys, Margarita Chli
In this paper, we propose a depth completion and uncertainty estimation approach that better handles the challenges of aerial platforms, such as large viewpoint and depth variations, and limited computing resources.
no code implementations • 14 Jan 2020 • Audrey Richard, Ian Cherabier, Martin R. Oswald, Vagia Tsiminaki, Marc Pollefeys, Konrad Schindler
We present a super-resolution method capable of creating a high-resolution texture map for a virtual 3D object from a set of lower-resolution images of that object.