no code implementations • 22 Apr 2024 • Eric Brachmann, Jamie Wynn, Shuai Chen, Tommaso Cavallari, Áron Monszpart, Daniyar Turmukhambetov, Victor Adrian Prisacariu
We address the task of estimating camera parameters from a set of images depicting a scene.
1 code implementation • CVPR 2023 • Axel Barroso-Laguna, Eric Brachmann, Victor Adrian Prisacariu, Gabriel J. Brostow, Daniyar Turmukhambetov
As a remedy, we propose the Fundamental Scoring Network (FSNet), which infers a score for a pair of overlapping images and any proposed fundamental matrix.
1 code implementation • CVPR 2023 • Jamie Wynn, Daniyar Turmukhambetov
During NeRF training, random RGBD patches are rendered and the estimated gradient of the log-likelihood is backpropagated to the color and density fields.
1 code implementation • 11 Oct 2022 • Eduardo Arnold, Jamie Wynn, Sara Vicente, Guillermo Garcia-Hernando, Áron Monszpart, Victor Adrian Prisacariu, Daniyar Turmukhambetov, Eric Brachmann
Can we relocalize in a scene represented by a single reference image?
1 code implementation • CVPR 2021 • Michaël Ramamonjisoa, Michael Firman, Jamie Watson, Vincent Lepetit, Daniyar Turmukhambetov
We present a novel method for predicting accurate depths from monocular images with high efficiency.
no code implementations • 8 May 2021 • Anh-Dzung Doan, Daniyar Turmukhambetov, Yasir Latif, Tat-Jun Chin, Soohyun Bae
Many robotics applications require interest points that are highly repeatable under varying viewpoints and lighting conditions.
1 code implementation • 21 Aug 2020 • Carl Toft, Daniyar Turmukhambetov, Torsten Sattler, Fredrik Kahl, Gabriel Brostow
Good local features improve the robustness of many 3D re-localization and multi-view reconstruction pipelines.
no code implementations • 16 Aug 2020 • Iaroslav Melekhov, Gabriel J. Brostow, Juho Kannala, Daniyar Turmukhambetov
Local features that are robust to both viewpoint and appearance changes are crucial for many computer vision tasks.
1 code implementation • ECCV 2020 • Anita Rau, Guillermo Garcia-Hernando, Danail Stoyanov, Gabriel J. Brostow, Daniyar Turmukhambetov
Even when this is a known scene, the answer typically requires an expensive search across scale space, with matching and geometric verification of large sets of local features.
2 code implementations • ECCV 2020 • Jamie Watson, Oisin Mac Aodha, Daniyar Turmukhambetov, Gabriel J. Brostow, Michael Firman
We propose that it is unnecessary to have such a high reliance on ground truth depths or even corresponding stereo pairs.
1 code implementation • ICCV 2019 • Jamie Watson, Michael Firman, Gabriel J. Brostow, Daniyar Turmukhambetov
Monocular depth estimators can be trained with various forms of self-supervision from binocular-stereo data to circumvent the need for high-quality laser scans or other ground-truth data.
Ranked #2 on Monocular Depth Estimation on VA (Virtual Apartment)
no code implementations • ICCV 2017 • Daniel E. Worrall, Stephan J. Garbin, Daniyar Turmukhambetov, Gabriel J. Brostow
We propose a simple method to construct a deep feature space, with explicitly disentangled representations of several known transformations.
1 code implementation • CVPR 2017 • Daniel E. Worrall, Stephan J. Garbin, Daniyar Turmukhambetov, Gabriel J. Brostow
This is not the case for rotations.
no code implementations • CVPR 2015 • Daniyar Turmukhambetov, Neill D. F. Campbell, Simon J. D. Prince, Jan Kautz
In this work we remove the image space alignment limitations of existing subspace models by conditioning the models on a shape dependent context that allows for the complex, non-linear structure of the appearance of the visual object to be captured and shared.