no code implementations • 14 Dec 2023 • Krishna Mullia, Fujun Luan, Xin Sun, Miloš Hašan
We combine an MLP decoder with a feature grid.
no code implementations • 6 Jul 2023 • Kai Yan, Fujun Luan, Miloš Hašan, Thibault Groueix, Valentin Deschaintre, Shuang Zhao
A 3D digital scene contains many components: lights, materials and geometries, interacting to reach the desired appearance.
no code implementations • 20 May 2023 • Xilong Zhou, Miloš Hašan, Valentin Deschaintre, Paul Guerrero, Yannick Hold-Geoffroy, Kalyan Sunkavalli, Nima Khademi Kalantari
Instead, we train a generator for a neural material representation that is rendered with a learned relighting module to create arbitrarily lit RGB images; these are compared against real photos using a discriminator.
1 code implementation • CVPR 2022 • Yu-Ying Yeh, Zhengqin Li, Yannick Hold-Geoffroy, Rui Zhu, Zexiang Xu, Miloš Hašan, Kalyan Sunkavalli, Manmohan Chandraker
Most indoor 3D scene reconstruction methods focus on recovering 3D geometry and scene layout.
no code implementations • 12 Jun 2022 • Xilong Zhou, Miloš Hašan, Valentin Deschaintre, Paul Guerrero, Kalyan Sunkavalli, Nima Kalantari
The resulting materials are tileable, can be larger than the target image, and are editable by varying the condition.
no code implementations • 10 Jun 2022 • Sai Praveen Bangaru, Michaël Gharbi, Tzu-Mao Li, Fujun Luan, Kalyan Sunkavalli, Miloš Hašan, Sai Bi, Zexiang Xu, Gilbert Bernstein, Frédo Durand
Our method leverages the distance to surface encoded in an SDF and uses quadrature on sphere tracer points to compute this warping function.
no code implementations • 19 May 2022 • Zhengqin Li, Jia Shi, Sai Bi, Rui Zhu, Kalyan Sunkavalli, Miloš Hašan, Zexiang Xu, Ravi Ramamoorthi, Manmohan Chandraker
We tackle this problem using two novel components: 1) a holistic scene reconstruction method that estimates scene reflectance and parametric 3D lighting, and 2) a neural rendering framework that re-renders the scene from our predictions.
no code implementations • 6 Nov 2021 • Jiahui Fan, Beibei Wang, Miloš Hašan, Jian Yang, Ling-Qi Yan
Bidirectional reflectance distribution functions (BRDFs) are pervasively used in computer graphics to produce realistic physically-based appearance.
no code implementations • 6 Apr 2021 • Alexandr Kuznetsov, Krishna Mullia, Zexiang Xu, Miloš Hašan, Ravi Ramamoorthi
We also introduce neural offsets, a novel method which allows rendering materials with intricate parallax effects without any tessellation.
1 code implementation • CVPR 2021 • Fanbo Xiang, Zexiang Xu, Miloš Hašan, Yannick Hold-Geoffroy, Kalyan Sunkavalli, Hao Su
We achieve this by introducing a 3D-to-2D texture mapping (or surface parameterization) network into volumetric representations.
no code implementations • 30 Sep 2020 • Yu Guo, Cameron Smith, Miloš Hašan, Kalyan Sunkavalli, Shuang Zhao
We address the problem of reconstructing spatially-varying BRDFs from a small set of image measurements.
no code implementations • 9 Aug 2020 • Sai Bi, Zexiang Xu, Pratul Srinivasan, Ben Mildenhall, Kalyan Sunkavalli, Miloš Hašan, Yannick Hold-Geoffroy, David Kriegman, Ravi Ramamoorthi
We combine this representation with a physically-based differentiable ray marching framework that can render images from a neural reflectance field under any viewpoint and light.
no code implementations • 25 Jul 2020 • Zhengqin Li, Ting-Wei Yu, Shen Sang, Sarah Wang, Meng Song, YuHan Liu, Yu-Ying Yeh, Rui Zhu, Nitesh Gundavarapu, Jia Shi, Sai Bi, Zexiang Xu, Hong-Xing Yu, Kalyan Sunkavalli, Miloš Hašan, Ravi Ramamoorthi, Manmohan Chandraker
Finally, we demonstrate that our framework may also be integrated with physics engines, to create virtual robotics environments with unique ground truth such as friction coefficients and correspondence to real scenes.
no code implementations • ECCV 2020 • Sai Bi, Zexiang Xu, Kalyan Sunkavalli, Miloš Hašan, Yannick Hold-Geoffroy, David Kriegman, Ravi Ramamoorthi
We also show that our learned reflectance volumes are editable, allowing for modifying the materials of the captured scenes.