no code implementations • 29 Nov 2024 • Xiaoyan Xing, Konrad Groh, Sezer Karaoglu, Theo Gevers, Anand Bhattad
Our approach makes two key contributions: a data curation strategy from the StyleGAN-based relighting model for our training, and a modified diffusion-based ControlNet that processes both latent intrinsic properties from the source image and latent extrinsic properties from the target image.
no code implementations • 27 Nov 2024 • Frédéric Fortier-Chouinard, Zitian Zhang, Louis-Etienne Messier, Mathieu Garon, Anand Bhattad, Jean-François Lalonde
In this paper, we show that precise lighting control can be achieved for object relighting simply by specifying the desired shadows of the object.
no code implementations • 26 Nov 2024 • Jun Myeong Choi, Annie Wang, Pieter Peers, Anand Bhattad, Roni Sengupta
Our key technical novelty is an Albedo-conditioned Stable Image Diffusion model that preserves the intrinsic color and texture of the original image after relighting and an encoder-decoder-based ControlNet architecture that enables geometry-preserving lighting effects with normal map and scribble annotations.
no code implementations • 10 Oct 2024 • Zitian Zhang, Frédéric Fortier-Chouinard, Mathieu Garon, Anand Bhattad, Jean-François Lalonde
We present ZeroComp, an effective zero-shot 3D object compositing approach that does not require paired composite-scene images during training.
no code implementations • 31 May 2024 • Xiao Zhang, William Gao, Seemandhar Jain, Michael Maire, David. A. Forsyth, Anand Bhattad
Image relighting is the task of showing what a scene from a source image would look like if illuminated differently.
no code implementations • 21 Mar 2024 • Xiang Fan, Anand Bhattad, Ranjay Krishna
We introduce Videoshop, a training-free video editing algorithm for localized semantic edits.
no code implementations • CVPR 2024 • Anand Bhattad, James Soole, D.A. Forsyth
We describe a novel method StyLitGAN for relighting and resurfacing images in the absence of labeled data.
no code implementations • 28 Nov 2023 • Xiaodan Du, Nicholas Kolkin, Greg Shakhnarovich, Anand Bhattad
Generative models excel at mimicking real scenes, suggesting they might inherently encode important intrinsic scene properties.
1 code implementation • CVPR 2024 • Ayush Sarkar, Hanlin Mai, Amitabh Mahapatra, Svetlana Lazebnik, D. A. Forsyth, Anand Bhattad
All three classifiers are denied access to image pixels, and look only at derived geometric features.
1 code implementation • ICCV 2023 • Yuanyi Zhong, Anand Bhattad, Yu-Xiong Wang, David Forsyth
Dense depth and surface normal predictors should possess the equivariant property to cropping-and-resizing -- cropping the input image should result in cropping the same output image.
no code implementations • 7 Jul 2023 • Vaibhav Vavilala, Seemandhar Jain, Rahul Vasanth, Anand Bhattad, David Forsyth
We present Blocks2World, a novel method for 3D scene rendering and editing that leverages a two-step process: convex decomposition of images and conditioned synthesis.
1 code implementation • 27 Jun 2023 • Kalyani Marathe, Mahtab Bigverdi, Nishat Khan, Tuhin Kundu, Patrick Howe, Sharan Ranjit S, Anand Bhattad, Aniruddha Kembhavi, Linda G. Shapiro, Ranjay Krishna
We train multiple models with different masked image modeling objectives to showcase the following findings: Representations trained on our automatically generated MIMIC-3M outperform those learned from expensive crowdsourced datasets (ImageNet-1K) and those learned from synthetic environments (MULTIVIEW-HABITAT) on two dense geometric tasks: depth estimation on NYUv2 (1. 7%), and surface normals estimation on Taskonomy (2. 05%).
no code implementations • 15 Jun 2023 • Zhi-Hao Lin, Bohan Liu, Yi-Ting Chen, Kuan-Sheng Chen, David Forsyth, Jia-Bin Huang, Anand Bhattad, Shenlong Wang
We present UrbanIR (Urban Scene Inverse Rendering), a new inverse graphics model that enables realistic, free-viewpoint renderings of scenes under various lighting conditions with a single video.
no code implementations • 27 Apr 2023 • Anand Bhattad, Viraj Shah, Derek Hoiem, D. A. Forsyth
StyleGAN's disentangled style representation enables powerful image editing by manipulating the latent variables, but accurately mapping real-world images to their latent variables (GAN inversion) remains a challenge.
no code implementations • 20 May 2022 • Anand Bhattad, D. A. Forsyth
We propose a novel method, StyLitGAN, for relighting and resurfacing generated images in the absence of labeled data.
no code implementations • 8 Dec 2021 • D. A. Forsyth, Anand Bhattad, Pranav Asthana, Yuanyi Zhong, YuXiong Wang
Novel theory shows that one can use similar scenes to estimate the different lightings that apply to a given scene, with bounded expected error.
2 code implementations • CVPR 2022 • Liwen Wu, Jae Yong Lee, Anand Bhattad, YuXiong Wang, David Forsyth
DIVeR's representation is a voxel based field of features.
no code implementations • CVPR 2021 • Anand Bhattad, Aysegul Dundar, Guilin Liu, Andrew Tao, Bryan Catanzaro
We describe a cycle consistency loss that encourages model textures to be aligned, so as to encourage sharing.
no code implementations • 12 Oct 2020 • Anand Bhattad, David A. Forsyth
We show how to insert an object from one image to another and get realistic results in the hard case, where the shading of the inserted object clashes with the shading of the scene.
1 code implementation • 21 Oct 2019 • Mao-Chuang Yeh, Shuai Tang, Anand Bhattad, Chuhang Zou, David Forsyth
Style transfer methods produce a transferred image which is a rendering of a content image in the manner of a style image.
1 code implementation • ICLR 2020 • Anand Bhattad, Min Jin Chong, Kaizhao Liang, Bo Li, D. A. Forsyth
Machine learning models, especially deep neural networks (DNNs), have been shown to be vulnerable against adversarial examples which are carefully crafted samples with a small magnitude of the perturbation.
no code implementations • 31 Mar 2018 • Mao-Chuang Yeh, Shuai Tang, Anand Bhattad, D. A. Forsyth
Style transfer methods produce a transferred image which is a rendering of a content image in the manner of a style image.
no code implementations • 15 Feb 2018 • Anand Bhattad, Jason Rock, David Forsyth
We describe a method for detecting an anomalous face image that meets these requirements.