Search Results for author: Anand Bhattad

Found 23 papers, 6 papers with code

LumiNet: Latent Intrinsics Meets Diffusion Models for Indoor Scene Relighting

no code implementations29 Nov 2024 Xiaoyan Xing, Konrad Groh, Sezer Karaoglu, Theo Gevers, Anand Bhattad

Our approach makes two key contributions: a data curation strategy from the StyleGAN-based relighting model for our training, and a modified diffusion-based ControlNet that processes both latent intrinsic properties from the source image and latent extrinsic properties from the target image.

SpotLight: Shadow-Guided Object Relighting via Diffusion

no code implementations27 Nov 2024 Frédéric Fortier-Chouinard, Zitian Zhang, Louis-Etienne Messier, Mathieu Garon, Anand Bhattad, Jean-François Lalonde

In this paper, we show that precise lighting control can be achieved for object relighting simply by specifying the desired shadows of the object.

Neural Rendering Object

ScribbleLight: Single Image Indoor Relighting with Scribbles

no code implementations26 Nov 2024 Jun Myeong Choi, Annie Wang, Pieter Peers, Anand Bhattad, Roni Sengupta

Our key technical novelty is an Albedo-conditioned Stable Image Diffusion model that preserves the intrinsic color and texture of the original image after relighting and an encoder-decoder-based ControlNet architecture that enables geometry-preserving lighting effects with normal map and scribble annotations.

Decoder

ZeroComp: Zero-shot Object Compositing from Image Intrinsics via Diffusion

no code implementations10 Oct 2024 Zitian Zhang, Frédéric Fortier-Chouinard, Mathieu Garon, Anand Bhattad, Jean-François Lalonde

We present ZeroComp, an effective zero-shot 3D object compositing approach that does not require paired composite-scene images during training.

Latent Intrinsics Emerge from Training to Relight

no code implementations31 May 2024 Xiao Zhang, William Gao, Seemandhar Jain, Michael Maire, David. A. Forsyth, Anand Bhattad

Image relighting is the task of showing what a scene from a source image would look like if illuminated differently.

Image Relighting

Videoshop: Localized Semantic Video Editing with Noise-Extrapolated Diffusion Inversion

no code implementations21 Mar 2024 Xiang Fan, Anand Bhattad, Ranjay Krishna

We introduce Videoshop, a training-free video editing algorithm for localized semantic edits.

Video Editing

StyLitGAN: Image-Based Relighting via Latent Control

no code implementations CVPR 2024 Anand Bhattad, James Soole, D.A. Forsyth

We describe a novel method StyLitGAN for relighting and resurfacing images in the absence of labeled data.

Diversity

Generative Models: What Do They Know? Do They Know Things? Let's Find Out!

no code implementations28 Nov 2023 Xiaodan Du, Nicholas Kolkin, Greg Shakhnarovich, Anand Bhattad

Generative models excel at mimicking real scenes, suggesting they might inherently encode important intrinsic scene properties.

Improving Equivariance in State-of-the-Art Supervised Depth and Normal Predictors

1 code implementation ICCV 2023 Yuanyi Zhong, Anand Bhattad, Yu-Xiong Wang, David Forsyth

Dense depth and surface normal predictors should possess the equivariant property to cropping-and-resizing -- cropping the input image should result in cropping the same output image.

Data Augmentation

Blocks2World: Controlling Realistic Scenes with Editable Primitives

no code implementations7 Jul 2023 Vaibhav Vavilala, Seemandhar Jain, Rahul Vasanth, Anand Bhattad, David Forsyth

We present Blocks2World, a novel method for 3D scene rendering and editing that leverages a two-step process: convex decomposition of images and conditioned synthesis.

Data Augmentation

MIMIC: Masked Image Modeling with Image Correspondences

1 code implementation27 Jun 2023 Kalyani Marathe, Mahtab Bigverdi, Nishat Khan, Tuhin Kundu, Patrick Howe, Sharan Ranjit S, Anand Bhattad, Aniruddha Kembhavi, Linda G. Shapiro, Ranjay Krishna

We train multiple models with different masked image modeling objectives to showcase the following findings: Representations trained on our automatically generated MIMIC-3M outperform those learned from expensive crowdsourced datasets (ImageNet-1K) and those learned from synthetic environments (MULTIVIEW-HABITAT) on two dense geometric tasks: depth estimation on NYUv2 (1. 7%), and surface normals estimation on Taskonomy (2. 05%).

Depth Estimation Pose Estimation +3

UrbanIR: Large-Scale Urban Scene Inverse Rendering from a Single Video

no code implementations15 Jun 2023 Zhi-Hao Lin, Bohan Liu, Yi-Ting Chen, Kuan-Sheng Chen, David Forsyth, Jia-Bin Huang, Anand Bhattad, Shenlong Wang

We present UrbanIR (Urban Scene Inverse Rendering), a new inverse graphics model that enables realistic, free-viewpoint renderings of scenes under various lighting conditions with a single video.

Inverse Rendering

Make It So: Steering StyleGAN for Any Image Inversion and Editing

no code implementations27 Apr 2023 Anand Bhattad, Viraj Shah, Derek Hoiem, D. A. Forsyth

StyleGAN's disentangled style representation enables powerful image editing by manipulating the latent variables, but accurately mapping real-world images to their latent variables (GAN inversion) remains a challenge.

StyLitGAN: Prompting StyleGAN to Produce New Illumination Conditions

no code implementations20 May 2022 Anand Bhattad, D. A. Forsyth

We propose a novel method, StyLitGAN, for relighting and resurfacing generated images in the absence of labeled data.

Diversity

SIRfyN: Single Image Relighting from your Neighbors

no code implementations8 Dec 2021 D. A. Forsyth, Anand Bhattad, Pranav Asthana, Yuanyi Zhong, YuXiong Wang

Novel theory shows that one can use similar scenes to estimate the different lightings that apply to a given scene, with bounded expected error.

Data Augmentation Image Relighting

View Generalization for Single Image Textured 3D Models

no code implementations CVPR 2021 Anand Bhattad, Aysegul Dundar, Guilin Liu, Andrew Tao, Bryan Catanzaro

We describe a cycle consistency loss that encourages model textures to be aligned, so as to encourage sharing.

3D geometry

Cut-and-Paste Object Insertion by Enabling Deep Image Prior for Reshading

no code implementations12 Oct 2020 Anand Bhattad, David A. Forsyth

We show how to insert an object from one image to another and get realistic results in the hard case, where the shading of the inserted object clashes with the shading of the scene.

Image Harmonization Neural Rendering +1

Improving Style Transfer with Calibrated Metrics

1 code implementation21 Oct 2019 Mao-Chuang Yeh, Shuai Tang, Anand Bhattad, Chuhang Zou, David Forsyth

Style transfer methods produce a transferred image which is a rendering of a content image in the manner of a style image.

Style Transfer

Unrestricted Adversarial Examples via Semantic Manipulation

1 code implementation ICLR 2020 Anand Bhattad, Min Jin Chong, Kaizhao Liang, Bo Li, D. A. Forsyth

Machine learning models, especially deep neural networks (DNNs), have been shown to be vulnerable against adversarial examples which are carefully crafted samples with a small magnitude of the perturbation.

Colorization Image Captioning +1

Quantitative Evaluation of Style Transfer

no code implementations31 Mar 2018 Mao-Chuang Yeh, Shuai Tang, Anand Bhattad, D. A. Forsyth

Style transfer methods produce a transferred image which is a rendering of a content image in the manner of a style image.

Style Transfer

Detecting Anomalous Faces with 'No Peeking' Autoencoders

no code implementations15 Feb 2018 Anand Bhattad, Jason Rock, David Forsyth

We describe a method for detecting an anomalous face image that meets these requirements.

Cannot find the paper you are looking for? You can Submit a new open access paper.