Search Results for author: Divya Kothandaraman

Found 18 papers, 12 papers with code

Text2Story: Advancing Video Storytelling with Text Guidance

no code implementations8 Mar 2025 Taewon Kang, Divya Kothandaraman, Ming C. Lin

We introduce a novel storytelling approach to enable seamless video generation with natural action transitions and structured narratives.

Form Image Generation +1

ImPoster: Text and Frequency Guidance for Subject Driven Action Personalization using Diffusion Models

1 code implementation24 Sep 2024 Divya Kothandaraman, Kuldeep Kulkarni, Sumit Shekhar, Balaji Vasan Srinivasan, Dinesh Manocha

We propose a novel diffusion guidance formulation, image frequency guidance, to steer the generation towards the manifold of the source subject and the driving action at every step of the inference denoising.

Denoising

3D-free meets 3D priors: Novel View Synthesis from a Single Image with Pretrained Diffusion Guidance

no code implementations12 Aug 2024 Taewon Kang, Divya Kothandaraman, Dinesh Manocha, Ming C. Lin

Recent 3D novel view synthesis (NVS) methods often require extensive 3D data for training, and also typically lack generalization beyond the training distribution.

Image Generation Novel View Synthesis

Prompt Mixing in Diffusion Models using the Black Scholes Algorithm

1 code implementation22 May 2024 Divya Kothandaraman, Ming Lin, Dinesh Manocha

We introduce a novel approach for prompt mixing, aiming to generate images at the intersection of multiple text prompts using pre-trained text-to-image diffusion models.

Denoising

HawkI: Homography & Mutual Information Guidance for 3D-free Single Image to Aerial View

2 code implementations27 Nov 2023 Divya Kothandaraman, Tianyi Zhou, Ming Lin, Dinesh Manocha

It seamlessly blends the visual features from the input image within a pretrained text-to-2Dimage stable diffusion model with a test-time optimization process for a careful bias-variance trade-off, which uses an Inverse Perspective Mapping (IPM) homography transformation to provide subtle cues for aerialview synthesis.

Novel View Synthesis

Differentiable Frequency-based Disentanglement for Aerial Video Action Recognition

no code implementations15 Sep 2022 Divya Kothandaraman, Ming Lin, Dinesh Manocha

We build a differentiable static-dynamic frequency mask prior to model the salient static and dynamic pixels in the video, crucial for the underlying task of action recognition.

Action Recognition Activity Recognition In Videos +2

Placing Human Animations into 3D Scenes by Learning Interaction- and Geometry-Driven Keyframes

no code implementations13 Sep 2022 James F. Mullen Jr, Divya Kothandaraman, Aniket Bera, Dinesh Manocha

We compare our method, which we call PAAK, with prior approaches, including POSA, PROX ground truth, and a motion synthesis method, and highlight the benefits of our method with a perceptual study.

Human Animation Motion Synthesis

SALAD: Source-free Active Label-Agnostic Domain Adaptation for Classification, Segmentation and Detection

1 code implementation24 May 2022 Divya Kothandaraman, Sumit Shekhar, Abhilasha Sancheti, Manoj Ghuhan, Tripti Shukla, Dinesh Manocha

SALAD has three key benefits: (i) it is task-agnostic, and can be applied across various visual tasks such as classification, segmentation and detection; (ii) it can handle shifts in output label space from the pre-trained source network to the target domain; (iii) it does not require access to source data for adaptation.

Active Learning Domain Adaptation +2

FAR: Fourier Aerial Video Recognition

1 code implementation21 Mar 2022 Divya Kothandaraman, Tianrui Guan, Xijun Wang, Sean Hu, Ming Lin, Dinesh Manocha

Our formulation uses a novel Fourier object disentanglement method to innately separate out the human agent (which is typically small) from the background.

Action Recognition Disentanglement +1

GANav: Efficient Terrain Segmentation for Robot Navigation in Unstructured Outdoor Environments

1 code implementation7 Mar 2021 Tianrui Guan, Divya Kothandaraman, Rohan Chandra, Adarsh Jagan Sathyamoorthy, Kasun Weerakoon, Dinesh Manocha

We interface GANav with a deep reinforcement learning-based navigation algorithm and highlight its benefits in terms of navigation in real-world unstructured terrains.

Deep Reinforcement Learning Robot Navigation +1

Deep Atrous Guided Filter for Image Restoration in Under Display Cameras

1 code implementation14 Aug 2020 Varun Sundar, Sumanth Hegde, Divya Kothandaraman, Kaushik Mitra

In this work, we present Deep Atrous Guided Filter (DAGF), a two-stage, end-to-end approach for image restoration in UDC systems.

Image Restoration

Cannot find the paper you are looking for? You can Submit a new open access paper.