Search Results for author: Risheek Garrepalli

Found 17 papers, 2 papers with code

Distilling Multi-modal Large Language Models for Autonomous Driving

no code implementations16 Jan 2025 Deepti Hegde, Rajeev Yasarla, Hong Cai, Shizhong Han, Apratim Bhattacharyya, Shweta Mahajan, Litian Liu, Risheek Garrepalli, Vishal M. Patel, Fatih Porikli

Training with DiMA results in a 37% reduction in the L2 trajectory error and an 80% reduction in the collision rate of the vision-based planner, as well as a 44% trajectory error reduction in longtail scenarios.

Autonomous Driving Motion Planning +1

DDIL: Improved Diffusion Distillation With Imitation Learning

no code implementations15 Oct 2024 Risheek Garrepalli, Shweta Mahajan, Munawar Hayat, Fatih Porikli

Efforts such as progressive distillation or consistency distillation have shown promise by reducing the number of passes at the expense of quality of the generated samples.

Denoising Imitation Learning

Rapid Switching and Multi-Adapter Fusion via Sparse High Rank Adapters

no code implementations22 Jul 2024 Kartikeya Bhardwaj, Nilesh Prasad Pandey, Sweta Priyadarshi, Viswanath Ganapathy, Rafael Esteves, Shreya Kadambi, Shubhankar Borse, Paul Whatmough, Risheek Garrepalli, Mart van Baalen, Harris Teague, Markus Nagel

In this paper, we propose Sparse High Rank Adapters (SHiRA) that directly finetune 1-2% of the base model weights while leaving others unchanged, thus, resulting in a highly sparse adapter.

Sparse High Rank Adapters

no code implementations19 Jun 2024 Kartikeya Bhardwaj, Nilesh Prasad Pandey, Sweta Priyadarshi, Viswanath Ganapathy, Shreya Kadambi, Rafael Esteves, Shubhankar Borse, Paul Whatmough, Risheek Garrepalli, Mart van Baalen, Harris Teague, Markus Nagel

However, from a mobile deployment standpoint, we can either avoid inference overhead in the fused mode but lose the ability to switch adapters rapidly, or suffer significant (up to 30% higher) inference latency while enabling rapid switching in the unfused mode.

FouRA: Fourier Low Rank Adaptation

no code implementations13 Jun 2024 Shubhankar Borse, Shreya Kadambi, Nilesh Prasad Pandey, Kartikeya Bhardwaj, Viswanath Ganapathy, Sweta Priyadarshi, Risheek Garrepalli, Rafael Esteves, Munawar Hayat, Fatih Porikli

While Low-Rank Adaptation (LoRA) has proven beneficial for efficiently fine-tuning large models, LoRA fine-tuned text-to-image diffusion models lack diversity in the generated images, as the model tends to copy data from the observed training samples.

Diversity

FutureDepth: Learning to Predict the Future Improves Video Depth Estimation

no code implementations19 Mar 2024 Rajeev Yasarla, Manish Kumar Singh, Hong Cai, Yunxiao Shi, Jisoo Jeong, Yinhao Zhu, Shizhong Han, Risheek Garrepalli, Fatih Porikli

In this paper, we propose a novel video depth estimation approach, FutureDepth, which enables the model to implicitly leverage multi-frame and motion cues to improve depth estimation by making it learn to predict the future at training.

Decoder Future prediction +1

DIFT: Dynamic Iterative Field Transforms for Memory Efficient Optical Flow

no code implementations9 Jun 2023 Risheek Garrepalli, Jisoo Jeong, Rajeswaran C Ravindran, Jamie Menjay Lin, Fatih Porikli

Also, we present a novel dynamic coarse-to-fine cost volume processing during various stages of refinement to avoid multiple levels of cost volumes.

Optical Flow Estimation

DistractFlow: Improving Optical Flow Estimation via Realistic Distractions and Pseudo-Labeling

no code implementations CVPR 2023 Jisoo Jeong, Hong Cai, Risheek Garrepalli, Fatih Porikli

We propose a novel data augmentation approach, DistractFlow, for training optical flow estimation models by introducing realistic distractions to the input frames.

Data Augmentation Optical Flow Estimation

DejaVu: Conditional Regenerative Learning to Enhance Dense Prediction

no code implementations CVPR 2023 Shubhankar Borse, Debasmit Das, Hyojin Park, Hong Cai, Risheek Garrepalli, Fatih Porikli

Next, we use a conditional regenerator, which takes the redacted image and the dense predictions as inputs, and reconstructs the original image by filling in the missing structural information.

Depth Estimation Prediction

TransAdapt: A Transformative Framework for Online Test Time Adaptive Semantic Segmentation

no code implementations24 Feb 2023 Debasmit Das, Shubhankar Borse, Hyojin Park, Kambiz Azarian, Hong Cai, Risheek Garrepalli, Fatih Porikli

Test-time adaptive (TTA) semantic segmentation adapts a source pre-trained image semantic segmentation model to unlabeled batches of target domain test images, different from real-world, where samples arrive one-by-one in an online fashion.

Segmentation Semantic Segmentation +1

Oracle Analysis of Representations for Deep Open Set Detection

no code implementations22 Sep 2022 Risheek Garrepalli

The second is to introduce Oracle representation learning, which produces a representation that is guaranteed to be sufficient for accurate anomaly detection.

Anomaly Detection Autonomous Driving +1

Open Category Detection with PAC Guarantees

1 code implementation ICML 2018 Si Liu, Risheek Garrepalli, Thomas G. Dietterich, Alan Fern, Dan Hendrycks

Further, while there are algorithms for open category detection, there are few empirical results that directly report alien detection rates.

Cannot find the paper you are looking for? You can Submit a new open access paper.