Search Results for author: M. Jehanzeb Mirza

Found 8 papers, 6 papers with code

Into the Fog: Evaluating Multiple Object Tracking Robustness

no code implementations12 Apr 2024 Nadezda Kirillova, M. Jehanzeb Mirza, Horst Possegger, Horst Bischof

To address these limitations, we propose a pipeline for physic-based volumetric fog simulation in arbitrary real-world MOT dataset utilizing frame-by-frame monocular depth estimation and a fog formation optical model.

Monocular Depth Estimation Multiple Object Tracking +1

Towards Multimodal In-Context Learning for Vision & Language Models

no code implementations19 Mar 2024 Sivan Doveh, Shaked Perek, M. Jehanzeb Mirza, Amit Alfassy, Assaf Arbelle, Shimon Ullman, Leonid Karlinsky

Inspired by the emergence of Large Language Models (LLMs) that can truly understand human language, significant progress has been made in aligning other, non-language, modalities to be `understandable' by an LLM, primarily via converting their samples into a sequence of embedded language-like tokens directly fed into the LLM (decoder) input stream.

In-Context Learning

Meta-Prompting for Automating Zero-shot Visual Recognition with LLMs

1 code implementation18 Mar 2024 M. Jehanzeb Mirza, Leonid Karlinsky, Wei Lin, Sivan Doveh, Jakub Micorek, Mateusz Kozinski, Hilde Kuhene, Horst Possegger

Prompt ensembling of Large Language Model (LLM) generated category-specific prompts has emerged as an effective method to enhance zero-shot recognition ability of Vision-Language Models (VLMs).

Language Modelling Large Language Model +1

TAP: Targeted Prompting for Task Adaptive Generation of Textual Training Instances for Visual Classification

1 code implementation13 Sep 2023 M. Jehanzeb Mirza, Leonid Karlinsky, Wei Lin, Horst Possegger, Rogerio Feris, Horst Bischof

Vision and Language Models (VLMs), such as CLIP, have enabled visual recognition of a potentially unlimited set of categories described by text prompts.

Zero-Shot Learning

Sit Back and Relax: Learning to Drive Incrementally in All Weather Conditions

1 code implementation30 May 2023 Stefan Leitner, M. Jehanzeb Mirza, Wei Lin, Jakub Micorek, Marc Masana, Mateusz Kozinski, Horst Possegger, Horst Bischof

We propose to store these affine parameters as a memory bank for each weather condition and plug-in their weather-specific parameters during driving (i. e. test time) when the respective weather conditions are encountered.

Autonomous Driving Incremental Learning +2

MATE: Masked Autoencoders are Online 3D Test-Time Learners

1 code implementation ICCV 2023 M. Jehanzeb Mirza, Inkyu Shin, Wei Lin, Andreas Schriebl, Kunyang Sun, Jaesung Choe, Horst Possegger, Mateusz Kozinski, In So Kweon, Kun-Jin Yoon, Horst Bischof

Our MATE is the first Test-Time-Training (TTT) method designed for 3D data, which makes deep networks trained for point cloud classification robust to distribution shifts occurring in test data.

3D Object Classification Point Cloud Classification

An Efficient Domain-Incremental Learning Approach to Drive in All Weather Conditions

1 code implementation19 Apr 2022 M. Jehanzeb Mirza, Marc Masana, Horst Possegger, Horst Bischof

This catastrophic forgetting is typically addressed via incremental learning approaches which usually re-train the model by either keeping a memory bank of training samples or keeping a copy of the entire model or model parameters for each scenario.

Autonomous Driving Incremental Learning +2

The Norm Must Go On: Dynamic Unsupervised Domain Adaptation by Normalization

1 code implementation CVPR 2022 M. Jehanzeb Mirza, Jakub Micorek, Horst Possegger, Horst Bischof

This can be a hurdle in fields which require continuous dynamic adaptation or suffer from scarcity of data, e. g. autonomous driving in challenging weather conditions.

Autonomous Driving object-detection +3

Cannot find the paper you are looking for? You can Submit a new open access paper.