Search Results for author: Shervin Ardeshir

Found 17 papers, 2 papers with code

Representation Reliability and Its Impact on Downstream Tasks

no code implementations31 May 2023 Young-Jin Park, Hao Wang, Shervin Ardeshir, Navid Azizan

Self-supervised pre-trained models extract general-purpose representations from data, and quantifying how reliable they are is crucial because many downstream models use these representations as input for their own tasks.

Uncertainty Quantification

LLM2Loss: Leveraging Language Models for Explainable Model Diagnostics

no code implementations4 May 2023 Shervin Ardeshir

We then extract a semantically meaningful representation for each training data point (such as CLIP embeddings from its visual encoder) and train a lightweight diagnosis model which maps this semantically meaningful representation of a data point to its task loss.

Zero-Shot Learning

Improving Identity-Robustness for Face Models

no code implementations7 Apr 2023 Qi Qi, Shervin Ardeshir

When it comes to models directly trained on human faces, a sensitive confounder is that of human identities.

Face Recognition

Fairness via Adversarial Attribute Neighbourhood Robust Learning

no code implementations12 Oct 2022 Qi Qi, Shervin Ardeshir, Yi Xu, Tianbao Yang

Improving fairness between privileged and less-privileged sensitive attribute groups (e. g, {race, gender}) has attracted lots of attention.

Attribute Fairness

Uncertainty in Contrastive Learning: On the Predictability of Downstream Performance

no code implementations19 Jul 2022 Shervin Ardeshir, Navid Azizan

In this work, we study whether the uncertainty of such a representation can be quantified for a single datapoint in a meaningful way.

Contrastive Learning Decision Making

On Negative Sampling for Audio-Visual Contrastive Learning from Movies

no code implementations29 Apr 2022 Mahdi M. Kalayeh, Shervin Ardeshir, Lingyi Liu, Nagendra Kamath, Ashok Chandrashekar

The abundance and ease of utilizing sound, along with the fact that auditory clues reveal a plethora of information about what happens in a scene, make the audio-visual space an intuitive choice for representation learning.

Action Recognition Audio Classification +3

Character-focused Video Thumbnail Retrieval

no code implementations13 Apr 2022 Shervin Ardeshir, Nagendra Kamath, Hossein Taghavi

Prominence and interactions: Character(s) in the thumbnail should be important character(s) in the video, to prevent the algorithm from suggesting non-representative frames as candidates.

Face Clustering Retrieval

Estimating Structural Disparities for Face Models

no code implementations CVPR 2022 Shervin Ardeshir, Cristina Segalin, Nathan Kallus

Performance of the model for each group is calculated by comparing $\hat{y}$ and $y$ for the datapoints within a specific group, and as a result, disparity of performance across the different groups can be calculated.

Attribute Face Recognition

On Attention Modules for Audio-Visual Synchronization

no code implementations14 Dec 2018 Naji Khosravan, Shervin Ardeshir, Rohit Puri

To judge whether audio and video signals of a multimedia presentation are synchronized, we as humans often pay close attention to discriminative spatio-temporal blocks of the video (e. g. synchronizing the lip movement with the utterance of words, or the sound of a bouncing ball at the moment it hits the ground).

Audio-Visual Synchronization

From Third Person to First Person: Dataset and Baselines for Synthesis and Retrieval

1 code implementation1 Dec 2018 Mohamed Elfeki, Krishna Regmi, Shervin Ardeshir, Ali Borji

In this work, we introduce two datasets (synthetic and natural/real) containing simultaneously recorded egocentric and exocentric videos.

Domain Adaptation Generative Adversarial Network +2

Integrating Egocentric Videos in Top-view Surveillance Videos: Joint Identification and Temporal Alignment

no code implementations ECCV 2018 Shervin Ardeshir, Ali Borji

Videos recorded from first person (egocentric) perspective have little visual appearance in common with those from third person perspective, especially with videos captured by top-view surveillance cameras.

EgoReID: Cross-view Self-Identification and Human Re-identification in Egocentric and Surveillance Videos

no code implementations24 Dec 2016 Shervin Ardeshir, Sandesh Sharma, Ali Broji

Human identification remains to be one of the challenging tasks in computer vision community due to drastic changes in visual features across different viewpoints, lighting conditions, occlusion, etc.

Person Re-Identification Visual Reasoning

EgoTransfer: Transferring Motion Across Egocentric and Exocentric Domains using Deep Neural Networks

no code implementations17 Dec 2016 Shervin Ardeshir, Krishna Regmi, Ali Borji

On one hand, the abundance of egocentric cameras in the past few years has offered the opportunity to study a lot of vision problems from the first-person perspective.

Egocentric Meets Top-view

no code implementations30 Aug 2016 Shervin Ardeshir, Ali Borji

First, having a set of egocentric videos and a top-view video, can we verify if the top-view video contains all, or some of the egocentric viewers present in the egocentric set?

Graph Matching

Ego2Top: Matching Viewers in Egocentric and Top-view Videos

no code implementations24 Jul 2016 Shervin Ardeshir, Ali Borji

At the same time, surveillance cameras and drones offer an abundance of visual information, often captured from top-view.

Graph Matching

Geo-Semantic Segmentation

1 code implementation CVPR 2015 Shervin Ardeshir, Kofi Malcolm Collins-Sibley, Mubarak Shah

In this paper, we propose a method which leverages information acquired from GIS databases to perform semantic segmentation of the image alongside with geo-referencing each semantic segment with its address and geo-location.

Segmentation Semantic Segmentation

GPS-Tag Refinement using Random Walks with an Adaptive Damping Factor

no code implementations CVPR 2014 Amir Roshan Zamir, Shervin Ardeshir, Mubarak Shah

We develop a robust method for identification and refinement of this subset using the rest of the images in the dataset.

TAG

Cannot find the paper you are looking for? You can Submit a new open access paper.