Search Results for author: Moshiur Farazi

Found 7 papers, 3 papers with code

VReBERT: A Simple and Flexible Transformer for Visual Relationship Detection

no code implementations18 Jun 2022 Yu Cui, Moshiur Farazi

Visual Relationship Detection (VRD) impels a computer vision model to 'see' beyond an individual object instance and 'understand' how different objects in a scene are related.

Object Relationship Detection +1

How You Start Matters for Generalization

no code implementations17 Jun 2022 Sameera Ramasinghe, Lachlan MacDonald, Moshiur Farazi, Hemanth Saratchandran, Simon Lucey

Characterizing the remarkable generalization properties of over-parameterized neural networks remains an open problem.

Recursive Training for Zero-Shot Semantic Segmentation

no code implementations26 Feb 2021 Ce Wang, Moshiur Farazi, Nick Barnes

We propose a recursive training scheme to supervise the retraining of a semantic segmentation model for a zero-shot setting using a pseudo-feature representation.

Segmentation Semantic Segmentation +1

Improving Action Quality Assessment using Weighted Aggregation

1 code implementation21 Feb 2021 Shafkat Farabi, Hasibul Himel, Fakhruddin Gazzali, Md. Bakhtiar Hasan, Md. Hasanul Kabir, Moshiur Farazi

We assess the effects of the depth and input clip size of the convolutional neural network on the quality of action score predictions.

Action Quality Assessment

Rethinking conditional GAN training: An approach using geometrically structured latent manifolds

1 code implementation NeurIPS 2021 Sameera Ramasinghe, Moshiur Farazi, Salman Khan, Nick Barnes, Stephen Gould

Conditional GANs (cGAN), in their rudimentary form, suffer from critical drawbacks such as the lack of diversity in generated outputs and distortion between the latent and output manifolds.

Image-to-Image Translation Translation

Attention Guided Semantic Relationship Parsing for Visual Question Answering

no code implementations5 Oct 2020 Moshiur Farazi, Salman Khan, Nick Barnes

Humans explain inter-object relationships with semantic labels that demonstrate a high-level understanding required to perform complex Vision-Language tasks such as Visual Question Answering (VQA).

Object Question Answering +1

Cannot find the paper you are looking for? You can Submit a new open access paper.