Search Results for author: Hyeonggon Ryu

Found 5 papers, 1 papers with code

Sound Source Localization is All about Cross-Modal Alignment

no code implementations ICCV 2023 Arda Senocak, Hyeonggon Ryu, Junsik Kim, Tae-Hyun Oh, Hanspeter Pfister, Joon Son Chung

However, prior arts and existing benchmarks do not account for a more important aspect of the problem, cross-modal semantic understanding, which is essential for genuine sound source localization.

Cross-Modal Retrieval Retrieval

Hindi as a Second Language: Improving Visually Grounded Speech with Semantically Similar Samples

no code implementations30 Mar 2023 Hyeonggon Ryu, Arda Senocak, In So Kweon, Joon Son Chung

The objective of this work is to explore the learning of visually grounded speech models (VGS) from multilingual perspective.

Cross-Modal Retrieval Retrieval

Generative Bias for Robust Visual Question Answering

1 code implementation CVPR 2023 Jae Won Cho, Dong-Jin Kim, Hyeonggon Ryu, In So Kweon

In this work, in order to better learn the bias a target VQA model suffers from, we propose a generative method to train the bias model directly from the target model, called GenB.

Knowledge Distillation Question Answering +1

Audio-Visual Fusion Layers for Event Type Aware Video Recognition

no code implementations12 Feb 2022 Arda Senocak, Junsik Kim, Tae-Hyun Oh, Hyeonggon Ryu, DIngzeyu Li, In So Kweon

Human brain is continuously inundated with the multisensory information and their complex interactions coming from the outside world at any given moment.

Multi-Task Learning Video Recognition +1

Learning Sound Localization Better From Semantically Similar Samples

no code implementations7 Feb 2022 Arda Senocak, Hyeonggon Ryu, Junsik Kim, In So Kweon

Thus, these semantically correlated pairs, "hard positives", are mistakenly grouped as negatives.

Contrastive Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.