no code implementations • ICCV 2023 • Arda Senocak, Hyeonggon Ryu, Junsik Kim, Tae-Hyun Oh, Hanspeter Pfister, Joon Son Chung
However, prior arts and existing benchmarks do not account for a more important aspect of the problem, cross-modal semantic understanding, which is essential for genuine sound source localization.
no code implementations • 30 Mar 2023 • Hyeonggon Ryu, Arda Senocak, In So Kweon, Joon Son Chung
The objective of this work is to explore the learning of visually grounded speech models (VGS) from multilingual perspective.
1 code implementation • CVPR 2023 • Jae Won Cho, Dong-Jin Kim, Hyeonggon Ryu, In So Kweon
In this work, in order to better learn the bias a target VQA model suffers from, we propose a generative method to train the bias model directly from the target model, called GenB.
no code implementations • 12 Feb 2022 • Arda Senocak, Junsik Kim, Tae-Hyun Oh, Hyeonggon Ryu, DIngzeyu Li, In So Kweon
Human brain is continuously inundated with the multisensory information and their complex interactions coming from the outside world at any given moment.
no code implementations • 7 Feb 2022 • Arda Senocak, Hyeonggon Ryu, Junsik Kim, In So Kweon
Thus, these semantically correlated pairs, "hard positives", are mistakenly grouped as negatives.