no code implementations • 1 Apr 2024 • Chull Hwan Song, Taebaek Hwang, Jooyoung Yoon, Shunghyun Choi, Yeong Hyeon Gu
Vision-language models (VLMs) have made significant strides in cross-modal understanding through large-scale paired datasets.
1 code implementation • 1 Apr 2024 • Chull Hwan Song, Jooyoung Yoon, Taebaek Hwang, Shunghyun Choi, Yeong Hyeon Gu, Yannis Avrithis
How important is it for training and evaluation sets to not have class overlap in image retrieval?
no code implementations • ICCV 2023 • Chull Hwan Song, Taebaek Hwang, Jooyoung Yoon, Shunghyun Choi, Yeong Hyeon Gu
Many studies in vision tasks have aimed to create effective embedding spaces for single-label object prediction within an image.
1 code implementation • 21 Oct 2022 • Chull Hwan Song, Jooyoung Yoon, Shunghyun Choi, Yannis Avrithis
(4) We enhance locality of interactions at the deeper layers of the encoder, which is the relative weakness of vision transformers.
no code implementations • ACL 2020 • Hyeryun Park, Kyungmo Kim, Jooyoung Yoon, Seongkeun Park, Jinwook Choi
Medical image captioning can reduce the workload of physicians and save time and expense by automatically generating reports.
no code implementations • 2 Oct 2019 • Wangjin Lee, Hyeryun Park, Jooyoung Yoon, Kyeongmo Kim, Jinwook Choi
With a neural sequence generation model, this study aims to develop a method of writing the patient clinical texts given a brief medical history.