no code implementations • 22 Oct 2024 • Shota Onohara, Atsuyuki Miyai, Yuki Imajuku, Kazuki Egashira, Jeonghun Baek, Xiang Yue, Graham Neubig, Kiyoharu Aizawa
Accelerating research on Large Multimodal Models (LMMs) in non-English languages is crucial for enhancing user experiences across broader populations.
1 code implementation • 29 Mar 2024 • Atsuyuki Miyai, Jingkang Yang, Jingyang Zhang, Yifei Ming, Qing Yu, Go Irie, Yixuan Li, Hai Li, Ziwei Liu, Kiyoharu Aizawa
This paper introduces a novel and significant challenge for Vision Language Models (VLMs), termed Unsolvable Problem Detection (UPD).
1 code implementation • 2 Oct 2023 • Atsuyuki Miyai, Qing Yu, Go Irie, Kiyoharu Aizawa
We consider that such data may significantly affect the performance of large pre-trained networks because the discriminability of these OOD data depends on the pre-training algorithm.
Out-of-Distribution Detection Out of Distribution (OOD) Detection
1 code implementation • NeurIPS 2023 • Atsuyuki Miyai, Qing Yu, Go Irie, Kiyoharu Aizawa
CLIP's local features have a lot of ID-irrelevant nuisances (e. g., backgrounds), and by learning to push them away from the ID class text embeddings, we can remove the nuisances in the ID class text embeddings and enhance the separation between ID and OOD.
Out-of-Distribution Detection Out of Distribution (OOD) Detection
2 code implementations • 10 Apr 2023 • Atsuyuki Miyai, Qing Yu, Go Irie, Kiyoharu Aizawa
First, images should be collected using only the name of the ID class without training on the ID data.
1 code implementation • 23 Oct 2022 • Atsuyuki Miyai, Qing Yu, Daiki Ikami, Go Irie, Kiyoharu Aizawa
The semantics of an image can be rotation-invariant or rotation-variant, so whether the rotated image is treated as positive or negative should be determined based on the content of the image.