Search Results for author: Tzu-Jui Julius Wang

Found 4 papers, 1 papers with code

PiTL: Cross-modal Retrieval with Weakly-supervised Vision-language Pre-training via Prompting

no code implementations14 Jul 2023 Zixin Guo, Tzu-Jui Julius Wang, Selen Pehlivan, Abduljalil Radman, Jorma Laaksonen

To further reduce the amount of supervision, we propose Prompts-in-The-Loop (PiTL) that prompts knowledge from large language models (LLMs) to describe images.

Cross-Modal Retrieval Object +1

Learning by Hallucinating: Vision-Language Pre-training with Weak Supervision

no code implementations24 Oct 2022 Tzu-Jui Julius Wang, Jorma Laaksonen, Tomas Langer, Heikki Arponen, Tom E. Bishop

Moreover, in other V-L downstream tasks considered, our WFH models are on par with models trained with paired V-L data, revealing the utility of unpaired data.

Cross-Modal Retrieval Image Retrieval +3

CLIP4IDC: CLIP for Image Difference Captioning

1 code implementation1 Jun 2022 Zixin Guo, Tzu-Jui Julius Wang, Jorma Laaksonen

Different from directly fine-tuning CLIP to generate sentences, we introduce an adaptation training process to adapt CLIP's visual encoder to capture and align differences in image pairs based on the textual descriptions.

Domain Adaptation Image Classification

Tackling the Unannotated: Scene Graph Generation with Bias-Reduced Models

no code implementations18 Aug 2020 Tzu-Jui Julius Wang, Selen Pehlivan, Jorma Laaksonen

Recent scene graph generation (SGG) models have shown their capability of capturing the most frequent relations among visual entities.

Graph Generation Scene Graph Generation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.