Search Results for author: Shinya Wada

Found 2 papers, 0 papers with code

VideoAdviser: Video Knowledge Distillation for Multimodal Transfer Learning

no code implementations27 Sep 2023 Yanan Wang, Donghuo Zeng, Shinya Wada, Satoshi Kurihara

In this work, to achieve high efficiency-performance multimodal transfer learning, we propose VideoAdviser, a video knowledge distillation method to transfer multimodal knowledge of video-enhanced prompts from a multimodal fundamental model (teacher) to a specific modal fundamental model (student).

Knowledge Distillation regression +2

VQA-GNN: Reasoning with Multimodal Knowledge via Graph Neural Networks for Visual Question Answering

no code implementations ICCV 2023 Yanan Wang, Michihiro Yasunaga, Hongyu Ren, Shinya Wada, Jure Leskovec

Visual question answering (VQA) requires systems to perform concept-level reasoning by unifying unstructured (e. g., the context in question and answer; "QA context") and structured (e. g., knowledge graph for the QA context and scene; "concept graph") multimodal knowledge.

Knowledge Graphs Question Answering +1

Cannot find the paper you are looking for? You can Submit a new open access paper.