Search Results for author: Takashi Miyazaki

Found 7 papers, 0 papers with code

A Visually-grounded First-person Dialogue Dataset with Verbal and Non-verbal Responses

no code implementations EMNLP 2020 Hisashi Kamezawa, Noriki Nishida, Nobuyuki Shimizu, Takashi Miyazaki, Hideki Nakayama

The results demonstrate that first-person vision helps neural network models correctly understand human intentions, and the production of non-verbal responses is a challenging task like that of verbal responses.

How do people talk about images? A study on open-domain conversations with images.

no code implementations NAACL (ACL) 2022 Yi-Pei Chen, Nobuyuki Shimizu, Takashi Miyazaki, Hideki Nakayama

This paper explores how humans conduct conversations with images by investigating an open-domain image conversation dataset, ImageChat.

Deep Learning Based Multi-modal Addressee Recognition in Visual Scenes with Utterances

no code implementations12 Sep 2018 Thao Minh Le, Nobuyuki Shimizu, Takashi Miyazaki, Koichi Shinoda

With the widespread use of intelligent systems, such as smart speakers, addressee recognition has become a concern in human-computer interaction, as more and more people expect such systems to understand complicated social scenes, including those outdoors, in cafeterias, and hospitals.

Cannot find the paper you are looking for? You can Submit a new open access paper.