Search Results for author: Komei Sugiura

Found 23 papers, 8 papers with code

Polos: Multimodal Metric Learning from Human Feedback for Image Captioning

1 code implementation28 Feb 2024 Yuiga Wada, Kanta Kaneda, Daichi Saito, Komei Sugiura

Establishing an automatic evaluation metric that closely aligns with human judgments is essential for effectively developing image captioning models.

Contrastive Learning Image Captioning +1

Learning-To-Rank Approach for Identifying Everyday Objects Using a Physical-World Search Engine

1 code implementation26 Dec 2023 Kanta Kaneda, Shunya Nagashima, Ryosuke Korekata, Motonari Kambara, Komei Sugiura

Therefore, we focus on the task of retrieving target objects from open-vocabulary user instructions in a human-in-the-loop setting, which we define as the learning-to-rank physical objects (LTRPO) task.

Learning-To-Rank

DialMAT: Dialogue-Enabled Transformer with Moment-Based Adversarial Training

1 code implementation12 Nov 2023 Kanta Kaneda, Ryosuke Korekata, Yuiga Wada, Shunya Nagashima, Motonari Kambara, Yui Iioka, Haruka Matsuo, Yuto Imai, Takayuki Nishimura, Komei Sugiura

This paper focuses on the DialFRED task, which is the task of embodied instruction following in a setting where an agent can actively ask questions about the task.

Instruction Following Position

Fully Automated Task Management for Generation, Execution, and Evaluation: A Framework for Fetch-and-Carry Tasks with Natural Language Instructions in Continuous Space

no code implementations7 Nov 2023 Motonari Kambara, Komei Sugiura

This paper aims to develop a framework that enables a robot to execute tasks based on visual information, in response to natural language instructions for Fetch-and-Carry with Object Grounding (FCOG) tasks.

Multimodal Diffusion Segmentation Model for Object Segmentation from Manipulation Instructions

no code implementations17 Jul 2023 Yui Iioka, Yu Yoshida, Yuiga Wada, Shumpei Hatanaka, Komei Sugiura

In this study, we aim to develop a model that comprehends a natural language instruction (e. g., "Go to the living room and get the nearest pillow to the radio art on the wall") and generates a segmentation mask for the target everyday object.

Segmentation Semantic Segmentation +1

Prototypical Contrastive Transfer Learning for Multimodal Language Understanding

no code implementations12 Jul 2023 Seitaro Otsuki, Shintaro Ishikawa, Komei Sugiura

Most conventional models have been trained on real-world datasets that are labor-intensive to collect, and they have not fully leveraged simulation data through a transfer learning framework.

Transfer Learning

Relational Future Captioning Model for Explaining Likely Collisions in Daily Tasks

1 code implementation19 Jul 2022 Motonari Kambara, Komei Sugiura

Domestic service robots that support daily tasks are a promising solution for elderly or disabled people.

Text Generation

Moment-based Adversarial Training for Embodied Language Comprehension

1 code implementation2 Apr 2022 Shintaro Ishikawa, Komei Sugiura

This is challenging because the robot needs to break down the instruction sentences into subgoals and execute them in the correct order.

Target-dependent UNITER: A Transformer-Based Multimodal Language Comprehension Model for Domestic Service Robots

no code implementations2 Jul 2021 Shintaro Ishikawa, Komei Sugiura

Currently, domestic service robots have an insufficient ability to interact naturally through language.

Visual Explanation using Attention Mechanism in Actor-Critic-based Deep Reinforcement Learning

no code implementations6 Mar 2021 Hidenori Itaya, Tsubasa Hirakawa, Takayoshi Yamashita, Hironobu Fujiyoshi, Komei Sugiura

A3C consists of a feature extractor that extracts features from an image, a policy branch that outputs the policy, and a value branch that outputs the state value.

Decision Making reinforcement-learning +1

CrossMap Transformer: A Crossmodal Masked Path Transformer Using Double Back-Translation for Vision-and-Language Navigation

no code implementations1 Mar 2021 Aly Magassouba, Komei Sugiura, Hisashi Kawai

Navigation guided by natural language instructions is particularly suitable for Domestic Service Robots that interacts naturally with users.

Translation Vision and Language Navigation

A Multimodal Target-Source Classifier with Attention Branches to Understand Ambiguous Instructions for Fetching Daily Objects

no code implementations23 Dec 2019 Aly Magassouba, Komei Sugiura, Hisashi Kawai

To solve such a task, we propose the multimodal target-source classifier model with attention branches (MTCM-AB), which is an extension of the MTCM.

Sentence

Multimodal Attention Branch Network for Perspective-Free Sentence Generation

no code implementations10 Sep 2019 Aly Magassouba, Komei Sugiura, Hisashi Kawai

In this paper, we address the automatic sentence generation of fetching instructions for domestic service robots.

Sentence

Cannot find the paper you are looking for? You can Submit a new open access paper.