1 code implementation • 7 Jan 2024 • Zane Durante, Qiuyuan Huang, Naoki Wake, Ran Gong, Jae Sung Park, Bidipta Sarkar, Rohan Taori, Yusuke Noda, Demetri Terzopoulos, Yejin Choi, Katsushi Ikeuchi, Hoi Vo, Li Fei-Fei, Jianfeng Gao
To accelerate research on agent-based multimodal intelligence, we define "Agent AI" as a class of interactive systems that can perceive visual stimuli, language inputs, and other environmentally-grounded data, and can produce meaningful embodied actions.
no code implementations • 20 Nov 2023 • Naoki Wake, Atsushi Kanehira, Kazuhiro Sasabuchi, Jun Takamatsu, Katsushi Ikeuchi
The computation starts by analyzing the videos with GPT-4V to convert environmental and action details into text, followed by a GPT-4-empowered task planner.
no code implementations • 18 Oct 2023 • Naoki Wake, Atsushi Kanehira, Kazuhiro Sasabuchi, Jun Takamatsu, Katsushi Ikeuchi
This technical report explores the ability of ChatGPT in recognizing emotions from text, which can be the basis of various applications like interactive chatbots, data annotation, and mental health analysis.
no code implementations • 7 Jul 2021 • Akihiko Sayo, Diego Thomas, Hiroshi Kawasaki, Yuta Nakashima, Katsushi Ikeuchi
We propose a new 2D pose refinement network that learns to predict the human bias in the estimated 2D pose.
Ranked #67 on 3D Human Pose Estimation on Human3.6M
1 code implementation • 27 Feb 2021 • Naoki Wake, Daichi Saito, Kazuhiro Sasabuchi, Hideki Koike, Katsushi Ikeuchi
These findings highlight the significance of object affordance in multimodal robot teaching, regardless of whether real objects are present in the images.
no code implementations • 9 Dec 2020 • Iori Yanokura, Naoki Wake, Kazuhiro Sasabuchi, Katsushi Ikeuchi, Masayuki Inaba
We propose a Learning-from-Observation framework that splits and understands a video of a human demonstration with verbal instructions to extract accurate action sequences.
no code implementations • 4 Aug 2020 • Naoki Wake, Riku Arakawa, Iori Yanokura, Takuya Kiyokawa, Kazuhiro Sasabuchi, Jun Takamatsu, Katsushi Ikeuchi
In the context of one-shot robot teaching, the contributions of the paper are: 1) to propose a framework that 1) covers various tasks in grasp-manipulation-release class household operations and 2) mimics human postures during the operations.
Robotics Human-Computer Interaction
no code implementations • 7 Jul 2018 • Ryosuke Kimura, Akihiko Sayo, Fabian Lorenzo Dayrit, Yuta Nakashima, Hiroshi Kawasaki, Ambrosio Blanco, Katsushi Ikeuchi
For full-body reconstruction with loose clothes, we propose to use lower dimensional embeddings of texture and deformation referred to as eigen-texturing and eigen-deformation, to reproduce views of even unobserved surfaces.
no code implementations • 14 Apr 2018 • Ryoichi Ishikawa, Takeshi Oishi, Katsushi Ikeuchi
In this paper, we propose a method of targetless and automatic Camera-LiDAR calibration.
1 code implementation • 13 Apr 2018 • Ryoichi Ishikawa, Takeshi Oishi, Katsushi Ikeuchi
In the experiments, we confirm the parameters obtained by two types of offline calibration according to the degree of freedom of robot movement and validate the effectiveness of online correction method by plotting localized position error during robot's intense movement.
no code implementations • CVPR 2017 • Chen Li, Stephen Lin, Kun Zhou, Katsushi Ikeuchi
An important practical feature of the proposed method is that the skin color model is utilized in a way that does not require color calibration of the camera.
no code implementations • CVPR 2017 • Chen Li, Stephen Lin, Kun Zhou, Katsushi Ikeuchi
We present a method for radiometric calibration of cameras from a single image that contains a human face.
no code implementations • CVPR 2016 • Yoshie Kobayashi, Tetsuro Morimoto, Imari Sato, Yasuhiro Mukaigawa, Takao Tomono, Katsushi Ikeuchi
First, we narrow down candidates of zenith angle by degree of polarization and determine it by the intensity of thin film which increases monotonically along the zenith angle.
no code implementations • 1 Jun 2016 • Shaodi You, Yasuyuki Matsushita, Sudipta Sinha, Yusuke Bou, Katsushi Ikeuchi
Digitally unwrapping images of paper sheets is crucial for accurate document scanning and text recognition.
no code implementations • 4 Apr 2016 • Shaodi You, Robby T. Tan, Rei Kawakami, Yasuhiro Mukaigawa, Katsushi Ikeuchi
(2) The imagery inside a water-drop is determined by the water-drop 3D shape and total reflection at the boundary.
no code implementations • CVPR 2014 • Xiangqi Huang, Bo Zheng, Takeshi Masuda, Katsushi Ikeuchi
Our feature description is designed as two steps: 1) we normalize the detected local regions to canonical shapes for robust matching; 2) we encode each key point with multiple vectors at different Morse function values.
no code implementations • CVPR 2013 • Shaodi You, Robby T. Tan, Rei Kawakami, Katsushi Ikeuchi
First, it detects raindrops based on the motion and the intensity temporal derivatives of the input video.
no code implementations • CVPR 2013 • Bo Zheng, Yibiao Zhao, Joey C. Yu, Katsushi Ikeuchi, Song-Chun Zhu
In this paper, we present an approach for scene understanding by reasoning physical stability of objects from point cloud.