no code implementations • 15 Dec 2023 • Minxue Niu, Zhaobo Zheng, Kumar Akash, Teruhisa Misu
Humans' internal states play a key role in human-machine interaction, leading to the rise of human state estimation as a prominent field.
no code implementations • 9 Oct 2023 • Kaiwen Zhou, Kwonjoon Lee, Teruhisa Misu, Xin Eric Wang
For problems where the goal is to infer conclusions beyond image content, which we noted as visual commonsense inference (VCI), VLMs face difficulties, while LLMs, given sufficient visual evidence, can use commonsense to infer the answer well.
no code implementations • 21 Sep 2022 • Zhaobo K. Zheng, Kumar Akash, Teruhisa Misu, Vidya Krishmoorthy, Miaomiao Dong, Yuni Lee, Gaojian Huang
This work proposes identification of user driving style preference with multimodal signals, so the vehicle could match user preference in a continuous and automatic way.
no code implementations • 6 Jun 2022 • Xiaofeng Gao, Xingwei Wu, Samson Ho, Teruhisa Misu, Kumar Akash
To understand the effect of highlighting on drivers' SA for objects with different types and locations under various traffic densities, we conducted an in-person experiment with 20 participants on a driving simulator.
no code implementations • 28 Mar 2022 • Lingfeng Sun, Chen Tang, Yaru Niu, Enna Sachdeva, Chiho Choi, Teruhisa Misu, Masayoshi Tomizuka, Wei Zhan
To address these issues, we propose a novel approach to avoid KL vanishing and induce an interpretable interactive latent space with pseudo labels.
no code implementations • 15 Mar 2022 • Yuning Qiu, Teruhisa Misu, Carlos Busso
The experimental results reveal that recordings annotated with events that are likely to be anomalous, such as avoiding on-road pedestrians and traffic rule violations, have higher anomaly scores than recordings without any event annotation.
1 code implementation • 15 Jan 2022 • Kaihong Wang, Kumar Akash, Teruhisa Misu
In this work, we propose a novel paradigm that regularizes the spatiotemporal consistency by synthesizing motions in input videos with the generated optical flow instead of estimating them.
no code implementations • 28 Nov 2021 • Wenda Qin, Teruhisa Misu, Derry Wijaya
Vision-and-Language Navigation (VLN) is a challenging task in the field of artificial intelligence.
no code implementations • CVPR 2019 • Jinkyu Kim, Teruhisa Misu, Yi-Ting Chen, Ashish Tawari, John Canny
We show that taking advice improves the performance of the end-to-end network, while the network cues on a variety of visual features that are provided by advice.
no code implementations • 28 Jun 2019 • Vidyasagar Sadhu, Teruhisa Misu, Dario Pompili
In this paper, we present a novel multi-task learning based approach that leverages domain-knowledge (maneuver labels) for anomaly detection in driving data.
no code implementations • 7 Feb 2019 • Ahmed Taha, Yi-Ting Chen, Teruhisa Misu, Abhinav Shrivastava, Larry Davis
We introduce an unsupervised formulation to estimate heteroscedastic uncertainty in retrieval systems.
1 code implementation • 24 Jan 2019 • Ahmed Taha, Yi-Ting Chen, Teruhisa Misu, Abhinav Shrivastava, Larry Davis
We employ triplet loss as a feature embedding regularizer to boost classification performance.
no code implementations • 23 Jan 2019 • Ahmed Taha, Yi-Ting Chen, Xitong Yang, Teruhisa Misu, Larry Davis
We cast visual retrieval as a regression problem by posing triplet loss as a regression loss.
no code implementations • CVPR 2018 • Vasili Ramanishka, Yi-Ting Chen, Teruhisa Misu, Kate Saenko
We present the Honda Research Institute Driving Dataset (HDD), a challenging dataset to enable research on learning driver behavior in real-life environments.