no code implementations • 14 Dec 2023 • Yafei Hu, Quanting Xie, Vidhi Jain, Jonathan Francis, Jay Patrikar, Nikhil Keetha, Seungchan Kim, Yaqi Xie, Tianyi Zhang, Shibo Zhao, Yu Quan Chong, Chen Wang, Katia Sycara, Matthew Johnson-Roberson, Dhruv Batra, Xiaolong Wang, Sebastian Scherer, Zsolt Kira, Fei Xia, Yonatan Bisk
Motivated by the impressive open-set performance and content generation capabilities of web-scale, large-capacity pre-trained models (i. e., foundation models) in research fields such as Natural Language Processing (NLP) and Computer Vision (CV), we devote this survey to exploring (i) how these existing foundation models from NLP and CV can be applied to the field of robotics, and also exploring (ii) what a robotics-specific foundation model would look like.
no code implementations • 27 May 2022 • Lucy Nwosu, Xiangfang Li, Lijun Qian, Seungchan Kim, Xishuang Dong
However, prediction uncertainty of deep learning models for these tasks, which is very important to safety-critical applications like medical image processing, has not been comprehensively investigated.
1 code implementation • 3 Dec 2021 • Bowen Li, Chen Wang, Pranay Reddy, Seungchan Kim, Sebastian Scherer
Few-shot object detection has attracted increasing attention and rapidly progressed in recent years.
1 code implementation • 18 Nov 2021 • Chen Wang, Yuheng Qiu, Wenshan Wang, Yafei Hu, Seungchan Kim, Sebastian Scherer
Instead, we propose to develop a method that automatically adapts online to the environment to report interesting scenes quickly.
no code implementations • 27 Feb 2021 • Lucy Nwosu, Xiangfang Li, Lijun Qian, Seungchan Kim, Xishuang Dong
Coronavirus disease 2019 (COVID-19) is an ongoing global pandemic in over 200 countries and territories, which has resulted in a great public health concern across the international community.
no code implementations • 30 May 2019 • Kavosh Asadi, Dipendra Misra, Seungchan Kim, Michel L. Littman
In this paper, we address the compounding-error problem by introducing a multi-step model that directly outputs the outcome of executing a sequence of actions.
Model-based Reinforcement Learning reinforcement-learning +1