no code implementations • 8 Feb 2025 • Aayam Shrestha, Pan Liu, German Ros, Kai Yuan, Alan Fern
This work focuses on generating realistic, physically-based human behaviors from multi-modal inputs, which may only partially specify the desired motion.
no code implementations • 10 Dec 2024 • Kai Yuan, Xiaobing Pei, Haoran Yang
To address this problem, we propose a Adversarial Attacks on High-level Semantics in Graph Neural Networks (AHSG), which is a graph structure attack model that ensures the retention of primary semantics.
no code implementations • 22 Sep 2024 • Yuxing Zhi, Yuan Guo, Kai Yuan, Hesong Wang, Heng Xu, Haina Yao, Albert C Yang, Guangrui Huang, Yuping Duan
Future medical LLMs are suggested to be expert in medical domain knowledge to understand both natural languages and quantified medical data, and further make logical inferences.
no code implementations • 8 Aug 2024 • Zhou Qin, Kai Yuan, Pratik Lahiri, Wenyang Liu
In a typical e-commerce setting, Content Ranking Optimization (CRO) mechanisms are employed to surface content on the search page to fulfill customers' shopping missions.
1 code implementation • 30 May 2024 • Minttu Alakuijala, Reginald McLean, Isaac Woungang, Nariman Farsad, Samuel Kaski, Pekka Marttinen, Kai Yuan
Natural language is often the easiest and most convenient modality for humans to specify tasks for robots.
1 code implementation • 26 Mar 2024 • Kai Yuan, Christoph Bauinger, Xiangyi Zhang, Pascal Baehr, Matthias Kirchhart, Darius Dabert, Adrien Tousnakhoff, Pierre Boudier, Michael Paulitsch
We compare our approach to a similar CUDA implementation for MLPs and show that our implementation on the Intel Data Center GPU outperforms the CUDA implementation on Nvidia's H100 GPU by a factor up to 2. 84 in inference and 1. 75 in training.
no code implementations • 15 Aug 2023 • Kai Yuan, Noor Sajid, Karl Friston, Zhibin Li
We approach this problem by hierarchical generative modelling equipped with multi-level planning-for autonomous task completion-that mimics the deep temporal architecture of human motor control.
no code implementations • 28 Sep 2021 • Fernando Acero, Kai Yuan, Zhibin Li
To proactively navigate and traverse various terrains, active use of visual perception becomes indispensable.
no code implementations • 11 Aug 2021 • Kai Yuan, Da Kuang
Autocomplete (a. k. a "Query Auto-Completion", "AC") suggests full queries based on a prefix typed by customer.
no code implementations • 10 Dec 2020 • Chuanyu Yang, Kai Yuan, Qiuguo Zhu, Wanming Yu, Zhibin Li
Achieving versatile robot locomotion requires motor skills which can adapt to previously unseen situations.
no code implementations • 15 Feb 2020 • Zhaole Sun, Kai Yuan, Wenbin Hu, Chuanyu Yang, Zhibin Li
In robotic grasping, objects are often occluded in ungraspable configurations such that no pregrasp pose can be found, eg large flat boxes on the table that can only be grasped from the side.
Robotics
no code implementations • 11 Feb 2020 • Wenbin Hu, Chuanyu Yang, Kai Yuan, Zhibin Li
The performance of learned policy is evaluated on three different tasks: grasping a static target, grasping a dynamic target, and re-grasping.
Robotics
no code implementations • 7 Feb 2020 • Chuanyu Yang, Kai Yuan, Wolfgang Merkt, Taku Komura, Sethu Vijayakumar, Zhibin Li
This paper presents a hierarchical framework for Deep Reinforcement Learning that acquires motor skills for a variety of push recovery and balancing behaviors, i. e., ankle, hip, foot tilting, and stepping strategies.
no code implementations • 5 Feb 2020 • Ruoshi Wen, Kai Yuan, Qiang Wang, Shuai Heng, Zhibin Li
Regulating contact forces with high precision is crucial for grasping and manipulating fragile or deformable objects.
Robotics