no code implementations • 31 Oct 2023 • Fen Fang, Yi Cheng, Ying Sun, Qianli Xu
In this report, we present our approach to the EPIC-KITCHENS VISOR Hand Object Segmentation Challenge, which focuses on the estimation of the relation between the hands and the objects given a single frame as input.
1 code implementation • 14 Sep 2023 • Fen Fang, Yun Liu, Ali Koksal, Qianli Xu, Joo-Hwee Lim
The introduced mask acts akin to a task-oriented attention filter, enabling the diffusion/denoising process to concentrate on a subset of action types.
no code implementations • 13 Jul 2023 • Yi Cheng, Ziwei Xu, Fen Fang, Dongyun Lin, Hehe Fan, Yongkang Wong, Ying Sun, Mohan Kankanhalli
Our research focuses on the innovative application of a differentiable logic loss in the training to leverage the co-occurrence relations between verb and noun, as well as the pre-trained Large Language Models (LLMs) to generate the logic rules for the adaptation to unseen action labels.
no code implementations • 29 Jan 2023 • Yi Cheng, Dongyun Lin, Fen Fang, Hao Xuan Woon, Qianli Xu, Ying Sun
In this report, we present the technical details of our submission to the EPIC-KITCHENS-100 Unsupervised Domain Adaptation (UDA) Challenge for Action Recognition 2022.
no code implementations • 3 Jun 2022 • Yi Cheng, Fen Fang, Ying Sun
Based on an existing method for video domain adaptation, i. e., TA3N, we propose to learn hand-centric features by leveraging the hand bounding box information for UDA on fine-grained action recognition.
no code implementations • 24 May 2022 • Qianli Xu, Nicolas Gauthier, Wenyu Liang, Fen Fang, Hui Li Tan, Ying Sun, Yan Wu, Liyuan Li, Joo-Hwee Lim
When deploying a robot to a new task, one often has to train it to detect novel objects, which is time-consuming and labor-intensive.
1 code implementation • NeurIPS 2021 • Qianli Xu, Fen Fang, Ana Molino, Vigneshwaran Subbaraju, Joo-Hwee Lim
In this study, we investigate factors that affect event memorability according to a cued recall process.