Search Results for author: Kun-Yu Lin

Found 12 papers, 6 papers with code

Rethinking CLIP-based Video Learners in Cross-Domain Open-Vocabulary Action Recognition

1 code implementation3 Mar 2024 Kun-Yu Lin, Henghui Ding, Jiaming Zhou, Yi-Xing Peng, Zhilin Zhao, Chen Change Loy, Wei-Shi Zheng

To answer this, we establish a CROSS-domain Open-Vocabulary Action recognition benchmark named XOV-Action, and conduct a comprehensive evaluation of five state-of-the-art CLIP-based video learners under various types of domain gaps.

Open Vocabulary Action Recognition

ActionHub: A Large-scale Action Video Description Dataset for Zero-shot Action Recognition

no code implementations22 Jan 2024 Jiaming Zhou, Junwei Liang, Kun-Yu Lin, Jinrui Yang, Wei-Shi Zheng

With the proposed ActionHub dataset, we further propose a novel Cross-modality and Cross-action Modeling (CoCo) framework for ZSAR, which consists of a Dual Cross-modality Alignment module and a Cross-action Invariance Mining module.

Action Recognition Video Description +1

Event-Guided Procedure Planning from Instructional Videos with Text Supervision

no code implementations ICCV 2023 An-Lan Wang, Kun-Yu Lin, Jia-Run Du, Jingke Meng, Wei-Shi Zheng

In this work, we focus on the task of procedure planning from instructional videos with text supervision, where a model aims to predict an action sequence to transform the initial visual state into the goal visual state.

DilateFormer: Multi-Scale Dilated Transformer for Visual Recognition

1 code implementation3 Feb 2023 Jiayu Jiao, Yu-Ming Tang, Kun-Yu Lin, Yipeng Gao, Jinhua Ma, YaoWei Wang, Wei-Shi Zheng

In this work, we explore effective Vision Transformers to pursue a preferable trade-off between the computational complexity and size of the attended receptive field.

Instance Segmentation object-detection +2

AsyFOD: An Asymmetric Adaptation Paradigm for Few-Shot Domain Adaptive Object Detection

1 code implementation CVPR 2023 Yipeng Gao, Kun-Yu Lin, Junkai Yan, YaoWei Wang, Wei-Shi Zheng

Critically, in FSDAOD, the data-scarcity in the target domain leads to an extreme data imbalance between the source and target domains, which potentially causes over-adaptation in traditional feature alignment.

object-detection Object Detection

Generating Anomalies for Video Anomaly Detection With Prompt-Based Feature Mapping

no code implementations CVPR 2023 Zuhao Liu, Xiao-Ming Wu, Dian Zheng, Kun-Yu Lin, Wei-Shi Zheng

There also exists a scene gap between virtual and real scenarios, including scene-specific anomalies (events that are abnormal in one scene but normal in another) and scene-specific attributes, such as the viewpoint of the surveillance camera.

Anomaly Detection In Surveillance Videos Video Anomaly Detection

Out-of-distribution Detection by Cross-class Vicinity Distribution of In-distribution Data

1 code implementation19 Jun 2022 Zhilin Zhao, Longbing Cao, Kun-Yu Lin

We thus improve the discriminability of a pretrained network by finetuning it with out-of-distribution samples drawn from the cross-class vicinity distribution, where each out-of-distribution input corresponds to a complementary label.

Image Classification Out-of-Distribution Detection

Supervision Adaptation Balancing In-distribution Generalization and Out-of-distribution Detection

no code implementations19 Jun 2022 Zhilin Zhao, Longbing Cao, Kun-Yu Lin

To tackle this issue, several state-of-the-art methods include adding extra OOD samples to training and assign them with manually-defined labels.

Out-of-Distribution Detection Out of Distribution (OOD) Detection

Revealing the Distributional Vulnerability of Discriminators by Implicit Generators

1 code implementation23 Aug 2021 Zhilin Zhao, Longbing Cao, Kun-Yu Lin

According to the Shannon entropy, an energy-based implicit generator is inferred from a discriminator without extra training costs.

Out of Distribution (OOD) Detection

Graph-Based High-Order Relation Modeling for Long-Term Action Recognition

no code implementations CVPR 2021 Jiaming Zhou, Kun-Yu Lin, Haoxin Li, Wei-Shi Zheng

In this paper, we propose a Graph-based High-order Relation Modeling (GHRM) module to exploit the high-order relations in the long-term actions for long-term action recognition.

Action Recognition Long-video Activity Recognition +3

Cannot find the paper you are looking for? You can Submit a new open access paper.