Search Results for author: Hui Lu

Found 15 papers, 5 papers with code

Enhancing Video Transformers for Action Understanding with VLM-aided Training

no code implementations24 Mar 2024 Hui Lu, Hu Jian, Ronald Poppe, Albert Ali Salah

The FTP framework adds four feature processors that focus on specific aspects of human action in videos: action category, action components, action description, and context information.

Action Understanding

TCNet: Continuous Sign Language Recognition from Trajectories and Correlated Regions

1 code implementation18 Mar 2024 Hui Lu, Albert Ali Salah, Ronald Poppe

A key challenge in continuous sign language recognition (CSLR) is to efficiently capture long-range spatial interactions over time from the video input.

Sign Language Recognition

Compensation Sampling for Improved Convergence in Diffusion Models

1 code implementation11 Dec 2023 Hui Lu, Albert Ali Salah, Ronald Poppe

We argue that the denoising process is crucially limited by an accumulation of the reconstruction error due to an initial inaccurate reconstruction of the target data.

Denoising Facial Inpainting

ASPEN: High-Throughput LoRA Fine-Tuning of Large Language Models with a Single GPU

1 code implementation5 Dec 2023 Zhengmao Ye, Dengchun Li, Jingqi Tian, Tingfeng Lan, Jie Zuo, Lei Duan, Hui Lu, Yexi Jiang, Jian Sha, Ke Zhang, Mingjie Tang

Transformer-based large language models (LLMs) have demonstrated outstanding performance across diverse domains, particularly when fine-turned for specific domains.

Large Language Model Scheduling

Heuristic-Driven Link-of-Analogy Prompting: Enhancing Large Language Models for Document-Level Event Argument Extraction

no code implementations11 Nov 2023 Hanzhang Zhou, Junlang Qian, Zijian Feng, Hui Lu, Zixiao Zhu, Kezhi Mao

In this study, we investigate in-context learning (ICL) in document-level event argument extraction (EAE) to alleviate the dependency on large-scale labeled data for this task.

Event Argument Extraction In-Context Learning +2

Efficient Temporal Sentence Grounding in Videos with Multi-Teacher Knowledge Distillation

no code implementations7 Aug 2023 Renjie Liang, Yiming Yang, Hui Lu, Li Li

To tackle this problem, we propose a novel efficient multi-teacher model (EMTM) based on knowledge distillation to transfer diverse knowledge from both heterogeneous and isomorphic networks.

Knowledge Distillation Sentence +1

Private Multiparty Perception for Navigation

no code implementations2 Dec 2022 Hui Lu, Mia Chiquier, Carl Vondrick

We introduce a framework for navigating through cluttered environments by connecting multiple cameras together while simultaneously preserving privacy.

Towards High-Quality Neural TTS for Low-Resource Languages by Learning Compact Speech Representations

1 code implementation27 Oct 2022 Haohan Guo, Fenglong Xie, Xixin Wu, Hui Lu, Helen Meng

Moreover, we optimize the training strategy by leveraging more audio to learn MSMCRs better for low-resource languages.

Transfer Learning

Disentangled Speech Representation Learning for One-Shot Cross-lingual Voice Conversion Using $β$-VAE

no code implementations25 Oct 2022 Hui Lu, Disong Wang, Xixin Wu, Zhiyong Wu, Xunying Liu, Helen Meng

We propose an unsupervised learning method to disentangle speech into content representation and speaker identity representation.

Disentanglement Voice Conversion

Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation

no code implementations18 Feb 2022 Disong Wang, Songxiang Liu, Xixin Wu, Hui Lu, Lifa Sun, Xunying Liu, Helen Meng

The primary task of ASA fine-tunes the SE with the speech of the target dysarthric speaker to effectively capture identity-related information, and the secondary task applies adversarial training to avoid the incorporation of abnormal speaking patterns into the reconstructed speech, by regularizing the distribution of reconstructed speech to be close to that of reference speech with high quality.

Multi-Task Learning Speaker Verification

Stabilized Likelihood-based Imitation Learning via Denoising Continuous Normalizing Flow

no code implementations29 Sep 2021 Xin Zhang, Yanhua Li, Ziming Zhang, Christopher Brinton, Zhenming Liu, Zhi-Li Zhang, Hui Lu, Zhihong Tian

State-of-the-art imitation learning (IL) approaches, e. g, GAIL, apply adversarial training to minimize the discrepancy between expert and learner behaviors, which is prone to unstable training and mode collapse.

Denoising Imitation Learning

Channel-wise Gated Res2Net: Towards Robust Detection of Synthetic Speech Attacks

2 code implementations19 Jul 2021 Xu Li, Xixin Wu, Hui Lu, Xunying Liu, Helen Meng

This argument motivates the current work that presents a novel, channel-wise gated Res2Net (CG-Res2Net), which modifies Res2Net to enable a channel-wise gating mechanism in the connection between feature groups.

Speaker Verification

Cannot find the paper you are looking for? You can Submit a new open access paper.