Search Results for author: Hui Lu

Found 20 papers, 9 papers with code

Autonomous self-evolving research on biomedical data: the DREAM paradigm

no code implementations18 Jul 2024 Luojia Deng, Yijie Wu, Yongyong Ren, Hui Lu

Utilizing a clinical dataset and two omics datasets, DREAM demonstrated its ability to raise and deepen scientific questions, with difficulty scores for clinical data questions surpassing top published articles by 5. 7% and outperforming GPT-4 and bioinformatics graduate students by 58. 6% and 56. 0%, respectively.

Investigating Decoder-only Large Language Models for Speech-to-text Translation

no code implementations3 Jul 2024 Chao-Wei Huang, Hui Lu, Hongyu Gong, Hirofumi Inaguma, Ilia Kulikov, Ruslan Mavlyutov, Sravya Popuri

Large language models (LLMs), known for their exceptional reasoning capabilities, generalizability, and fluency across diverse domains, present a promising avenue for enhancing speech-related tasks.

Decoder parameter-efficient fine-tuning +2

Snakes and Ladders: Two Steps Up for VideoMamba

1 code implementation27 Jun 2024 Hui Lu, Albert Ali Salah, Ronald Poppe

Even without extensive pre-training, our models present an increasingly attractive and efficient alternative to current transformer models.

Action Recognition Mamba +2

MMCL: Boosting Deformable DETR-Based Detectors with Multi-Class Min-Margin Contrastive Learning for Superior Prohibited Item Detection

1 code implementation5 Jun 2024 Mingyuan Li, Tong Jia, Hui Lu, Bowen Ma, Hao Wang, Dongyue Chen

Prohibited Item detection in X-ray images is one of the most effective security inspection methods. However, differing from natural light images, the unique overlapping phenomena in X-ray images lead to the coupling of foreground and background features, thereby lowering the accuracy of general object detectors. Therefore, we propose a Multi-Class Min-Margin Contrastive Learning (MMCL) method that, by clarifying the category semantic information of content queries under the deformable DETR architecture, aids the model in extracting specific category foreground information from coupled features. Specifically, after grouping content queries by the number of categories, we employ the Multi-Class Inter-Class Exclusion (MIE) loss to push apart content queries from different groups.

Contrastive Learning

Enhancing Video Transformers for Action Understanding with VLM-aided Training

no code implementations24 Mar 2024 Hui Lu, Hu Jian, Ronald Poppe, Albert Ali Salah

The FTP framework adds four feature processors that focus on specific aspects of human action in videos: action category, action components, action description, and context information.

Action Classification Action Recognition +1

MSLM-S2ST: A Multitask Speech Language Model for Textless Speech-to-Speech Translation with Speaker Style Preservation

no code implementations19 Mar 2024 Yifan Peng, Ilia Kulikov, Yilin Yang, Sravya Popuri, Hui Lu, Changhan Wang, Hongyu Gong

There have been emerging research interest and advances in speech-to-speech translation (S2ST), translating utterances from one language to another.

Decoder Language Modelling +2

TCNet: Continuous Sign Language Recognition from Trajectories and Correlated Regions

1 code implementation18 Mar 2024 Hui Lu, Albert Ali Salah, Ronald Poppe

A key challenge in continuous sign language recognition (CSLR) is to efficiently capture long-range spatial interactions over time from the video input.

Sign Language Recognition

Compensation Sampling for Improved Convergence in Diffusion Models

1 code implementation11 Dec 2023 Hui Lu, Albert Ali Salah, Ronald Poppe

We argue that the denoising process is crucially limited by an accumulation of the reconstruction error due to an initial inaccurate reconstruction of the target data.

Denoising Facial Inpainting

LLMs Learn Task Heuristics from Demonstrations: A Heuristic-Driven Prompting Strategy for Document-Level Event Argument Extraction

1 code implementation11 Nov 2023 Hanzhang Zhou, Junlang Qian, Zijian Feng, Hui Lu, Zixiao Zhu, Kezhi Mao

In this study, we investigate in-context learning (ICL) in document-level event argument extraction (EAE) to alleviate the dependency on large-scale labeled data for this task.

Event Argument Extraction In-Context Learning +2

Efficient Temporal Sentence Grounding in Videos with Multi-Teacher Knowledge Distillation

1 code implementation7 Aug 2023 Renjie Liang, Yiming Yang, Hui Lu, Li Li

To tackle this problem, we propose a novel efficient multi-teacher model (EMTM) based on knowledge distillation to transfer diverse knowledge from both heterogeneous and isomorphic networks.

Knowledge Distillation Sentence +1

DLRover-RM: Resource Optimization for Deep Recommendation Models Training in the Cloud

no code implementations4 Apr 2023 Qinlong Wang, Tingfeng Lan, Yinghao Tang, Ziling Huang, Yiheng Du, HaiTao Zhang, Jian Sha, Hui Lu, Yuanchun Zhou, Ke Zhang, Mingjie Tang

To overcome them, we introduce DLRover-RM, an elastic training framework for DLRMs designed to increase resource utilization and handle the instability of a cloud environment.

Scheduling

Private Multiparty Perception for Navigation

no code implementations2 Dec 2022 Hui Lu, Mia Chiquier, Carl Vondrick

We introduce a framework for navigating through cluttered environments by connecting multiple cameras together while simultaneously preserving privacy.

Towards High-Quality Neural TTS for Low-Resource Languages by Learning Compact Speech Representations

1 code implementation27 Oct 2022 Haohan Guo, Fenglong Xie, Xixin Wu, Hui Lu, Helen Meng

Moreover, we optimize the training strategy by leveraging more audio to learn MSMCRs better for low-resource languages.

Transfer Learning

Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation

no code implementations18 Feb 2022 Disong Wang, Songxiang Liu, Xixin Wu, Hui Lu, Lifa Sun, Xunying Liu, Helen Meng

The primary task of ASA fine-tunes the SE with the speech of the target dysarthric speaker to effectively capture identity-related information, and the secondary task applies adversarial training to avoid the incorporation of abnormal speaking patterns into the reconstructed speech, by regularizing the distribution of reconstructed speech to be close to that of reference speech with high quality.

Multi-Task Learning Speaker Verification

Stabilized Likelihood-based Imitation Learning via Denoising Continuous Normalizing Flow

no code implementations29 Sep 2021 Xin Zhang, Yanhua Li, Ziming Zhang, Christopher Brinton, Zhenming Liu, Zhi-Li Zhang, Hui Lu, Zhihong Tian

State-of-the-art imitation learning (IL) approaches, e. g, GAIL, apply adversarial training to minimize the discrepancy between expert and learner behaviors, which is prone to unstable training and mode collapse.

Denoising Imitation Learning

Channel-wise Gated Res2Net: Towards Robust Detection of Synthetic Speech Attacks

2 code implementations19 Jul 2021 Xu Li, Xixin Wu, Hui Lu, Xunying Liu, Helen Meng

This argument motivates the current work that presents a novel, channel-wise gated Res2Net (CG-Res2Net), which modifies Res2Net to enable a channel-wise gating mechanism in the connection between feature groups.

Speaker Verification

Cannot find the paper you are looking for? You can Submit a new open access paper.