no code implementations • 4 Oct 2024 • Sen Fang, Sizhou Chen, Yalin Feng, Xiaofeng Zhang, Teik Toe Teoh
This paper presents an innovative approach called BGTAI to simplify multimodal understanding by utilizing gloss-based annotation as an intermediate step in aligning Text and Audio with Images.
no code implementations • 17 May 2024 • Sen Fang, Lei Wang, Ce Zheng, Chunyu Sui, Mingyu Zhao, Yapeng Tian, Chen Chen
In this paper, we propose SignLLM, a multilingual Sign Language Production (SLP) large language model, which includes two novel multilingual SLP modes MLSF and Prompt2LangGloss that allow sign language gestures generation from query texts input and question-style prompts input respectively.
1 code implementation • 31 Jan 2024 • Benoit Baudry, Khashayar Etemadi, Sen Fang, Yogya Gamage, Yi Liu, Yuxin Liu, Martin Monperrus, Javier Ron, André Silva, Deepika Tiwari
The results show that LLMs can successfully generate realistic test data generators in a wide range of domains at all three levels of integrability.
1 code implementation • 25 Dec 2023 • André Silva, Sen Fang, Martin Monperrus
This results in RepairLLaMA producing a highly effective `program repair adapter' for fixing bugs with AI.
1 code implementation • 26 Sep 2023 • Zimin Chen, Sen Fang, Martin Monperrus
Software optimization refines programs for resource efficiency while preserving functionality.
no code implementations • 14 Sep 2023 • Sizhou Chen, Songyang Gao, Sen Fang
The Transformer architecture has proven to be highly effective for Automatic Speech Recognition (ASR) tasks, becoming a foundational component for a plethora of research in the domain.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 30 Aug 2023 • Sen Fang, Chunyu Sui, Yanghao Zhou, Xuedong Zhang, Hongbin Zhong, Minyu Zhao, Yapeng Tian, Chen Chen
In this paper, we propose a dual-condition diffusion pre-training model named SignDiff that can generate human sign language speakers from a skeleton pose.
no code implementations • 29 Jul 2023 • Sen Fang, Bowen Gao, Yangjian Wu, Teik Toe Teoh
Multimodal large models have been recognized for their advantages in various performance and downstream tasks.
no code implementations • 8 Mar 2023 • Sen Fang, Yangjian Wu, Bowen Gao, Jingwen Cai, Teik Toe Teoh
Recently, researchers have gradually realized that in some cases, the self-supervised pre-training on large-scale Internet data is better than that of high-quality/manually labeled data sets, and multimodal/large models are better than single or bimodal/small models.
2 code implementations • 19 Nov 2022 • Youwei Huang, Sen Fang, Jianwen Li, Jiachun Tao, Bin Hu, Tao Zhang
In recent years, research in software security has concentrated on identifying vulnerabilities in smart contracts to prevent significant losses of crypto assets on blockchains.