Search Results for author: Qinglin Zhang

Found 27 papers, 10 papers with code

AISafetyLab: A Comprehensive Framework for AI Safety Evaluation and Improvement

2 code implementations24 Feb 2025 Zhexin Zhang, Leqi Lei, Junxiao Yang, Xijie Huang, Yida Lu, Shiyao Cui, Renmiao Chen, Qinglin Zhang, Xinyuan Wang, Hao Wang, Hao Li, Xianqi Lei, Chengwei Pan, Lei Sha, Hongning Wang, Minlie Huang

As AI models are increasingly deployed across diverse real-world scenarios, ensuring their safety remains a critical yet underexplored challenge.

Uni-Retrieval: A Multi-Style Retrieval Framework for STEM's Education

no code implementations9 Feb 2025 Yanhao Jia, Xinyi Wu, Hao Li, Qinglin Zhang, Yuxiao Hu, Shuai Zhao, Wenqi Fan

In this paper, we propose a diverse expression retrieval task tailored to educational scenarios, supporting retrieval based on multiple query styles and expressions.

Image Retrieval Language Modeling +2

OmniFlatten: An End-to-end GPT Model for Seamless Voice Conversation

1 code implementation23 Oct 2024 Qinglin Zhang, Luyao Cheng, Chong Deng, Qian Chen, Wen Wang, Siqi Zheng, Jiaqing Liu, Hai Yu, Chaohong Tan, Zhihao Du, Shiliang Zhang

However, achieving low latency and natural interactions in full-duplex dialogue systems remains a significant challenge, especially considering human conversation dynamics such as interruptions, backchannels, and overlapping speech.

Large Language Model Spoken Dialogue Systems

Recording for Eyes, Not Echoing to Ears: Contextualized Spoken-to-Written Conversion of ASR Transcripts

no code implementations19 Aug 2024 Jiaqing Liu, Chong Deng, Qinglin Zhang, Shilin Zhou, Qian Chen, Hai Yu, Wen Wang

To improve readability, we propose a Contextualized Spoken-to-Written conversion (CoS2W) task to address ASR and grammar errors and also transfer the informal text into the formal style with content preserved, utilizing contexts and auxiliary information.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Multimodal Fusion and Coherence Modeling for Video Topic Segmentation

no code implementations1 Aug 2024 Hai Yu, Chong Deng, Qinglin Zhang, Jiaqing Liu, Qian Chen, Wen Wang

In this work, we improve supervised VTS by thoroughly exploring multimodal fusion and multimodal coherence modeling.

Contrastive Learning Scene Segmentation +2

Skip-Layer Attention: Bridging Abstract and Detailed Dependencies in Transformers

no code implementations17 Jun 2024 Qian Chen, Wen Wang, Qinglin Zhang, Siqi Zheng, Shiliang Zhang, Chong Deng, Hai Yu, Jiaqing Liu, Yukun Ma, Chong Zhang

The Transformer architecture has significantly advanced deep learning, particularly in natural language processing, by effectively managing long-range dependencies.

Diversity Language Modeling +1

Risk Taxonomy, Mitigation, and Assessment Benchmarks of Large Language Model Systems

no code implementations11 Jan 2024 Tianyu Cui, Yanling Wang, Chuanpu Fu, Yong Xiao, Sijia Li, Xinhao Deng, Yunpeng Liu, Qinglin Zhang, Ziyi Qiu, Peiyang Li, Zhixing Tan, Junwu Xiong, Xinyu Kong, Zujie Wen, Ke Xu, Qi Li

Based on this, we propose a comprehensive taxonomy, which systematically analyzes potential risks associated with each module of an LLM system and discusses the corresponding mitigation strategies.

Language Modeling Language Modelling +1

Loss Masking Is Not Needed in Decoder-only Transformer for Discrete-token-based ASR

1 code implementation8 Nov 2023 Qian Chen, Wen Wang, Qinglin Zhang, Siqi Zheng, Shiliang Zhang, Chong Deng, Yukun Ma, Hai Yu, Jiaqing Liu, Chong Zhang

We find that applying the conventional cross-entropy loss on input speech tokens does not consistently improve the ASR performance over the Loss Masking approach.

Decoder

Improving Long Document Topic Segmentation Models With Enhanced Coherence Modeling

1 code implementation18 Oct 2023 Hai Yu, Chong Deng, Qinglin Zhang, Jiaqing Liu, Qian Chen, Wen Wang

Our approach improve $F_1$ of old SOTA by 3. 42 (73. 74 -> 77. 16) and reduces $P_k$ by 1. 11 points (15. 0 -> 13. 89) on WIKI-727K and achieves an average relative reduction of 4. 3% on $P_k$ on WikiSection.

Information Retrieval Segmentation +3

Improving BERT with Hybrid Pooling Network and Drop Mask

no code implementations14 Jul 2023 Qian Chen, Wen Wang, Qinglin Zhang, Chong Deng, Ma Yukun, Siqi Zheng

Transformer-based pre-trained language models, such as BERT, achieve great success in various natural language understanding tasks.

Language Modeling Language Modelling +3

Advancing Precise Outline-Conditioned Text Generation with Task Duality and Explicit Outline Control

no code implementations23 May 2023 Yunzhe Li, Qian Chen, Weixiang Yan, Wen Wang, Qinglin Zhang, Hari Sundaram

Furthermore, we identify an issue of imbalanced utilization of the outline information in the precise outline-conditioned generation, which is ubiquitously observed across fine-tuned models and zero-shot inference models.

Sentence Text Generation

Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings

1 code implementation18 May 2023 Qian Chen, Wen Wang, Qinglin Zhang, Siqi Zheng, Chong Deng, Hai Yu, Jiaqing Liu, Yukun Ma, Chong Zhang

Prior studies diagnose the anisotropy problem in sentence representations from pre-trained language models, e. g., BERT, without fine-tuning.

Language Modeling Language Modelling +5

Meeting Action Item Detection with Regularized Context Modeling

no code implementations27 Mar 2023 Jiaqing Liu, Chong Deng, Qinglin Zhang, Qian Chen, Wen Wang

We construct and release the first Chinese meeting corpus with manual action item annotations.

Contrastive Learning

Overview of the ICASSP 2023 General Meeting Understanding and Generation Challenge (MUG)

no code implementations24 Mar 2023 Qinglin Zhang, Chong Deng, Jiaqing Liu, Hai Yu, Qian Chen, Wen Wang, Zhijie Yan, Jinglin Liu, Yi Ren, Zhou Zhao

ICASSP2023 General Meeting Understanding and Generation Challenge (MUG) focuses on prompting a wide range of spoken language processing (SLP) research on meeting transcripts, as SLP applications are critical to improve users' efficiency in grasping important information in meetings.

Extractive Summarization Keyphrase Extraction

MUG: A General Meeting Understanding and Generation Benchmark

1 code implementation24 Mar 2023 Qinglin Zhang, Chong Deng, Jiaqing Liu, Hai Yu, Qian Chen, Wen Wang, Zhijie Yan, Jinglin Liu, Yi Ren, Zhou Zhao

To prompt SLP advancement, we establish a large-scale general Meeting Understanding and Generation Benchmark (MUG) to benchmark the performance of a wide range of SLP tasks, including topic segmentation, topic-level and session-level extractive summarization and topic title generation, keyphrase extraction, and action item detection.

Extractive Summarization Keyphrase Extraction +1

DopplerBAS: Binaural Audio Synthesis Addressing Doppler Effect

no code implementations14 Dec 2022 Jinglin Liu, Zhenhui Ye, Qian Chen, Siqi Zheng, Wen Wang, Qinglin Zhang, Zhou Zhao

Recently, binaural audio synthesis (BAS) has emerged as a promising research field for its applications in augmented and virtual realities.

Audio Synthesis

Sequence Model with Self-Adaptive Sliding Window for Efficient Spoken Document Segmentation

1 code implementation20 Jul 2021 Qinglin Zhang, Qian Chen, YaLi Li, Jiaqing Liu, Wen Wang

Evaluations are conducted on the English Wiki-727K document segmentation benchmark, a Chinese Wikipedia-based document segmentation dataset we created, and an in-house Chinese spoken document dataset.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

Discriminative Self-training for Punctuation Prediction

no code implementations21 Apr 2021 Qian Chen, Wen Wang, Mengzhe Chen, Qinglin Zhang

Punctuation prediction for automatic speech recognition (ASR) output transcripts plays a crucial role for improving the readability of the ASR transcripts and for improving the performance of downstream natural language processing applications.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Pre-training for Spoken Language Understanding with Joint Textual and Phonetic Representation Learning

no code implementations21 Apr 2021 Qian Chen, Wen Wang, Qinglin Zhang

In this paper, we propose a novel joint textual-phonetic pre-training approach for learning spoken language representations, aiming at exploring the full potentials of phonetic information to improve SLU robustness to ASR errors.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +7

An Application of ASP Theories of Intentions to Understanding Restaurant Scenarios: Insights and Narrative Corpus

no code implementations30 Sep 2018 Qinglin Zhang, Chris Benton, Daniela Inclezan

This paper presents a practical application of Answer Set Programming to the understanding of narratives about restaurants.

An ASP Methodology for Understanding Narratives about Stereotypical Activities

no code implementations26 Apr 2018 Daniela Inclezan, Qinglin Zhang, Marcello Balduccini, Ankush Israney

We exemplify the application of this methodology by answering questions about a number of restaurant stories.

Question Answering

Cannot find the paper you are looking for? You can Submit a new open access paper.