Search Results for author: Binghuai Lin

Found 19 papers, 8 papers with code

基于熵的二语语音习得评价研究—以日本学习者习得汉语声母为例(An Entropy-based Evaluation of L2 Speech Acquisition: The Preliminary Report on Chinese Initials Produced by Japanese Learners)

no code implementations CCL 2022 Xiaoli Feng, Yingming Gao, Binghuai Lin, Jinson Zhang

“本文引入“熵”对学习者二语音素发音错误的分布情况进行了量化研究。通过对不同音素及不同二语水平学习者音素错误率和错误分散度的分析发现:1. 错误率与错误分散度有较高的相关性, 二者的差异反映出错误分布的差异性;2. 错误率类似的音素中, 与母语音素相似度越高的音素错误分散度越小;3. 较初级水平, 中级水平学习者音素错误率下降而错误分散度上升。由此可见, 熵可以在错误率基础上可以进一步揭示学习者母语音系及二语水平对音素发音错误分散度的影响。”

Making Large Language Models Better Reasoners with Alignment

no code implementations5 Sep 2023 Peiyi Wang, Lei LI, Liang Chen, Feifan Song, Binghuai Lin, Yunbo Cao, Tianyu Liu, Zhifang Sui

To address this problem, we introduce an \textit{Alignment Fine-Tuning (AFT)} paradigm, which involves three steps: 1) fine-tuning LLMs with COT training data; 2) generating multiple COT responses for each question, and categorizing them into positive and negative ones based on whether they achieve the correct answer; 3) calibrating the scores of positive and negative responses given by LLMs with a novel constraint alignment loss.

Soft Language Clustering for Multilingual Model Pre-training

no code implementations13 Jun 2023 Jiali Zeng, Yufan Jiang, Yongjing Yin, Yi Jing, Fandong Meng, Binghuai Lin, Yunbo Cao, Jie zhou

Multilingual pre-trained language models have demonstrated impressive (zero-shot) cross-lingual transfer abilities, however, their performance is hindered when the target language has distant typology from source languages or when pre-training data is limited in size.

Clustering Question Answering +5

Rhythm-controllable Attention with High Robustness for Long Sentence Speech Synthesis

no code implementations5 Jun 2023 Dengfeng Ke, Yayue Deng, Yukang Jia, Jinlong Xue, Qi Luo, Ya Li, Jianqing Sun, Jiaen Liang, Binghuai Lin

Regressive Text-to-Speech (TTS) system utilizes attention mechanism to generate alignment between text and acoustic feature sequence.

Sentence Speech Synthesis

Large Language Models are not Fair Evaluators

1 code implementation29 May 2023 Peiyi Wang, Lei LI, Liang Chen, Zefan Cai, Dawei Zhu, Binghuai Lin, Yunbo Cao, Qi Liu, Tianyu Liu, Zhifang Sui

In this paper, we uncover a systematic bias in the evaluation paradigm of adopting large language models~(LLMs), e. g., GPT-4, as a referee to score and compare the quality of responses generated by candidate models.

Language Modelling Large Language Model +1

DialogVCS: Robust Natural Language Understanding in Dialogue System Upgrade

no code implementations24 May 2023 Zefan Cai, Xin Zheng, Tianyu Liu, Xu Wang, Haoran Meng, Jiaqi Han, Gang Yuan, Binghuai Lin, Baobao Chang, Yunbo Cao

In the constant updates of the product dialogue systems, we need to retrain the natural language understanding (NLU) model as new data from the real users would be merged into the existent data accumulated in the last updates.

Intent Detection Multi-Label Classification +1

Denoising Bottleneck with Mutual Information Maximization for Video Multimodal Fusion

1 code implementation24 May 2023 Shaoxiang Wu, Damai Dai, Ziwei Qin, Tianyu Liu, Binghuai Lin, Yunbo Cao, Zhifang Sui

However, unlike other image-text multimodal tasks, video has longer multimodal sequences with more redundancy and noise in both visual and audio modalities.

Denoising Multimodal Sentiment Analysis

Enhancing Continual Relation Extraction via Classifier Decomposition

1 code implementation8 May 2023 Heming Xia, Peiyi Wang, Tianyu Liu, Binghuai Lin, Yunbo Cao, Zhifang Sui

In this work, we point out that there exist two typical biases after training of this vanilla strategy: classifier bias and representation bias, which causes the previous knowledge that the model learned to be shaded.

Continual Relation Extraction Relation

DialogQAE: N-to-N Question Answer Pair Extraction from Customer Service Chatlog

no code implementations14 Dec 2022 Xin Zheng, Tianyu Liu, Haoran Meng, Xu Wang, Yufan Jiang, Mengliang Rao, Binghuai Lin, Zhifang Sui, Yunbo Cao

Harvesting question-answer (QA) pairs from customer service chatlog in the wild is an efficient way to enrich the knowledge base for customer service chatbots in the cold start or continuous integration scenarios.

Retrieval

DualNER: A Dual-Teaching framework for Zero-shot Cross-lingual Named Entity Recognition

no code implementations15 Nov 2022 Jiali Zeng, Yufan Jiang, Yongjing Yin, Xu Wang, Binghuai Lin, Yunbo Cao

We present DualNER, a simple and effective framework to make full use of both annotated source language corpus and unlabeled target language text for zero-shot cross-lingual named entity recognition (NER).

named-entity-recognition Named Entity Recognition +1

Learning Robust Representations for Continual Relation Extraction via Adversarial Class Augmentation

1 code implementation10 Oct 2022 Peiyi Wang, YiFan Song, Tianyu Liu, Binghuai Lin, Yunbo Cao, Sujian Li, Zhifang Sui

In this paper, through empirical studies we argue that this assumption may not hold, and an important reason for catastrophic forgetting is that the learned representations do not have good robustness against the appearance of analogous relations in the subsequent learning process.

Continual Relation Extraction Relation

Text-Aware End-to-end Mispronunciation Detection and Diagnosis

1 code implementation15 Jun 2022 Linkai Peng, Yingming Gao, Binghuai Lin, Dengfeng Ke, Yanlu Xie, Jinsong Zhang

In the field of assessing the pronunciation quality of constrained speech, the given transcriptions can play the role of a teacher.

Contrastive Learning

HPT: Hierarchy-aware Prompt Tuning for Hierarchical Text Classification

1 code implementation28 Apr 2022 Zihan Wang, Peiyi Wang, Tianyu Liu, Binghuai Lin, Yunbo Cao, Zhifang Sui, Houfeng Wang

However, in this paradigm, there exists a huge gap between the classification tasks with sophisticated label hierarchy and the masked language model (MLM) pretraining tasks of PLMs and thus the potentials of PLMs can not be fully tapped.

Language Modelling Multi-Label Classification +2

Speech Enhancement using Separable Polling Attention and Global Layer Normalization followed with PReLU

no code implementations6 May 2021 Dengfeng Ke, Jinsong Zhang, Yanlu Xie, Yanyan Xu, Binghuai Lin

With all these modifications, the size of the PHASEN model is shrunk from 33M parameters to 5M parameters, while the performance on VoiceBank+DEMAND is improved to the CSIG score of 4. 30, the PESQ score of 3. 07 and the COVL score of 3. 73.

Speech Enhancement

A Full Text-Dependent End to End Mispronunciation Detection and Diagnosis with Easy Data Augmentation Techniques

1 code implementation17 Apr 2021 Kaiqi Fu, Jones Lin, Dengfeng Ke, Yanlu Xie, Jinsong Zhang, Binghuai Lin

Recently, end-to-end mispronunciation detection and diagnosis (MD&D) systems has become a popular alternative to greatly simplify the model-building process of conventional hybrid DNN-HMM systems by representing complicated modules with a single deep network architecture.

Data Augmentation

Cannot find the paper you are looking for? You can Submit a new open access paper.