no code implementations • 4 Dec 2024 • Jiaxuan Liu, Zhaoci Liu, Yajun Hu, Yingying Gao, Shilei Zhang, ZhenHua Ling
Human speech exhibits rich and flexible prosodic variations.
no code implementations • 19 Sep 2024 • Keying Zuo, Qingtian Xu, Jie Zhang, ZhenHua Ling
Brain-assisted speech enhancement (BASE) aims to extract the target speaker in complex multi-talker scenarios using electroencephalogram (EEG) signals as an assistive modality, as the auditory attention of the listener can be decoded from electroneurographic signals of the brain.
no code implementations • 12 Jun 2024 • Hengyu Li, Kangdi Mei, Zhaoci Liu, Yang Ai, Liping Chen, Jie Zhang, ZhenHua Ling
It was shown in literature that speech representations extracted by self-supervised pre-trained models exhibit similarities with brain activations of human for speech perception and fine-tuning speech representation models on downstream tasks can further improve the similarity.
no code implementations • 22 Jan 2024 • Shihao Chen, Liping Chen, Jie Zhang, KongAik Lee, ZhenHua Ling, LiRong Dai
For validation, we employ the open-source pre-trained YourTTS model for speech generation and protect the target speaker's speech in the white-box scenario.
no code implementations • 6 May 2023 • Beiduo Chen, Shaohan Huang, Zihan Zhang, Wu Guo, ZhenHua Ling, Haizhen Huang, Furu Wei, Weiwei Deng, Qi Zhang
Besides, two self-correction courses are proposed to bridge the chasm between the two encoders by creating a "correction notebook" for secondary-supervision.
no code implementations • 14 Sep 2022 • Yukun Peng, ZhenHua Ling
In our experiments, our proposed method effectively improved the intelligibility and naturalness of multilingual speech synthesis comparing with the baseline meta-learning synthesis method.
no code implementations • 2 Mar 2022 • Pengyu Cheng, ZhenHua Ling
In this paper, we propose a method of speaker adaption with intuitive prosodic features for statistical parametric speech synthesis.
no code implementations • 1 Mar 2022 • Lu Dong, ZhenHua Ling, Qiang Ling, Zefeng Lai
Then, based on the estimated student vectors, the probabilistic part of DINA can be modified to a student dependent model that the slip and guess rates are related to student vectors.
no code implementations • 9 Oct 2021 • Cheng Gong, Longbiao Wang, ZhenHua Ling, Ju Zhang, Jianwu Dang
The end-to-end speech synthesis model can directly take an utterance as reference audio, and generate speech from the text with prosody and speaker characteristics similar to the reference audio.
1 code implementation • SEMEVAL 2021 • Boyuan Zheng, Xiaoyu Yang, Yu-Ping Ruan, ZhenHua Ling, Quan Liu, Si Wei, Xiaodan Zhu
Given a passage and the corresponding question, a participating system is expected to choose the correct answer from five candidates of abstract concepts in a cloze-style machine reading comprehension setup.
no code implementations • COLING 2020 • Zhiqiang Guo, Zhaoci Liu, ZhenHua Ling, Shijin Wang, Lingjing Jin, Yunxia Li
Finally, a best detection accuracy of 81. 6{\%} is obtained by our proposed methods on the Mandarin AD corpus.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Lei Zhang, Runze Wang, Jingbo Zhou, Jingsong Yu, ZhenHua Ling, Hui Xiong
Continuous efforts have been devoted to language understanding (LU) for conversational queries with the fast and wide-spread popularity of voice assistants.