no code implementations • 22 Jan 2024 • Shihao Chen, Liping Chen, Jie Zhang, KongAik Lee, ZhenHua Ling, LiRong Dai
For validation, we employ the open-source pre-trained YourTTS model for speech generation and protect the target speaker's speech in the white-box scenario.
no code implementations • 6 May 2023 • Beiduo Chen, Shaohan Huang, Zihan Zhang, Wu Guo, ZhenHua Ling, Haizhen Huang, Furu Wei, Weiwei Deng, Qi Zhang
Besides, two self-correction courses are proposed to bridge the chasm between the two encoders by creating a "correction notebook" for secondary-supervision.
no code implementations • 14 Sep 2022 • Yukun Peng, ZhenHua Ling
In our experiments, our proposed method effectively improved the intelligibility and naturalness of multilingual speech synthesis comparing with the baseline meta-learning synthesis method.
no code implementations • 2 Mar 2022 • Pengyu Cheng, ZhenHua Ling
In this paper, we propose a method of speaker adaption with intuitive prosodic features for statistical parametric speech synthesis.
no code implementations • 1 Mar 2022 • Lu Dong, ZhenHua Ling, Qiang Ling, Zefeng Lai
Then, based on the estimated student vectors, the probabilistic part of DINA can be modified to a student dependent model that the slip and guess rates are related to student vectors.
no code implementations • 9 Oct 2021 • Cheng Gong, Longbiao Wang, ZhenHua Ling, Ju Zhang, Jianwu Dang
The end-to-end speech synthesis model can directly take an utterance as reference audio, and generate speech from the text with prosody and speaker characteristics similar to the reference audio.
1 code implementation • SEMEVAL 2021 • Boyuan Zheng, Xiaoyu Yang, Yu-Ping Ruan, ZhenHua Ling, Quan Liu, Si Wei, Xiaodan Zhu
Given a passage and the corresponding question, a participating system is expected to choose the correct answer from five candidates of abstract concepts in a cloze-style machine reading comprehension setup.
no code implementations • COLING 2020 • Zhiqiang Guo, Zhaoci Liu, ZhenHua Ling, Shijin Wang, Lingjing Jin, Yunxia Li
Finally, a best detection accuracy of 81. 6{\%} is obtained by our proposed methods on the Mandarin AD corpus.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Lei Zhang, Runze Wang, Jingbo Zhou, Jingsong Yu, ZhenHua Ling, Hui Xiong
Continuous efforts have been devoted to language understanding (LU) for conversational queries with the fast and wide-spread popularity of voice assistants.