1 code implementation • 3 Apr 2018 • Lauri Juvela, Bajibabu Bollepalli, Xin Wang, Hirokazu Kameoka, Manu Airaksinen, Junichi Yamagishi, Paavo Alku
This paper proposes a method for generating speech from filterbank mel frequency cepstral coefficients (MFCC), which are widely used in speech applications, such as ASR, but are generally considered unusable for speech synthesis.
no code implementations • 7 Apr 2018 • Xin Wang, Jaime Lorenzo-Trueba, Shinji Takaki, Lauri Juvela, Junichi Yamagishi
Recent advances in speech synthesis suggest that limitations such as the lossy nature of the amplitude spectrum with minimum phase approximation and the over-smoothing effect in acoustic modeling can be overcome by using advanced machine learning approaches.
no code implementations • 25 Apr 2018 • Lauri Juvela, Vassilis Tsiaras, Bajibabu Bollepalli, Manu Airaksinen, Junichi Yamagishi, Paavo Alku
Recent speech technology research has seen a growing interest in using WaveNets as statistical vocoders, i. e., generating speech waveforms from acoustic features.
no code implementations • 29 Oct 2018 • Bajibabu Bollepalli, Lauri Juvela, Paavo Alku
Moreover, we experiment with a WaveNet vocoder in synthesis of Lombard speech.
no code implementations • 30 Oct 2018 • Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku
The state-of-the-art in text-to-speech synthesis has recently improved considerably due to novel neural waveform generation methods, such as WaveNet.
no code implementations • 14 Mar 2019 • Bajibabu Bollepalli, Lauri Juvela, Paavo Alku
The results show that the newly proposed GANs achieve synthesis quality comparable to that of widely-used DNNs, without using an additive noise component.
1 code implementation • 8 Apr 2019 • Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku
Recent advances in neural network -based text-to-speech have reached human level naturalness in synthetic speech.
no code implementations • 27 Oct 2019 • Yi Zhao, Xin Wang, Lauri Juvela, Junichi Yamagishi
Recent neural waveform synthesizers such as WaveNet, WaveGlow, and the neural-source-filter (NSF) model have shown good performance in speech synthesis despite their different methods of waveform generation.
no code implementations • 5 Nov 2019 • Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Hector Delgado, Andreas Nautsch, Nicholas Evans, Md Sahidullah, Ville Vestman, Tomi Kinnunen, Kong Aik Lee, Lauri Juvela, Paavo Alku, Yu-Huai Peng, Hsin-Te Hwang, Yu Tsao, Hsin-Min Wang, Sebastien Le Maguer, Markus Becker, Fergus Henderson, Rob Clark, Yu Zhang, Quan Wang, Ye Jia, Kai Onuma, Koji Mushika, Takashi Kaneda, Yuan Jiang, Li-Juan Liu, Yi-Chiao Wu, Wen-Chin Huang, Tomoki Toda, Kou Tanaka, Hirokazu Kameoka, Ingmar Steiner, Driss Matrouf, Jean-Francois Bonastre, Avashna Govender, Srikanth Ronanki, Jing-Xuan Zhang, Zhen-Hua Ling
Spoofing attacks within a logical access (LA) scenario are generated with the latest speech synthesis and voice conversion technologies, including state-of-the-art neural acoustic and waveform model techniques.
no code implementations • 2 Nov 2022 • Alec Wright, Vesa Välimäki, Lauri Juvela
We propose an audio effects processing framework that learns to emulate a target electric guitar tone from a recording.
no code implementations • 2 Jun 2023 • Pablo Pérez Zarazaga, Zofia Malisz, Gustav Eje Henter, Lauri Juvela
We also find that the small set of phonetically relevant speech parameters we use is sufficient to allow for speaker-independent synthesis (a. k. a.
no code implementations • 26 Sep 2023 • Lauri Juvela, Xin Wang
Advances in neural speech synthesis have brought us technology that is not only close to human naturalness, but is also capable of instant voice cloning with little data, and is highly accessible with pre-trained models available.