no code implementations • 20 Dec 2023 • Jiachen Lian, Carly Feng, Naasir Farooqi, Steve Li, Anshul Kashyap, Cheol Jun Cho, Peter Wu, Robbie Netzorg, Tingle Li, Gopala Krishna Anumanchipalli
Dysfluent speech modeling requires time-accurate and silence-aware transcription at both the word-level and phonetic-level.
no code implementations • 13 Dec 2023 • Robin Netzorg, Ajil Jalal, Luna McNulty, Gopala Krishna Anumanchipalli
Perceptual modification of voice is an elusive goal.
no code implementations • 6 Jun 2022 • Jiachen Lian, Chunlei Zhang, Gopala Krishna Anumanchipalli, Dong Yu
We leverage recent advancements in self-supervised speech representation learning as well as speech synthesis front-end techniques for system development.
1 code implementation • 11 May 2022 • Jiachen Lian, Chunlei Zhang, Gopala Krishna Anumanchipalli, Dong Yu
In our experiment on the VCTK dataset, we demonstrate that content embeddings derived from the conditional DSVAE overcome the randomness and achieve a much better phoneme classification accuracy, a stabilized vocalization and a better zero-shot VC performance compared with the competitive DSVAE baseline.
1 code implementation • 1 Apr 2022 • Jiachen Lian, Alan W Black, Louis Goldstein, Gopala Krishna Anumanchipalli
Most of the research on data-driven speech representation learning has focused on raw audios in an end-to-end manner, paying little attention to their internal phonological or gestural structure.