Search Results for author: Yukiya Hono

Found 12 papers, 2 papers with code

Release of Pre-Trained Models for the Japanese Language

no code implementations2 Apr 2024 Kei Sawada, Tianyu Zhao, Makoto Shing, Kentaro Mitsui, Akio Kaga, Yukiya Hono, Toshiaki Wakatsuki, Koh Mitsuda

AI democratization aims to create a world in which the average person can utilize AI techniques.

PeriodGrad: Towards Pitch-Controllable Neural Vocoder Based on a Diffusion Probabilistic Model

no code implementations22 Feb 2024 Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda

This paper presents a neural vocoder based on a denoising diffusion probabilistic model (DDPM) incorporating explicit periodic signals as auxiliary conditioning signals.

Denoising Pitch control +1

An Integration of Pre-Trained Speech and Language Models for End-to-End Speech Recognition

no code implementations6 Dec 2023 Yukiya Hono, Koh Mitsuda, Tianyu Zhao, Kentaro Mitsui, Toshiaki Wakatsuki, Kei Sawada

Advances in machine learning have made it possible to perform various text and speech processing tasks, including automatic speech recognition (ASR), in an end-to-end (E2E) manner.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +5

Towards human-like spoken dialogue generation between AI agents from written dialogue

no code implementations2 Oct 2023 Kentaro Mitsui, Yukiya Hono, Kei Sawada

The advent of large language models (LLMs) has made it possible to generate natural written dialogues between two agents.

Dialogue Generation

UniFLG: Unified Facial Landmark Generator from Text or Speech

no code implementations28 Feb 2023 Kentaro Mitsui, Yukiya Hono, Kei Sawada

The two primary frameworks used for talking face generation comprise a text-driven framework, which generates synchronized speech and talking faces from text, and a speech-driven framework, which generates talking faces from speech.

Speech Synthesis Talking Face Generation

Singing Voice Synthesis Based on a Musical Note Position-Aware Attention Mechanism

no code implementations28 Dec 2022 Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda

This paper proposes a novel sequence-to-sequence (seq2seq) model with a musical note position-aware attention mechanism for singing voice synthesis (SVS).

Position Singing Voice Synthesis

Embedding a Differentiable Mel-cepstral Synthesis Filter to a Neural Speech Synthesis System

1 code implementation21 Nov 2022 Takenori Yoshimura, Shinji Takaki, Kazuhiro Nakamura, Keiichiro Oura, Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda

This paper integrates a classic mel-cepstral synthesis filter into a modern neural speech synthesis system towards end-to-end controllable speech synthesis.

Speech Synthesis

Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System

1 code implementation5 Aug 2021 Yukiya Hono, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda

To better model a singing voice, the proposed system incorporates improved approaches to modeling pitch and vibrato and better training criteria into the acoustic model.

Singing Voice Synthesis

Hierarchical Multi-Grained Generative Model for Expressive Speech Synthesis

no code implementations17 Sep 2020 Yukiya Hono, Kazuna Tsuboi, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda

This framework consists of a multi-grained variational autoencoder, a conditional prior, and a multi-level auto-regressive latent converter to obtain the different time-resolution latent variables and sample the finer-level latent variables from the coarser-level ones by taking into account the input text.

Expressive Speech Synthesis Text-To-Speech Synthesis

Cannot find the paper you are looking for? You can Submit a new open access paper.