no code implementations • 2 Apr 2024 • Haven Kim, Taketo Akama
In musical compositions that include vocals, lyrics significantly contribute to artistic expression.
no code implementations • 9 Jan 2024 • Zhe Zhang, Taketo Akama
GANStrument, exploiting GANs with a pitch-invariant feature extractor and instance conditioning technique, has shown remarkable capabilities in synthesizing realistic instrument sounds.
no code implementations • 16 Dec 2023 • Gakusei Sato, Taketo Akama
To tackle this issue, we propose a transcription model that does not require any MIDI-audio paired data through the utilization of scalable synthetic audio for pre-training and adversarial domain confusion using unannotated real audio.
1 code implementation • 10 Jul 2023 • Keisuke Toyama, Taketo Akama, Yukara Ikemiya, Yuhta Takida, Wei-Hsiang Liao, Yuki Mitsufuji
This is especially helpful when determining the precise onset and offset for each note in the polyphonic piano content.
no code implementations • 15 Apr 2023 • Taketo Akama, Hiroaki Kitano, Katsuhiro Takematsu, Yasushi Miyajima, Natalia Polouliakh
In the realm of music information retrieval, similarity-based retrieval and auto-tagging serve as essential components.
no code implementations • 10 Nov 2022 • Gaku Narita, Junichi Shimizu, Taketo Akama
In addition, we introduce an adversarial training scheme for a pitch-invariant feature extractor that significantly improves the pitch accuracy and timbre consistency.
no code implementations • 23 Nov 2021 • Taketo Akama
We propose a contextual latent space model (CLSM) in order for users to be able to explore subsequence generation with a sense of direction in the generation space, e. g., interpolation, as well as exploring variations -- semantically similar possible subsequences.