News
Eight papers accepted at INTERSPEECH 2024
July 03, 2024
Our papers have been accepted at INTERSPEECH 2024 (external link).
Universal Score-based Speech Enhancement with High Content Preservation
Robin Scheibler, Yusuke Fujita, Yuma Shirahata, Tatsuya Komatsu
URGENT Challenge: Universality, Robustness, and Generalizability For Speech Enhancement
Wangyou Zhang (Carnegie Mellon University, Shanghai Jiao Tong University), Robin Scheibler, Kohei Saijo (Waseda University), et al.
Audio-conditioned phonemic and prosodic annotation for building text-to-speech models from unlabeled speech data
Yuma Shirahata, Byeongseon Park, Ryuichi Yamamoto, Kentaro Tachibana
LibriTTS-P: A Corpus with Speaking Style and Speaker Identity Prompts for Text-to-Speech and Style Captioning
Masaya Kawamura, Ryuichi Yamamoto, Yuma Shirahata, Takuya Hasumi, Kentaro Tachibana
Song Data Cleansing for End-to-End Neural Singer Diarization Using Neural Analysis and Synthesis Framework
Hokuto Munakata, Ryo Terashima, Yusuke Fujita
Audio Fingerprinting with Holographic Reduced Representations
Yusuke Fujita, Tatsuya Komatsu
SRC4VC: Smartphone-Recorded Corpus for Voice Conversion Benchmark
Yuki Saito (The University of Tokyo), Takuto Igarashi (The University of Tokyo), Kentaro Seki (The University of Tokyo), Shinnosuke Takamichi (The University of Tokyo, Keio University), Ryuichi Yamamoto, Kentaro Tachibana, Hiroshi Saruwatari (The University of Tokyo)
Noise-Robust Voice Conversion by Conditional Denoising Training Using Latent Variables of Recording Quality and Environment
Takuto Igarashi (The University of Tokyo), Yuki Saito (The University of Tokyo), Kentaro Seki (The University of Tokyo), Shinnosuke Takamichi (The University of Tokyo, Keio University), Ryuichi Yamamoto, Kentaro Tachibana, Hiroshi Saruwatari (The University of Tokyo)