News

Eight papers accepted at INTERSPEECH 2024

July 03, 2024

Our papers have been accepted at INTERSPEECH 2024 (external link).

Universal Score-based Speech Enhancement with High Content Preservation
Robin Scheibler, Yusuke Fujita, Yuma Shirahata, Tatsuya Komatsu

URGENT Challenge: Universality, Robustness, and Generalizability For Speech Enhancement
Wangyou Zhang (Carnegie Mellon University, Shanghai Jiao Tong University), Robin Scheibler, Kohei Saijo (Waseda University), et al.

Audio-conditioned phonemic and prosodic annotation for building text-to-speech models from unlabeled speech data
Yuma Shirahata, Byeongseon Park, Ryuichi Yamamoto, Kentaro Tachibana

LibriTTS-P: A Corpus with Speaking Style and Speaker Identity Prompts for Text-to-Speech and Style Captioning
Masaya Kawamura, Ryuichi Yamamoto, Yuma Shirahata, Takuya Hasumi, Kentaro Tachibana

Song Data Cleansing for End-to-End Neural Singer Diarization Using Neural Analysis and Synthesis Framework
Hokuto Munakata, Ryo Terashima, Yusuke Fujita

Audio Fingerprinting with Holographic Reduced Representations
Yusuke Fujita, Tatsuya Komatsu

SRC4VC: Smartphone-Recorded Corpus for Voice Conversion Benchmark
Yuki Saito (The University of Tokyo), Takuto Igarashi (The University of Tokyo), Kentaro Seki (The University of Tokyo), Shinnosuke Takamichi (The University of Tokyo, Keio University), Ryuichi Yamamoto, Kentaro Tachibana, Hiroshi Saruwatari (The University of Tokyo)

Noise-Robust Voice Conversion by Conditional Denoising Training Using Latent Variables of Recording Quality and Environment
Takuto Igarashi (The University of Tokyo), Yuki Saito (The University of Tokyo), Kentaro Seki (The University of Tokyo), Shinnosuke Takamichi (The University of Tokyo, Keio University), Ryuichi Yamamoto, Kentaro Tachibana, Hiroshi Saruwatari (The University of Tokyo)