Singing voice synthesis based on generative adversarial networks Y Hono, K Hashimoto, K Oura, Y Nankaku, K Tokuda ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019 | 71 | 2019 |
Sinsy: A deep neural network-based singing voice synthesis system Y Hono, K Hashimoto, K Oura, Y Nankaku, K Tokuda IEEE/ACM Transactions on Audio, Speech, and Language Processing 29, 2803-2815, 2021 | 41 | 2021 |
Recent development of the DNN-based singing voice synthesis system—Sinsy Y Hono, S Murata, K Nakamura, K Hashimoto, K Oura, Y Nankaku, ... 2018 Asia-Pacific Signal and Information Processing Association Annual …, 2018 | 41 | 2018 |
Hierarchical multi-grained generative model for expressive speech synthesis Y Hono, K Tsuboi, K Sawada, K Hashimoto, K Oura, Y Nankaku, ... ISCA Interspeech 2020, 3441-3445, 2020 | 30 | 2020 |
End-to-end text-to-speech based on latent representation of speaking styles using spontaneous dialogue K Mitsui, T Zhao, K Sawada, Y Hono, Y Nankaku, K Tokuda arXiv preprint arXiv:2206.12040, 2022 | 21 | 2022 |
PeriodNet: A non-autoregressive waveform generation model with a structure separating periodic and aperiodic components Y Hono, S Takaki, K Hashimoto, K Oura, Y Nankaku, K Tokuda ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and …, 2021 | 18 | 2021 |
An integration of pre-trained speech and language models for end-to-end speech recognition Y Hono, K Mitsuda, T Zhao, K Mitsui, T Wakatsuki, K Sawada arXiv preprint arXiv:2312.03668, 2023 | 13 | 2023 |
Release of pre-trained models for the Japanese language K Sawada, T Zhao, M Shing, K Mitsui, A Kaga, Y Hono, T Wakatsuki, ... arXiv preprint arXiv:2404.01657, 2024 | 11 | 2024 |
Towards human-like spoken dialogue generation between AI agents from written dialogue K Mitsui, Y Hono, K Sawada arXiv preprint arXiv:2310.01088, 2023 | 11 | 2023 |
PeriodNet: A non-autoregressive raw waveform generative model with a structure separating periodic and aperiodic components Y Hono, S Takaki, K Hashimoto, K Oura, Y Nankaku, K Tokuda IEEE Access 9, 137599-137612, 2021 | 9 | 2021 |
Embedding a differentiable mel-cepstral synthesis filter to a neural speech synthesis system T Yoshimura, S Takaki, K Nakamura, K Oura, Y Hono, K Hashimoto, ... ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023 | 7 | 2023 |
PSLM: Parallel Generation of Text and Speech with LLMs for Low-Latency Spoken Dialogue Systems K Mitsui, K Mitsuda, T Wakatsuki, Y Hono, K Sawada arXiv preprint arXiv:2406.12428, 2024 | 5 | 2024 |
UniFLG: Unified facial landmark generator from text or speech K Mitsui, Y Hono, K Sawada arXiv preprint arXiv:2302.14337, 2023 | 5 | 2023 |
PeriodGrad: Towards Pitch-Controllable Neural Vocoder Based on a Diffusion Probabilistic Model Y Hono, K Hashimoto, Y Nankaku, K Tokuda ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and …, 2024 | 2 | 2024 |
Singing Voice Synthesis Based on a Musical Note Position-Aware Attention Mechanism Y Hono, K Hashimoto, Y Nankaku, K Tokuda ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023 | 2 | 2023 |
Singing voice synthesis based on frame-level sequence-to-sequence models considering vocal timing deviation M Nishihara, Y Hono, K Hashimoto, Y Nankaku, K Tokuda arXiv preprint arXiv:2301.02262, 2023 | 2 | 2023 |
Singing Voice Conversion Using Posted Waveform Data on Music social media K Senda, Y Hono, K Sawada, K Hashimoto, K Oura, Y Nankaku, K Tokuda 2018 Asia-Pacific Signal and Information Processing Association Annual …, 2018 | 2 | 2018 |
Singing voice synthesis based on a frame-driven attention mechanism considering vocal timing deviation M Nishihara, Y Hono, K Hashimoto, Y Nankaku, K Tokuda IEICE Technical Report; IEICE Tech. Rep. 122 (389), 19-24, 2023 | | 2023 |
A study on vocal timing modeling for sequence-to-sequence singing voice synthesis M NISHIHARA, Y HONO, KEI HASHIMOTO, Y NANKAKU, K TOKUDA 日本音響学会研究発表会講演論文集 (CD-ROM) 2022, 3-19, 2022 | | 2022 |
深層ニューラルネットワークに基づく歌声合成のための音響・波形モデリング Y Hono, Y NANKAKU, K HASHIMOTO | | 2022 |