Подписаться
Tao Li
Tao Li
Audio, Speech and Language Processing Group (ASLP@NPU), School of Computer Science
Подтвержден адрес электронной почты в домене npu-aslp.org
Название
Процитировано
Процитировано
Год
Controllable emotion transfer for end-to-end speech synthesis
T Li, S Yang, L Xue, L Xie
2021 12th International Symposium on Chinese Spoken Language Processing …, 2021
1002021
Cross-speaker emotion disentangling and transfer for end-to-end speech synthesis
T Li, X Wang, Q Xie, Z Wang, L Xie
IEEE/ACM Transactions on Audio, Speech, and Language Processing 30, 1448-1460, 2022
442022
Multi-speaker expressive speech synthesis via multiple factors decoupling
X Zhu, Y Lei, K Song, Y Zhang, T Li, L Xie
ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023
212023
Enriching source style transfer in recognition-synthesis based non-parallel voice conversion
Z Wang, X Zhou, F Yang, T Li, H Du, L Xie, W Gan, H Chen, H Li
arXiv preprint arXiv:2106.08741, 2021
212021
One-shot voice conversion for style transfer based on speaker adaptation
Z Wang, Q Xie, T Li, H Du, L Xie, P Zhu, M Bi
ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and …, 2022
162022
Cross-speaker Emotion Transfer Based On Prosody Compensation for End-to-End Speech Synthesis
T Li, X Wang, Q Xie, Z Wang, M Jiang, L Xie
arXiv preprint arXiv:2207.01198, 2022
142022
Metts: Multilingual emotional text-to-speech by cross-speaker and cross-lingual emotion transfer
X Zhu, Y Lei, T Li, Y Zhang, H Zhou, H Lu, L Xie
IEEE/ACM Transactions on Audio, Speech, and Language Processing 32, 1506-1518, 2024
132024
Vec-tok speech: speech vectorization and tokenization for neural speech generation
X Zhu, Y Lv, Y Lei, T Li, W He, H Zhou, H Lu, L Xie
arXiv preprint arXiv:2310.07246, 2023
132023
Multi-speaker multi-style text-to-speech synthesis with single-speaker single-style training data scenarios
Q Xie, T Li, X Wang, Z Wang, L Xie, G Yu, G Wan
2022 13th International Symposium on Chinese Spoken Language Processing …, 2022
122022
Diclet-tts: Diffusion model based cross-lingual emotion transfer for text-to-speech—a study between english and mandarin
T Li, C Hu, J Cong, X Zhu, J Li, Q Tian, Y Wang, L Xie
IEEE/ACM Transactions on Audio, Speech, and Language Processing 31, 3418-3430, 2023
102023
Msm-vc: High-fidelity source style transfer for non-parallel voice conversion by multi-scale style modeling
Z Wang, X Wang, Q Xie, T Li, L Xie, Q Tian, Y Wang
IEEE/ACM Transactions on Audio, Speech, and Language Processing 31, 3883-3895, 2023
42023
U-Style: Cascading U-Nets With Multi-Level Speaker and Style Modeling for Zero-Shot Voice Cloning
T Li, Z Wang, X Zhu, J Cong, Q Tian, Y Wang, L Xie
IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2024
22024
HIGNN-TTS: Hierarchical Prosody Modeling With Graph Neural Networks for Expressive Long-Form TTS
D Guo, X Zhu, L Xue, T Li, Y Lv, Y Jiang, L Xie
2023 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), 1-7, 2023
22023
Towards Expressive Zero-Shot Speech Synthesis with Hierarchical Prosody Modeling
Y Jiang, T Li, F Yang, L Xie, M Meng, Y Wang
arXiv preprint arXiv:2406.05681, 2024
2024
В данный момент система не может выполнить эту операцию. Повторите попытку позднее.
Статьи 1–14