An Embarrassingly Simple Approach for LLM with Strong ASR Capacity Z Ma, G Yang, Y Yang, Z Gao, J Wang, Z Du, F Yu, Q Chen, S Zheng, ... arXiv preprint arXiv:2402.08846, 2024 | 7 | 2024 |
Fast-Hubert: an Efficient Training Framework for Self-Supervised Speech Representation Learning G Yang, Z Ma, Z Zheng, Y Song, Z Niu, X Chen 2023 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), 1-7, 2023 | 4 | 2023 |
Pushing the limits of unsupervised unit discovery for SSL speech representation Z Ma, Z Zheng, G Yang, Y Wang, C Zhang, X Chen arXiv preprint arXiv:2306.08920, 2023 | 4 | 2023 |
MaLa-ASR: Multimedia-Assisted LLM-Based ASR G Yang, Z Ma, F Yu, Z Gao, S Zhang, X Chen arXiv preprint arXiv:2406.05839, 2024 | 1 | 2024 |
TacoLM: GaTed Attention Equipped Codec Language Model are Efficient Zero-Shot Text to Speech Synthesizers Y Song, Z Chen, X Wang, Z Ma, G Yang, X Chen arXiv preprint arXiv:2406.15752, 2024 | | 2024 |