In this post, I’m reviewing Making LLMs Better Many-to-Many Speech-to-Text Translators with Curriculum Learning (ACL 2025 ) without exotic training tricks. The recipe—called LLM-SRT—keeps a frozen Whisper encoder, adds a lightweight speech adapter (Q-Former + MLP), and fine-tunes in three stages: (1) ASR, (2) speech-aided MT (SMT/MMT), and