Title: On The Landscape of Spoken Language Models: A Comprehensive Survey
Authors: Siddhant Arora, Kai-Wei Chang, Chung-Ming Chien, Yifan Peng, Haibin Wu, Yossi Adi, Emmanuel Dupoux, Hung-Yi Lee, Karen Livescu, Shinji Watanabe
Published: 11th April 2025 (Friday) @ 13:40:53
Link: http://arxiv.org/abs/2504.08528v1
Abstract
The field of spoken language processing is undergoing a shift from training custom-built, task-specific models toward using and optimizing spoken language models (SLMs) which act as universal speech processing systems. This trend is similar to the progression toward universal language models that has taken place in the field of (text) natural language processing. SLMs include both âpureâ language models of speech â models of the distribution of tokenized speech sequences â and models that combine speech encoders with text language models, often including both spoken and written input or output. Work in this area is very diverse, with a range of terminology and evaluation settings. This paper aims to contribute an improved understanding of SLMs via a unifying literature survey of recent work in the context of the evolution of the field. Our survey categorizes the work in this area by model architecture, training, and evaluation choices, and describes some key challenges and directions for future work.
Notes
- âAdditionally, byte pair encoding (BPE) (Gage, 1994) is also sometimes applied to the discrete token sequences (Wu et al., 2023a; Shen et al., 2024) to capture recurring patterns.â
- Wav2Seq Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
- Acoustic BPE for Speech Generation with Discrete Tokens
- they do exactly what we propose to do with BPE
- except they use - presumably single-codepoint - Chinese Unicode characters to represent tokens; and
- use HuBERTâs final layer and a smaller codebook size; IIRC
- âŠbut they donât look at ASR; they focus instead on âsyntax capturingâ
- basically tested by scrambling the words in a sentence and generating synthetically with TTS and doing a classification between these nonsense utterances and correct, grammatical utterances, if I understood correctly
- need to understand
- their rescoring method (some kind of decoding thing like the guys in the Sardine lab work on a lot �); and
- the entropy informativeness point they make
- they do exactly what we propose to do with BPE