Self-Supervised Speech Pre-training and Representation Learning Toolkit
[ICLR 2025] SOTA discrete acoustic codec models with 40/75 tokens per second for audio language modeling
[ACL 2024] Official PyTorch code for extracting features and training downstream models with emotion2vec: Self-Supervised Pre-Training for Speech Emotion Representation
A Survey of Spoken Dialogue Models (60 pages)
LightHuBERT: Lightweight and Configurable Speech Representation Learning with Once-for-All Hidden-Unit BERT
#大语言模型#A single-layer, streaming codec model providing SOTA audio quality and discrete tokens designed for superior downstream modelability.
Official Implementation of Mockingjay in Pytorch
A mini, simple, and fast end-to-end automatic speech recognition toolkit.
#计算机科学#Causal Speech Enhancement Based on a Two-Branch Nested U-Net Architecture Using Self-Supervised Speech Embeddings
Semi-supervised spoken language understanding (SLU) via self-supervised speech and language model pretraining
音学シンポジウム2025チュートリアル「マルチモーダル大規模言語モデル入門」資料