LiteASR: Efficient Automatic Speech Recognition with Low-Rank Approximation
Abstract
Modern automatic speech recognition (ASR) models, such as OpenAI's Whisper, rely on deep encoder-decoder architectures, and their encoders are a critical bottleneck for efficient deployment due to high computational intensity. We introduce LiteASR, a low-rank compression scheme for ASR encoders that significantly reduces inference costs while maintaining transcription accuracy. Our approach leverages the strong low-rank properties observed in intermediate activations: by applying principal component analysis (PCA) with a small calibration dataset, we approximate linear transformations with a chain of low-rank matrix multiplications, and further optimize self-attention to work in the reduced dimension. Evaluation results show that our method can compress Whisper large-v3's encoder size by over 50%, matching Whisper medium's size with better transcription accuracy, thereby establishing a new Pareto-optimal frontier of efficiency and performance. The code of LiteASR is available at https://github.com/efeslab/LiteASR.
Community
LiteASR is a novel compression scheme for speech encoders. Link for GitHub repo: https://github.com/efeslab/LiteASR
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Prepending or Cross-Attention for Speech-to-Text? An Empirical Comparison (2025)
- On the Robust Approximation of ASR Metrics (2025)
- FireRedASR: Open-Source Industrial-Grade Mandarin Speech Recognition Models from Encoder-Decoder to LLM Integration (2025)
- Samba-ASR: State-Of-The-Art Speech Recognition Leveraging Structured State-Space Models (2025)
- Understanding Zero-shot Rare Word Recognition Improvements Through LLM Integration (2025)
- DeltaLLM: Compress LLMs with Low-Rank Deltas between Shared Weights (2025)
- Return of the Encoder: Maximizing Parameter Efficiency for SLMs (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 6
Browse 6 models citing this paperDatasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper