Hugging Face introduces Distil-Whisper, a compact speech recognition model

TL;DR:

  • Hugging Face introduces Distil-Whisper, a compact speech recognition model for resource-constrained environments.
  • Created an open-source dataset using pseudo-labelling to develop Distil-Whisper.
  • Whisper model, pre-trained on 680,000 hours of data, forms the basis for Distil-Whisper.
  • Distil-Whisper maintains resilience in challenging acoustic conditions and mitigates errors in long-form audio.
  • Pseudo-labelling and knowledge distillation techniques applied for model compression.
  • Distil-Whisper offers a 5.8x speedup with a 51% parameter reduction and low WER.
  • Future research opportunities in audio domain knowledge distillation and model compression.

Main AI News:

In the ever-evolving landscape of AI-driven technology, Hugging Face researchers have embarked on a groundbreaking mission to bridge the gap between high-performance speech recognition and resource-constrained environments. Their solution? Introducing Distil-Whisper, a compact speech recognition model that promises to reshape the way we interact with low-resource settings.

Hugging Face’s journey began by recognizing the challenges posed by deploying large pre-trained speech recognition models in resource-constrained scenarios. Their ingenious solution was to craft a substantial open-source dataset using a novel approach—pseudo-labelling. This dataset served as the foundation for distilling a more compact version of the Whisper model, aptly named Distil-Whisper.

The Whisper model, a formidable speech recognition transformer, was initially pre-trained on a staggering 680,000 hours of noisy internet speech data. This model comprises transformer-based encoder and decoder components, showcasing competitive performance in zero-shot scenarios without fine-tuning. Enter Distil-Whisper, a nimble offspring crafted through the art of knowledge distillation, leveraging pseudo-labelling. What sets Distil-Whisper apart is its ability to uphold the Whisper model’s resilience in challenging acoustic conditions while mitigating hallucination errors during long-form audio processing. This innovative research introduces a large-scale pseudo-labelling method for speech data, a relatively unexplored yet promising avenue for knowledge distillation.

Automatic Speech Recognition (ASR) systems have made remarkable strides, achieving human-level accuracy. However, their growth is hindered by the burgeoning size of pre-trained models, particularly in resource-constrained environments. The Whisper model, with its remarkable capabilities, excels in diverse datasets but yearned for practicality in low-latency deployment. While knowledge distillation has effectively compressed NLP transformer models, its application in the realm of speech recognition remains an underexplored territory.

The proposed approach hinges on pseudo-labelling to construct a substantial open-source dataset, facilitating the art of knowledge distillation. To ensure impeccable training quality, a Word Error Rate (WER) heuristic is thoughtfully employed to select optimal pseudo-labels. The knowledge distillation objective introduces a dynamic combination of Kullback-Leibler divergence and pseudo-label terms infused with a mean-square error component to align the students’ hidden layer outputs with the teachers’. This cutting-edge distillation technique is seamlessly applied to the Whisper model within the Seq2Seq ASR framework, guaranteeing uniform transcription formatting and offering invaluable sequence-level distillation guidance.

Distil-Whisper, born from the crucible of knowledge distillation, heralds a new era of efficiency. It significantly enhances processing speed while drastically reducing parameters when compared to the original Whisper model. A remarkable 5.8x speedup accompanies a 51% parameter reduction, all while achieving a minuscule WER of less than 1% on out-of-distribution test data in a zero-shot scenario. The distil-medium.en model, although sporting a slightly higher WER, delivers a staggering 6.8x improvement in immediate inference and a remarkable 75% model compression. Notably, the Whisper model’s susceptibility to hallucination errors during long-form audio transcription is effectively addressed by Distil-Whisper, all while maintaining a competitive WER performance.

Conclusion:

Distil-Whisper’s emergence signifies a pivotal shift in the speech recognition market. Its compact design, speed, and efficiency make it a game-changer for resource-constrained environments. This innovation not only addresses current challenges but also opens up new avenues for research, positioning Hugging Face at the forefront of speech recognition technology.

Source