arXiv Open Access 2022

HuBERT-EE: Early Exiting HuBERT for Efficient Speech Recognition

Ji Won Yoon Beom Jun Woo Nam Soo Kim
Lihat Sumber

Abstrak

Pre-training with self-supervised models, such as Hidden-unit BERT (HuBERT) and wav2vec 2.0, has brought significant improvements in automatic speech recognition (ASR). However, these models usually require an expensive computational cost to achieve outstanding performance, slowing down the inference speed. To improve the model efficiency, we introduce an early exit scheme for ASR, namely HuBERT-EE, that allows the model to stop the inference dynamically. In HuBERT-EE, multiple early exit branches are added at the intermediate layers. When the intermediate prediction of the early exit branch is confident, the model stops the inference, and the corresponding result can be returned early. We investigate the proper early exiting criterion and fine-tuning strategy to effectively perform early exiting. Experimental results on the LibriSpeech show that HuBERT-EE can accelerate the inference of the HuBERT while simultaneously balancing the trade-off between the performance and the latency.

Topik & Kata Kunci

Penulis (3)

J

Ji Won Yoon

B

Beom Jun Woo

N

Nam Soo Kim

Format Sitasi

Yoon, J.W., Woo, B.J., Kim, N.S. (2022). HuBERT-EE: Early Exiting HuBERT for Efficient Speech Recognition. https://arxiv.org/abs/2204.06328

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2022
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓