Publication | Closed Access
Injecting Text in Self-Supervised Speech Pretraining
25
Citations
42
References
2021
Year
Llm Fine-tuningEngineeringSpoken Language ProcessingSelf-supervised Speech PretrainingMultilingual PretrainingSpeech RecognitionNatural Language ProcessingComputational LinguisticsAutomated Speech RecognitionLanguage StudiesReal-time LanguageMachine TranslationTranscribed SpeechSpeech SynthesisSpeech OutputDeep LearningSpeech CommunicationSpeech TechnologySelf-supervised PretrainingSpeech ProcessingSpeech InputSpeech PerceptionLinguistics
Self-supervised pretraining for Automated Speech Recognition (ASR) has shown varied degrees of success. In this paper, we propose to jointly learn representations during pretraining from two different modalities: speech and text. The proposed method, tts4pretrain complements the power of contrastive learning in self-supervision with linguistic/lexical representations derived from synthesized speech, effectively learning from untranscribed speech and unspoken text. Lexical learning in the speech encoder is enforced through an additional sequence loss term that is coupled with contrastive loss during pretraining. We demonstrate that this novel pretraining method yields Word Error Rate (WER) reductions of 10% relative on the well-benchmarked, Librispeech task over a state-of-the-art baseline pretrained with wav2vec2.0 only. The proposed method also serves as an effective strategy to compensate for the lack of transcribed speech, effectively matching the performance of 5000 hours of transcribed speech with just 100 hours of transcribed speech on the AMI meeting transcription task. Finally, we demonstrate WER reductions of up to 15% on an inhouse Voice Search task over traditional pretraining. Incorporating text into encoder pretraining is complimentary to rescoring with a larger or in-domain language model, resulting in additional 6% relative reduction in WER.
| Year | Citations | |
|---|---|---|
2023 | 73.5K | |
2013 | 8.7K | |
2015 | 5.7K | |
2006 | 5.3K | |
2019 | 3.4K | |
2020 | 2.5K | |
2020 | 2.4K | |
2020 | 2.3K | |
2020 | 2.2K | |
2012 | 1.3K |
Page 1
Page 1