Publication | Open Access
Curriculum Pre-training for End-to-End Speech Translation
99
Citations
32
References
2020
Year
Unknown Venue
EngineeringCross-lingual RepresentationMultilingualismSpoken Language ProcessingMultilingual PretrainingLanguage LearningCurriculum Pre-training MethodCorpus LinguisticsEnd-to-end Speech TranslationSpeech RecognitionNatural Language ProcessingComputational LinguisticsLanguage StudiesCurriculum Pre-trainingReal-time LanguageMachine TranslationDeep LearningNeural Machine TranslationSpeech TranslationSpeech FeaturesSpeech ProcessingLinguistics
End-to-end speech translation poses a heavy burden on the encoder because it has to transcribe, understand, and learn cross-lingual semantics simultaneously. To obtain a powerful encoder, traditional methods pre-train it on ASR data to capture speech features. However, we argue that pre-training the encoder only through simple speech recognition is not enough, and high-level linguistic knowledge should be considered. Inspired by this, we propose a curriculum pre-training method that includes an elementary course for transcription learning and two advanced courses for understanding the utterance and mapping words in two languages. The difficulty of these courses is gradually increasing. Experiments show that our curriculum pre-training method leads to significant improvements on En-De and En-Fr speech translation benchmarks.
| Year | Citations | |
|---|---|---|
2023 | 73.5K | |
2014 | 14.6K | |
2015 | 8.5K | |
2015 | 5.7K | |
2009 | 4.8K | |
2019 | 3.4K | |
2016 | 2.3K | |
2018 | 1.3K | |
2017 | 1.1K | |
2016 | 285 |
Page 1
Page 1