End-to-End Speech Translation with Knowledge Distillation
文献类型:会议论文
作者 | Yuchen Liu1,4![]() ![]() ![]() |
出版日期 | 2019-09 |
会议日期 | Sep. 15-19, 2019 |
会议地点 | Graz,Austria |
英文摘要 | End-to-end speech translation (ST), which directly translates from source language speech into target language text, has attracted intensive attentions in recent years. Compared to conventional pipeline systems, end-to-end ST model has potential benefits of lower latency, smaller model size and less error propagation. However, it is notoriously difficult to implement such model which combines automatic speech recognition (ASR) and machine translation (MT) together. In this paper, we propose a knowledge distillation approach to improve ST by transferring the knowledge from text translation. Specifically, we first train a text translation model, regarded as the teacher model, and then ST model is trained to learn the output probabilities of teacher model through knowledge distillation. Experiments on English-French Augmented LibriSpeech and English-Chinese TED corpus show that end-to-end ST is possible to implement on both similar and dissimilar language pairs. In addition, with the instruction of the teacher model, end-to end ST model can gain significant improvements by over 3.5 BLEU points. |
源URL | [http://ir.ia.ac.cn/handle/173211/44410] ![]() |
专题 | 模式识别国家重点实验室_自然语言处理 |
作者单位 | 1.University of Chinese Academy of Sciences 2.Baidu Inc. 3.CAS Center for Excellence in Brain Science and Intelligence Technology 4.NLPR, Institute of Automation, Chinese Academy of Sciences |
推荐引用方式 GB/T 7714 | Yuchen Liu,Hao Xiong,Jiajun Zhang,et al. End-to-End Speech Translation with Knowledge Distillation[C]. 见:. Graz,Austria. Sep. 15-19, 2019. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。