中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
End-to-End Speech Translation with Knowledge Distillation

文献类型:会议论文

作者Yuchen Liu1,4; Hao Xiong2; Jiajun Zhang1,4; Zhongjun He2; Hua Wu2; Haifeng Wang2; Chengqing Zong1,3,4
出版日期2019-09
会议日期Sep. 15-19, 2019
会议地点Graz,Austria
英文摘要

End-to-end speech translation (ST), which directly translates from source language speech into target language text, has attracted intensive attentions in recent years. Compared to conventional pipeline systems, end-to-end ST model has potential benefits of lower latency, smaller model size and less error propagation. However, it is notoriously difficult to implement such model which combines automatic speech recognition (ASR) and machine translation (MT) together. In this paper, we propose a knowledge distillation approach to improve ST by transferring the knowledge from text translation. Specifically, we first train a text translation model, regarded as the teacher model, and then ST model is trained to learn the output probabilities of teacher model through knowledge distillation. Experiments on English-French Augmented LibriSpeech and English-Chinese TED corpus show that end-to-end ST is possible to implement on both similar and dissimilar language pairs. In addition, with the instruction of the teacher model, end-to end ST model can gain significant improvements by over 3.5 BLEU points. 

源URL[http://ir.ia.ac.cn/handle/173211/44410]  
专题模式识别国家重点实验室_自然语言处理
作者单位1.University of Chinese Academy of Sciences
2.Baidu Inc.
3.CAS Center for Excellence in Brain Science and Intelligence Technology
4.NLPR, Institute of Automation, Chinese Academy of Sciences
推荐引用方式
GB/T 7714
Yuchen Liu,Hao Xiong,Jiajun Zhang,et al. End-to-End Speech Translation with Knowledge Distillation[C]. 见:. Graz,Austria. Sep. 15-19, 2019.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。