中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Shifted Chunk Encoder for Transformer Based Streaming End-to-End ASR

文献类型:会议论文

作者Wang FY(王方圆); Xu B(徐波)
出版日期2022
会议日期2022.11.28
会议地点Indore,India
英文摘要

Currently, there are mainly three kinds of Transformer encoder based streaming End to End (E2E) Automatic Speech Recognition (ASR) approaches, namely time-restricted methods, chunk-wise methods, and memory-based methods. Generally, all of them have limitations in aspects of linear computational complexity, global context modeling, and parallel training. In this work, we aim to build a model to take all these three advantages for streaming Transformer ASR. Particularly, we propose a shifted chunk mechanism for the chunk-wise Transformer which provides cross-chunk connections between chunks. Therefore, the global context modeling ability of chunk-wise models can be significantly enhanced while all the original merits inherited.We integrate this scheme with the chunk-wise Transformer and Conformer, and identify them as SChunk-Transformer and SChunk-Conformer, respectively. Experiments on AISHELL-1 show that the SChunk-Transformer and SChunk-Conformer can respectively achieve CER 6.43% and 5.77%. And the linear complexity makes them possible to train with large batches and infer more efficiently. Our models can significantly outperform their conventional chunk-wise counterparts, while being competitive, with only 0.22 absolute CER drop, when compared with U2 which has quadratic complexity. A better CER can be achieved if compared with existing chunkwise or memory-based methods, such as HS-DACS and MMA. Code is released. (see https://github.com/wangfangyuan/SChunk-Encoder.).

源URL[http://ir.ia.ac.cn/handle/173211/57382]  
专题数字内容技术与服务研究中心_听觉模型与认知计算
通讯作者Wang FY(王方圆)
作者单位中国科学院自动化研究所
推荐引用方式
GB/T 7714
Wang FY,Xu B. Shifted Chunk Encoder for Transformer Based Streaming End-to-End ASR[C]. 见:. Indore,India. 2022.11.28.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。