Shifted Chunk Encoder for Transformer Based Streaming End-to-End ASR
文献类型:会议论文
作者 | Wang FY(王方圆)![]() ![]() |
出版日期 | 2022 |
会议日期 | 2022.11.28 |
会议地点 | Indore,India |
英文摘要 | Currently, there are mainly three kinds of Transformer encoder based streaming End to End (E2E) Automatic Speech Recognition (ASR) approaches, namely time-restricted methods, chunk-wise methods, and memory-based methods. Generally, all of them have limitations in aspects of linear computational complexity, global context modeling, and parallel training. In this work, we aim to build a model to take all these three advantages for streaming Transformer ASR. Particularly, we propose a shifted chunk mechanism for the chunk-wise Transformer which provides cross-chunk connections between chunks. Therefore, the global context modeling ability of chunk-wise models can be significantly enhanced while all the original merits inherited.We integrate this scheme with the chunk-wise Transformer and Conformer, and identify them as SChunk-Transformer and SChunk-Conformer, respectively. Experiments on AISHELL-1 show that the SChunk-Transformer and SChunk-Conformer can respectively achieve CER 6.43% and 5.77%. And the linear complexity makes them possible to train with large batches and infer more efficiently. Our models can significantly outperform their conventional chunk-wise counterparts, while being competitive, with only 0.22 absolute CER drop, when compared with U2 which has quadratic complexity. A better CER can be achieved if compared with existing chunkwise or memory-based methods, such as HS-DACS and MMA. Code is released. (see https://github.com/wangfangyuan/SChunk-Encoder.). |
源URL | [http://ir.ia.ac.cn/handle/173211/57382] ![]() |
专题 | 数字内容技术与服务研究中心_听觉模型与认知计算 |
通讯作者 | Wang FY(王方圆) |
作者单位 | 中国科学院自动化研究所 |
推荐引用方式 GB/T 7714 | Wang FY,Xu B. Shifted Chunk Encoder for Transformer Based Streaming End-to-End ASR[C]. 见:. Indore,India. 2022.11.28. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。