中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Lead ASR Models to Generalize Better Using Approximated Bias-Variance Tradeof

文献类型:会议论文

作者Wang FY(王方圆)2; Ming Hao1; Yuhai Shi1; Bo Xu2
出版日期2023
会议日期2023.11.13
会议地点changsha,China
国家日本
英文摘要

The conventional recipe for Automatic Speech Recognition (ASR) models is to 1) train multiple checkpoints on a training set while relying on a validation set to prevent over fitting using early stopping and 2) average several last checkpoints or that of the lowest validation losses to obtain the final model. In this paper, we rethink and update the early stopping and checkpoint averaging from the perspective of the bias-variance tradeoff. Theoretically, the bias and variance represent the fitness and variability of a model and the tradeoff of them determines the overall generalization error. But, it’s impractical to evaluate them precisely. As an alternative, we take the training loss and validation loss as proxies of bias and variance and guide the early stopping and checkpoint averaging using their tradeoff, namely an Approximated Bias-Variance Tradeoff  ApproBiVT). When evaluating with advanced ASR models, our recipe provides 2.5%–3.7% and 3.1%–4.6% CER reduction on the AISHELL-1 and AISHELL-2, respectively (The code and sampled unaugmented training sets used in this paper will be public available on GitHub).

源URL[http://ir.ia.ac.cn/handle/173211/57379]  
专题数字内容技术与服务研究中心_听觉模型与认知计算
作者单位1.广播科学院互联网所
2.中国科学院自动化研究所
推荐引用方式
GB/T 7714
Wang FY,Ming Hao,Yuhai Shi,et al. Lead ASR Models to Generalize Better Using Approximated Bias-Variance Tradeof[C]. 见:. changsha,China. 2023.11.13.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。