Multi-Scale Permutation Entropy for Audio Deepfake Detection
文献类型:会议论文
作者 | Chenglong Wang1; He JY(何佳毅)1![]() ![]() ![]() |
出版日期 | 2024-04-14 |
会议日期 | 2024-4-14 |
会议地点 | 韩国首尔 |
英文摘要 | With the widespread application of Automatic Speaker Verification (ASV) technology in security authentication, the threat of fake audio attacks looms as a malicious means compromising system security. In this study, we employ the multi-scale permutation entropy (MPE) in audio deepfake detection, which could help measure the complexity and detect the dynamic characteristics of audio signals at different scales. Experimental results indicate that MPE can ffectively improve the performance of LFCC. For example, on the ASVspoof2019 LA test set, it successfully achieves an equal error rate (EER) of less than 2%, which is around 50% lower than that of LFCC. Notably, MPE exhibits extraordinary generalization perfor mance when applied to the In-the-Wild dataset, as its performance of EER is comparable to that of Wav2vec, without requiring pretrain ing. Therefore, we believe that MPE holds promising prospects in voice biometric recognition for anti-spoofing applications. Our code is available at https://github.com/ADDchallenge/MPE-for-audio deepfake-detection |
源URL | [http://ir.ia.ac.cn/handle/173211/57393] ![]() |
专题 | 自动化研究所_模式识别国家重点实验室_模式分析与学习团队 |
作者单位 | 1.中国科学院自动化研究所 2.清华大学 |
推荐引用方式 GB/T 7714 | Chenglong Wang,He JY,Jiangyan Yi,et al. Multi-Scale Permutation Entropy for Audio Deepfake Detection[C]. 见:. 韩国首尔. 2024-4-14. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。