Leros: Learning Explicit Reasoning on Synthesized Data for Commonsense Question Answering
文献类型:会议论文
作者 | Wang, Chenhao2,3![]() ![]() ![]() ![]() ![]() |
出版日期 | 2024 |
会议日期 | 2024-5 |
会议地点 | Torino, Italia |
英文摘要 | Recent work shows large language models can be prompted to generate useful rationales for commonsense question answering (CQA), which can improve the performance of both themselves and other models. However, the cost of deployment and further tuning is relatively expensive for the large models. Some work explores to distill the the rationale-generation ability to convenient small-sized models, yet it typically requires human-authored QA instances during the distillation. In this paper, we propose a novel framework that leverages both knowledge graphs and large language models to synthesize rationale-augmented CQA data. Based on it, we train Leros, a model that can generate helpful rationales to assist generic QA models to accomplish unseen CQA tasks. Empirical results demonstrate Leros can substantially enhance the performance of QA models on five unseen CQA benchmarks, providing better gains than both same-sized counterpart models trained with downstream data and 10x larger language models. Our work reveals a novel way to integrate knowledge from both knowledge graphs and large language models into smaller models. The codes and synthesized resources are publicly available at https://github.com/wchrepo/leros. |
源URL | [http://ir.ia.ac.cn/handle/173211/56702] ![]() |
专题 | 复杂系统认知与决策实验室 |
通讯作者 | Liu, Kang |
作者单位 | 1.China Merchants Bank 2.School of Artificial Intelligence, University of Chinese Academy of Sciences 3.The Laboratory of Cognition and Decision Intelligence for Complex Systems, Institute of Automation, Chinese Academy of Sciences 4.Shanghai Artificial Intelligence Laboratory |
推荐引用方式 GB/T 7714 | Wang, Chenhao,Cao, Pengfei,Li, Jiachun,et al. Leros: Learning Explicit Reasoning on Synthesized Data for Commonsense Question Answering[C]. 见:. Torino, Italia. 2024-5. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。