中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Probing Large Language Models from A Human Behavioral Perspective

文献类型:会议论文

作者Xintong Wang3; Xiaoyu Li2; Xingshan Li1; Chris Biemann3
出版日期2024
会议名称1st Workshop on Bridging Neurons and Symbols for Natural Language Processing and Knowledge Graphs Reasoning
会议日期2024
会议地点不详
关键词Large Language Models Interpretation and Understanding Eye-Tracking Human Behavioral
页码1-7
英文摘要

Large Language Models (LLMs) have emerged as dominant foundational models in modern NLP. However, the understanding of their prediction processes and internal mechanisms, such as feed-forward networks (FFN) and multi-head self-attention (MHSA), remains largely unexplored. In this work, we probe LLMs from a human behavioral perspective, correlating values from LLMs with eye-tracking measures, which are widely recognized as meaningful indicators of human reading patterns. Our findings reveal that LLMs exhibit a similar prediction pattern with humans but distinct from that of Shallow Language Models (SLMs). Moreover, with the escalation of LLM layers from the middle layers, the correlation coefficients also increase in FFN and MHSA, indicating that the logits within FFN increasingly encapsulate word semantics suitable for predicting tokens from the vocabulary.

收录类别EI
会议录LREC-COLING 2024 - Workshop Proceedings
语种英语
源URL[http://ir.psych.ac.cn/handle/311026/47857]  
专题中国科学院心理研究所
作者单位1.School of Computer Science and Technology, Beijing Institute of Technology
2.Institute of Psychology, Chinese Academy of Sciences
3.Department of Informatics, Universität Hamburg
推荐引用方式
GB/T 7714
Xintong Wang,Xiaoyu Li,Xingshan Li,et al. Probing Large Language Models from A Human Behavioral Perspective[C]. 见:1st Workshop on Bridging Neurons and Symbols for Natural Language Processing and Knowledge Graphs Reasoning. 不详. 2024.

入库方式: OAI收割

来源:心理研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。