论文标题
朝着适当的查询,键和价值计算以进行知识跟踪
Towards an Appropriate Query, Key, and Value Computation for Knowledge Tracing
论文作者
论文摘要
知识追踪是通过学习活动建模学生知识的行为,是计算机辅助教育领域的广泛研究问题。尽管具有注意机制的模型表现优于传统方法,例如贝叶斯知识追踪和协作过滤,但它们具有两个局限性。首先,这些模型依靠浅的注意层,并且随着时间的流逝,练习和反应之间的复杂关系。其次,没有广泛探索用于知识追踪的自我发项层的查询,键和值的不同组合。使用练习和互动(锻炼响应对)作为疑问和钥匙/值分别缺乏经验支持的通常实践。在本文中,我们提出了一个基于变压器的新型模型,用于知识追踪,圣徒:分开的自我攻击性神经知识追踪。 SAINT具有编码器解码器结构,其中锻炼和响应嵌入序列分别输入编码器和解码器,该序列可以多次堆叠注意力层。据我们所知,这是第一项提出一个用于知识追踪的编码模型模型的工作,该模型将深厚的自我牵键层应用于分别进行练习和响应。大规模知识追踪数据集的经验评估表明,与当前的最新模型相比,Saint在知识跟踪中实现了最新的知识跟踪表现。
Knowledge tracing, the act of modeling a student's knowledge through learning activities, is an extensively studied problem in the field of computer-aided education. Although models with attention mechanism have outperformed traditional approaches such as Bayesian knowledge tracing and collaborative filtering, they share two limitations. Firstly, the models rely on shallow attention layers and fail to capture complex relations among exercises and responses over time. Secondly, different combinations of queries, keys and values for the self-attention layer for knowledge tracing were not extensively explored. Usual practice of using exercises and interactions (exercise-response pairs) as queries and keys/values respectively lacks empirical support. In this paper, we propose a novel Transformer based model for knowledge tracing, SAINT: Separated Self-AttentIve Neural Knowledge Tracing. SAINT has an encoder-decoder structure where exercise and response embedding sequence separately enter the encoder and the decoder respectively, which allows to stack attention layers multiple times. To the best of our knowledge, this is the first work to suggest an encoder-decoder model for knowledge tracing that applies deep self-attentive layers to exercises and responses separately. The empirical evaluations on a large-scale knowledge tracing dataset show that SAINT achieves the state-of-the-art performance in knowledge tracing with the improvement of AUC by 1.8% compared to the current state-of-the-art models.