Skip to content

Commit cf3d5aa

Browse files
authored
Create 2024-09-18-extrakt.md
1 parent bf93496 commit cf3d5aa

File tree

1 file changed

+22
-0
lines changed

1 file changed

+22
-0
lines changed

_posts/2024-09-18-extrakt.md

Lines changed: 22 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,22 @@
1+
---
2+
layout: post
3+
title: 'extraKT: Extending Context Window of Attention Based Knowledge Tracing Models via Length Extrapolation'
4+
date: 2024-09-18T16:00:00.000+00:00
5+
tags: model
6+
categories: []
7+
author: ''
8+
post_image: "/assets/images/posts/extrakt.png"
9+
post_format: ''
10+
trending: true
11+
12+
---
13+
We added extraKT into our pyKT package.
14+
15+
The link is [here](https://pykt-toolkit.readthedocs.io/en/latest/models.html#extrakt) and the API is [here](https://pykt-toolkit.readthedocs.io/en/latest/pykt.models.html#module-pykt.models.extrakt).
16+
17+
Original paper can be found at [Li X, Bai Y, Guo T, et al. "Extending Context Window of Attention Based Knowledge Tracing Models via Length Extrapolation." Proceedings of the 26th European Conference on Artificial Intelligence. 2024.]([https://www.ijcai.org/proceedings/2024/654](https://www.ecai2024.eu/programme/accepted-papers#main-track))
18+
19+
Title: extraKT: Extending Context Window of Attention Based Knowledge Tracing Models via Length Extrapolation
20+
21+
22+
Abstract: Knowledge tracing (KT) is a prediction task that aims to predict students’ future performance based on their past learning data. The rapid progress in attention mechanisms has led to the emergence of various high-performing attention based KT models. However, in online or personalized education settings, students’ varying learning paths result in different lengths of student interaction sequences, which poses a significant challenge for attention based KT models as their context window sizes are fixed during both training and prediction stages. We refer to this as the length extrapolation of KT model. In this paper, we propose extraKT to facilitate better extrapolation that learn from student interactions with a short context window and continue to perform well across various longer context window sizes at prediction stage. Specifically, we negatively bias attention scores with linearly decreasing penalties that are proportional to query-key distance, which efficiently represents short-term forgetting characteristics of student knowledge states. We conduct comprehensive and rigorous experiments on three real-world educational datasets. The results show that our extraKT model exhibits robust length extrapolation capability and outperforms state-of-the-art baseline models in terms of AUC and accuracy.

0 commit comments

Comments
 (0)