Proceedings of the Seventh ACM Conference on Learning @ Scale 2020
DOI: 10.1145/3386527.3405945
|View full text |Cite
|
Sign up to set email alerts
|

Towards an Appropriate Query, Key, and Value Computation for Knowledge Tracing

Abstract: Knowledge tracing, the act of modeling a student's knowledge through learning activities, is an extensively studied problem in the field of computer-aided education. Armed with attention mechanisms focusing on relevant information for target prediction, recurrent neural networks and Transformer-based knowledge tracing models have outperformed traditional approaches such as Bayesian knowledge tracing and collaborative filtering. However, the attention mechanisms of current state-of-the-art knowledge tracing mod… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
58
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 112 publications
(58 citation statements)
references
References 9 publications
0
58
0
Order By: Relevance
“…In each self-attention layer of SAKT, each query is an exercise embedding vector, and key and value are interaction embedding vectors. SAINT [2] is the first Transformer based knowledge tracing model which leverages encoder-decoder architecture composed of stacked self-attention layers. Unlike SAKT, SAINT gets separated streams of exercises and responses as inputs where a sequence of exercises are fed to the encoder, and a sequence of encoder outputs and responses are fed to the decoder.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…In each self-attention layer of SAKT, each query is an exercise embedding vector, and key and value are interaction embedding vectors. SAINT [2] is the first Transformer based knowledge tracing model which leverages encoder-decoder architecture composed of stacked self-attention layers. Unlike SAKT, SAINT gets separated streams of exercises and responses as inputs where a sequence of exercises are fed to the encoder, and a sequence of encoder outputs and responses are fed to the decoder.…”
Section: Related Workmentioning
confidence: 99%
“…In this subsection, we give a brief review of SAINT, a Separated Self-AttentIve Neural Knowledge Tracing. We refer the paper [2] for those who want to lean detailed aspects of SAINT. SAINT is a knowledge tracing model based on Transformer [25] The most fundamental part of SAINT is a multi-head attention layer.…”
Section: Saint+ 41 Saint: Separated Self-attentive Neural Knowledge Tracingmentioning
confidence: 99%
See 1 more Smart Citation
“…However, unlike the single matrix in MANN, DKVMN uses two matrices a key matrix for latent concepts and value matrix for student mastery level. More recently, transformers have also been applied to train DKTs [225][226][227]. Since they employ a self-attention mechanism, information encoded by transformers may also be interpreted.…”
Section: Other Applications Of Deep Neural Network Inmentioning
confidence: 99%
“…Random forests [174,197,199], k-nearest neighbours [204], Neural networks [177,200,[208][209][210]219], Bayesian networks [197,210,214], Regression models [174,177,178,188,198,205,206,213,216,217], Nave Bayes [174,214], Rule-based systems [197,207,209,221], Decision trees [174, 183, 196-198, 210, 214], Correlational analysis [220], Support vector machines [198,214], Matrix factorization & collaborative filtering [200-203, 211, 212], Cox proportional hazard model [215] Deep knowledge tracing -Deep knowledge tracing [31,222,223], Memory-augmented neural networks [224], Transformers [225][226][227] Chapter 6…”
Section: Video Watching Behaviourmentioning
confidence: 99%