knowledge tracing machines factorization machines for
play

Knowledge Tracing Machines: Factorization Machines for Knowledge - PowerPoint PPT Presentation

Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Knowledge Tracing Machines: Factorization Machines for Knowledge Tracing Jill-Jnn Vie Hisashi Kashima KJMLW, February 22, 2019


  1. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Knowledge Tracing Machines: Factorization Machines for Knowledge Tracing Jill-Jênn Vie Hisashi Kashima KJMLW, February 22, 2019 https://arxiv.org/abs/1811.03388

  2. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Practical intro When exercises are too easy/difficult, students get bored/discouraged. To personalize assessment, ⇒ need a model of how people respond to exercises. Example To personalize this presentation, ⇒ need a model of how people respond to my slides.

  3. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Practical intro When exercises are too easy/difficult, students get bored/discouraged. To personalize assessment, → need a model of how people respond to exercises. Example To personalize this presentation, → need a model of how people respond to my slides.

  4. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Practical intro When exercises are too easy/difficult, students get bored/discouraged. To personalize assessment, → need a model of how people respond to exercises. Example To personalize this presentation, → need a model of how people respond to my slides. p(understanding) Practical: 0.9 Theoretical: 0.6

  5. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Theoretical intro Let us assume x is sparse.

  6. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Theoretical intro Let us assume x is sparse. Linear regression y = � w , x � Logistic regression y = σ ( � w , x � ) where σ is sigmoid. Neural network x ( L +1) = σ ( � w , x ( L ) � ) where σ is ReLU. What if σ : x �→ x 2 for example?

  7. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Theoretical intro Let us assume x is sparse. Linear regression y = � w , x � Logistic regression y = σ ( � w , x � ) where σ is sigmoid. Neural network x ( L +1) = σ ( � w , x ( L ) � ) where σ is ReLU. What if σ : x �→ x 2 for example? Polynomial kernel y = σ (1 + � w , x � ) where σ is a monomial. Factorization machine y = � w , x � + || V x || 2 Mathieu Blondel, Masakazu Ishihata, Akinori Fujino, and Naonori Ueda (2016). “Polynomial networks and factorization machines: new insights and efficient training algorithms”. In: Proceedings of the 33rd International Conference on International Conference on Machine Learning-Volume 48 . JMLR. org, pp. 850–858

  8. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Practical intro When exercises are too easy/difficult, students get bored/discouraged. To personalize assessment, → need a model of how people respond to exercises. Example To personalize this presentation, → need a model of how people respond to my slides.

  9. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Practical intro When exercises are too easy/difficult, students get bored/discouraged. To personalize assessment, → need a model of how people respond to exercises. Example To personalize this presentation, → need a model of how people respond to my slides. p(understanding) Practical: 0.9 Theoretical: 0.9

  10. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Students try exercises Math Learning Items 5 – 5 = ? New student ◦

  11. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Students try exercises Math Learning Items 5 – 5 = ? 17 – 3 = ? New student ◦ ◦

  12. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Students try exercises Math Learning Items 5 – 5 = ? 17 – 3 = ? 13 – 7 = ? New student ◦ ◦ ×

  13. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Students try exercises Math Learning Items 5 – 5 = ? 17 – 3 = ? 13 – 7 = ? ◦ ◦ New student × Language Learning

  14. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Students try exercises Math Learning Items 5 – 5 = ? 17 – 3 = ? 13 – 7 = ? ◦ ◦ New student × Language Learning Challenges Users can attempt a same item multiple times Users learn over time People can make mistakes that do not reflect their knowledge

  15. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Predicting student performance: knowledge tracing Data A population of users answering items Events: “User i answered item j correctly/incorrectly” Side information If we know the skills required to solve each item e.g., + , × Class ID, school ID, etc. Goal: classification problem Predict the performance of new users on existing items\ Metric: AUC Method Learn parameters of questions from historical data e.g., difficulty Measure parameters of new students e.g., expertise

  16. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Existing work Original Model Basically AUC Bayesian Knowledge Tracing Hidden Markov Model 0.67 (Corbett and Anderson 1994)

  17. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Existing work Original Model Basically AUC Bayesian Knowledge Tracing Hidden Markov Model 0.67 (Corbett and Anderson 1994) Deep Knowledge Tracing Recurrent Neural Network 0.86 (Piech et al. 2015)

  18. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Existing work Original Fixed Model Basically AUC AUC Bayesian Knowledge Tracing Hidden Markov Model 0.67 0.63 (Corbett and Anderson 1994) Deep Knowledge Tracing Recurrent Neural Network 0.86 0.75 (Piech et al. 2015)

  19. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Existing work Original Fixed Model Basically AUC AUC Bayesian Knowledge Tracing Hidden Markov Model 0.67 0.63 (Corbett and Anderson 1994) Deep Knowledge Tracing Recurrent Neural Network 0.86 0.75 (Piech et al. 2015) Item Response Theory (Rasch 1960) Online Logistic Regression 0.76 (Wilson et al., 2016)

  20. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Existing work Original Fixed Model Basically AUC AUC Bayesian Knowledge Tracing Hidden Markov Model 0.67 0.63 (Corbett and Anderson 1994) Deep Knowledge Tracing Recurrent Neural Network 0.86 0.75 (Piech et al. 2015) Item Response Theory (Rasch 1960) Online Logistic Regression 0.76 (Wilson et al., 2016) PFA ≤ DKT ≤ IRT ≤ KTM ���� ���� ���� ���� LogReg LSTM LogReg FM

  21. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Limitations and contributions Several models for knowledge tracing were developed independently In our paper, we prove that our approach is more generic Our contributions Knowledge Tracing Machines unify most existing models Encoding student data to sparse features Then running logistic regression or factorization machines Better models found It is better to estimate a bias per item, not only per skill Side information improves performance more than higher dim.

  22. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Our small dataset user item correct 1 1 1 User 1 answered Item 1 correct 1 2 0 User 1 answered Item 2 incorrect 2 1 0 User 2 answered Item 1 incorrect 2 1 1 User 2 answered Item 1 correct 2 2 ??? User 2 answered Item 2 ??? dummy.csv

  23. 0 1 1 0 0 1 0 0 1 1 0 1 1 0 1 0 1 1 0 1 0 0 0 0 0 1 1 0 1 1 0 0 1 0 0 0 0 0 0 1 1 0 1 0 0 1 1 2 0 2 0 0 1 1 0 1 0 0 1 0 0 1 0 0 2 1 1 0 2 ??? 1 1 ??? 2 1 1 3 2 0 3 2 0 2 0 Users 2 2 1 2 2 correct item user 0 0 encode IRT PFA KTM data.csv Items 0 1 0 0 0 0 0 0 0 1 1 0 1 0 1 0 0 Skills 1 0 0 0 0 0 0 0 0 0 2 1 Fails Wins 0 Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Our approach Encode data to sparse features Q 1 Q 2 Q 3 KC 1 KC 2 KC 3 KC 1 KC 2 KC 3 KC 1 KC 2 KC 3 sparse matrix X Run logistic regression or factorization machines ⇒ recover existing models or better models

  24. Introduction Knowledge Tracing Encoding existing models Knowledge Tracing Machines Results Conclusion Model 1: Item Response Theory Learn abilities θ i for each user i Learn easiness e j for each item j such that: Pr (User i Item j OK) = σ ( θ i + e j ) σ : x �→ 1 / (1 + exp( − x )) logit Pr (User i Item j OK) = θ i + e j Really popular model, used for the PISA assessment Logistic regression Learn w such that logit Pr ( x ) = � w , x � + b

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend