ET-805 Bayesian Knowledge Tracing Ramkumar.Rajendran@iitb.ac.in - - PowerPoint PPT Presentation

et 805 bayesian knowledge tracing
SMART_READER_LITE
LIVE PREVIEW

ET-805 Bayesian Knowledge Tracing Ramkumar.Rajendran@iitb.ac.in - - PowerPoint PPT Presentation

ET-805 Bayesian Knowledge Tracing Ramkumar.Rajendran@iitb.ac.in Activity - TPS Think individually and write down Bayes Theorem? (2 minutes) Pair with your neighbour and check your answer. If both are same write down why it is correct?


slide-1
SLIDE 1

ET-805 Bayesian Knowledge Tracing

Ramkumar.Rajendran@iitb.ac.in

slide-2
SLIDE 2

Activity - TPS

2

Think individually and write down Bayes’ Theorem? (2 minutes) Pair with your neighbour and check your answer. If both are same write down why it is correct? Else, discuss why the formula that you wrote is correct (3 minutes) Share to the class

slide-3
SLIDE 3

Activity - Class Response

3

slide-4
SLIDE 4

Bayes’ Theorem

  • Relating conditional probabilities
  • How probable is one event given that the other event
  • ccured
  • Example: Two events a) Passing the exam and b) Studying

hard

  • What is the probability that a student passed an exam given

that s/he studied for the exam? P (P/S) = P (P∩S)/ P(S)

4

Pass Study

slide-5
SLIDE 5

Bayes’ Theorem

P (P/S) = P (P∩S)/ P(S) (from last slide) Similarly P (S/P) = P (P∩S)/ P(P) P (P∩S)= P (P/S). P(S) = P (S/P). P(P) = P (P/S) = P (S/P). P(P) / P(S) Bayes’ Theorem: P (A/B) = P(B/A). P(A)/ P (B)

5

Pass Study

slide-6
SLIDE 6

Activity - Quiz

6

A couple has two children, one of which is a boy. What is the probability that they have two boys? P(A) = Both kids are boys = ¼ P(B) = one of the kids is boy = ¾ P (A/B) = P(B/A). P(A)/P(B) = 1. ¼ / ¾ = ⅓

https://brilliant.org/wiki/bayes-theorem/

slide-7
SLIDE 7

Activity - Answer

7

A couple has two children, one of which is a boy. What is the probability that they have two boys? Define two events: P(A) = Both children are boys = ¼ P (B) = one of their children is boy = ¾ P (A/B) = P(A).P (B/A)/ P(B) = ¼ . 1/ ¾ = ⅓

https://brilliant.org/wiki/bayes-theorem/

slide-8
SLIDE 8

Bayesian Network

  • A compact representation of joint probability distributions

among set of random variables (nodes) and their conditional dependencies (arcs)

  • Directed Acyclic Graph (DAG)
  • Conditional Probability Table (CPT)

8

slide-9
SLIDE 9

Confused!

9

slide-10
SLIDE 10

Bayesian Network

10

Slept after lunch Missed the Class

slide-11
SLIDE 11

Bayesian Network

11

Slept after lunch Missed the Class Had Meeting

slide-12
SLIDE 12

Bayesian Network

12

Slept after lunch (S) Missed the Class (C) Had Meeting (M) Watch Lecture Videos (V)

slide-13
SLIDE 13

Bayesian Network

13

S C M V

Directed acyclic graph

  • Static directions
  • Cycles are not allowed

Nodes are conditionally Independent

  • Example: Node S and V are dependent but if the value of

C is known, then they will become independent

  • Similarly node S and M are independent. But if the value
  • f C is known they will be dependent
  • The above relation is called conditional independency
slide-14
SLIDE 14

Bayesian Network

14

S C M V

Conditional Probability Table Assume the values of nodes are boolean (yes or no) For Node S, How many probability values required to represent? One: Probability of Sleeping after lunch = P(S) For Node C, how many probability values are required to represent? Four: P(C/S,M) = Combination of S and M

slide-15
SLIDE 15

Bayesian Network

15

S C M V

Conditional Probability Table P(X1, …,Xn) = ∏ P (Xi | Parents(Xi))

slide-16
SLIDE 16

Activity - TPS

Bayesian Network

  • A compact representation of joint probability distributions

among set of random variables (nodes) and their conditional dependencies (arcs)

  • Directed Acyclic Graph (DAG)
  • Conditional Probability Table (CPT)

Think Why the bayesian network is compact? (2 minutes) Share

16

slide-17
SLIDE 17

Bayesian Network

17

S C M V

  • Boolean values
  • Total 4 nodes, hence we will have to store

2 (pow) 4 values to represent the fully connected network (24 - 1)

  • Number of probability values in Bayesian

Network: 1 (S) + 1 (M) + 4 (C) + 2 (V) = 8 Compact representation of joint probability distributions over all the variables in the network

slide-18
SLIDE 18

Sample Bayesian Network

18

Conati, C. (2010). Bayesian student modeling. In Advances in intelligent tutoring systems (pp. 281-299). Springer, Berlin, Heidelberg.

slide-19
SLIDE 19

Bayesian Networks as Learner Model

Wayang Outpost

  • Time spent per problem, time spent per action, average

incorrect action to model student’s attitude

19

slide-20
SLIDE 20

Activity - One minute thinking

In the previous example, which node we can update at every time instance based on student’s response?

20

slide-21
SLIDE 21

Activity - One minute thinking

In the previous example, which node we can update at every time instance based on student’s response?

  • The node concept indicates whether the student understood

the concept or not.

  • Hence, we can update the conditional probability table (CPT)
  • f Concept based on student’s response.
  • Updating CPT of Bayesian network is Dynamic Bayesian

Network (DBN)

21

slide-22
SLIDE 22

Bayesian Knowledge Tracing (BKT)

We want to update a node called concept to indicate whether the student learned a skill or not. If we have a node for student’s response: Node = Student’s Response to a question. Boolean value. 1 = correct, 0 = incorrect

22

Corbett, A. T., & Anderson, J. R. (1994). Knowledge tracing: Modeling the acquisition of procedural knowledge. User modeling and user-adapted interaction, 4(4), 253-278. Baker, R. S., Corbett, A. T., & Aleven, V. (2008, June). More accurate student modeling through contextual estimation of slip and guess probabilities in bayesian knowledge tracing. In Intelligent Tutoring Systems (pp. 406-415). Springer, Berlin, Heidelberg.

C R

slide-23
SLIDE 23

Activity - TPS

Update the P(Concept/response) = Increase if the response = 1, decrease otherwise. Think Write one drawback of the above method the update student’s skill on concept (2 minutes) Pair with your neighbour and copy their answer. Discuss and rank the drawback in the order of importance (3 minutes) Share

23

slide-24
SLIDE 24

Activity - Class Response

  • Guess
  • By mistake he might selected wrong answer
  • Silly mistakes - Slip
  • 24
slide-25
SLIDE 25

Bayesian Knowledge Tracing

  • Student might give correct answer by guess
  • Student might have understood the concept but answered

wrongly by mistake (Slip)

25

slide-26
SLIDE 26

Bayesian Knowledge Tracing

p(Lo) Initial Learning - the probability of prior knowledge p(T) Acquisition - the probability a rule will make the transition from the unlearned to the learned state following an opportunity to apply the rule p(G) Guess - the probability a student will guess correctly if a rule is in the unlearned state p(S) Slip - the probability a student will slip (make a mistake) if a rule is in the learned state

26

slide-27
SLIDE 27

BKT

Initial Probability - Prior knowledge

27

slide-28
SLIDE 28

Activity - TPS

Think How do you measure the probability of learning P(L) if the response to the question is correct Share

28

slide-29
SLIDE 29

Updating P(Learning) at State t

If the answer is correct, it should be based on student’s prob(L) at state t-1 and not slip to the total probability of answering the question.

29

slide-30
SLIDE 30

Activity - TPS

Think How do you measure the probability of learning P(L) if the response to the question is incorrect Share P(L4/ observation = incorrect) = P (L3).P(S)/(P(L3).P(S) + 1 -P(L3))

30

slide-31
SLIDE 31

Updating P(Learning) at State t

If the answer is incorrect:

31

slide-32
SLIDE 32

Updating P(Skill Mastery)

32

To update the P(Learning) for next state t+1, use the probability of learning into probability of transition

slide-33
SLIDE 33

Activity - TPS

Think list down one drawback of BKT (2 minutes) Pair and copy your neighbours answer. Arrange the drawbacks in the order of importance (3 minutes) Share

33

slide-34
SLIDE 34

Activity - Class Response

34

slide-35
SLIDE 35

BKT Explained Using Examples

http://www.cs.williams.edu/~iris/res/bkt/

35

slide-36
SLIDE 36

Next Class

Class after Next:

  • Log Data Analytics
  • What is Learning Analytics

36

slide-37
SLIDE 37

Last Activity - Muddy Points

List down

  • two important and
  • two least clear

(muddy) points from today’s class

  • https://tinyurl.com/et8

05mp

37