On Robust Trimming of Bayesian Network Classifiers YooJung Choi and - - PowerPoint PPT Presentation

on robust trimming of bayesian network classifiers
SMART_READER_LITE
LIVE PREVIEW

On Robust Trimming of Bayesian Network Classifiers YooJung Choi and - - PowerPoint PPT Presentation

On Robust Trimming of Bayesian Network Classifiers YooJung Choi and Guy Van den Broeck UCLA Bayesian Network Classifiers Class Latent Test 1 Test 2 Test 3 Test 4 Features Bayesian Network Classifiers Class Latent Test 1 Test 2 Test


slide-1
SLIDE 1

On Robust Trimming of Bayesian Network Classifiers

YooJung Choi and Guy Van den Broeck UCLA

slide-2
SLIDE 2

Bayesian Network Classifiers

Class Latent Test 1 Test 2 Test 3 Test 4

Features

slide-3
SLIDE 3

Bayesian Network Classifiers

Class Latent Test 1 Test 2 Test 3 Test 4

Features

slide-4
SLIDE 4

Bayesian Network Classifiers

Class Latent Test 1 Test 2 Test 3 Test 4

Features

Can we make the same classifications with fewer features?

slide-5
SLIDE 5

Why Classification Similarity?

To preserve classification behavior on individual examples

  • Fairness
  • Deployed classifiers
slide-6
SLIDE 6

How to measure Similarity?

What is the expected probability that a classifier α will agree with its trimming β?

“Expected Classification Agreement”

slide-7
SLIDE 7

Robust Trimming

Trimmed classifier Original classifier

Similarity

slide-8
SLIDE 8

Trimming Algorithm

“Maximum Achievable Agreement” Feature subset selection Search Objective function

slide-9
SLIDE 9
  • Branch-and-Bound search

Trimming Algorithm

slide-10
SLIDE 10
  • Branch-and-Bound search
  • Need a bound for MAA

to prune subtrees

Trimming Algorithm

slide-11
SLIDE 11

Upper-bound for MAA

“Maximum Potential Agreement”

Maximum agreement between α and a hypothetical function that maps f’ to c

slide-12
SLIDE 12

Maximum Potential Agreement

  • 1. Upper-bounds the MAA
  • 2. Monotonically increasing

Great for pruning!

slide-13
SLIDE 13

Maximum Potential Agreement

  • 1. Upper-bounds the MAA
  • 2. Monotonically increasing
  • 3. Generally easier to compute than MAA
  • 4. Equal to MAA given some independence

condition (e.g. Naïve Bayes)

Great for pruning!

slide-14
SLIDE 14

𝐸 Pr 𝑆1 = + 𝐸) + 0.7 − 0.2

Computing the MPA and MAA

D

R1 R2 AC 𝑄

1 ⟺ ¬𝐸⋀¬𝑆1

𝑄2 ⟺ ¬𝐸⋀¬𝑆1 𝑄3 ⟺ ¬𝐸⋀¬𝑆1 𝑄

4 ⟺ ¬𝐸⋀¬𝑆1

𝑥 𝑄

1 = 0.7

𝑥(𝑄2) = 0.3 𝑥 𝑄3 = 0.2 𝑥(𝑄

4) = 0.8

𝑥 𝑚 = 1.0 for all other literal 𝑚

Pr(𝐸 = +) 0.2

∙∙∙ Prior works based on knowledge compilation

[Oztok,Choi,Darwiche 2016; C,Darwiche,VdB 2017]

slide-15
SLIDE 15

Evaluation

slide-16
SLIDE 16

Evaluation

Branch-and-bound improves efficiency (even with extra upper-bound computations)

slide-17
SLIDE 17

Evaluation

High information gain does not lead to high classification agreement Information-theoretic measures unaware of changes in classification threshold

slide-18
SLIDE 18

Thank you!

Questions?