on discriminative learning of prediction uncertainty
play

On Discriminative Learning of Prediction Uncertainty Vojtch Franc, - PowerPoint PPT Presentation

On Discriminative Learning of Prediction Uncertainty Vojtch Franc, Daniel Pra Department of Cybernetics Faculty of Electrical Engineering Czech Technical University in Prague On Discriminative Learning of Prediction Uncertainty h ( x


  1. On Discriminative Learning of Prediction Uncertainty Vojtěch Franc, Daniel Průša Department of Cybernetics Faculty of Electrical Engineering Czech Technical University in Prague

  2. On Discriminative Learning of Prediction Uncertainty � h ( x ) 2/2 with probability c ( x ) Selective classifier: ( h, c )( x ) = reject with probability 1 − c ( x ) where h : X → Y is a classifier and c : X → [0 , 1] is a selection function

  3. On Discriminative Learning of Prediction Uncertainty � h ( x ) 2/2 with probability c ( x ) Selective classifier: ( h, c )( x ) = reject with probability 1 − c ( x ) where h : X → Y is a classifier and c : X → [0 , 1] is a selection function Example: Linear SVM h ( x ) = sign( � φ ( x ) , w � + b ) c ( x ) = [ [ |� φ ( x ) , w � + b | ≥ θ ] ]

  4. On Discriminative Learning of Prediction Uncertainty � h ( x ) 2/2 with probability c ( x ) Selective classifier: ( h, c )( x ) = reject with probability 1 − c ( x ) where h : X → Y is a classifier and c : X → [0 , 1] is a selection function 8 Example: Linear SVM SVM + distance to hyperplane h ( x ) = sign( � φ ( x ) , w � + b ) 6 selective risk [%] c ( x ) = [ [ |� φ ( x ) , w � + b | ≥ θ ] ] 4 Coverage: � � R S = 2.1% φ ( c ) = E x ∼ p c ( x ) 2 Selective risk: � � 0 ℓ ( y,h ( x )) c ( x ) E ( x,y ) ∼ p R S ( h, c ) = 0 20 40 60 80 100 φ ( x ) coverage [%]

  5. On Discriminative Learning of Prediction Uncertainty � h ( x ) 2/2 with probability c ( x ) Selective classifier: ( h, c )( x ) = reject with probability 1 − c ( x ) where h : X → Y is a classifier and c : X → [0 , 1] is a selection function 8 Example: Linear SVM SVM + distance to hyperplane h ( x ) = sign( � φ ( x ) , w � + b ) 6 selective risk [%] c ( x ) = [ [ |� φ ( x ) , w � + b | ≥ θ ] ] 4 R S = 2.1% 2 In our paper we show: 1) What is the optimal c ( x ) 0 0 20 40 60 80 100 2) How to learn c ( x ) discriminatively coverage [%]

  6. On Discriminative Learning of Prediction Uncertainty � h ( x ) 2/2 with probability c ( x ) Selective classifier: ( h, c )( x ) = reject with probability 1 − c ( x ) where h : X → Y is a classifier and c : X → [0 , 1] is a selection function 8 Example: Linear SVM SVM + distance to hyperplane SVM + learned selection function h ( x ) = sign( � φ ( x ) , w � + b ) 6 selective risk [%] c ( x ) = [ [ |� φ ( x ) , w � + b | ≥ θ ] ] 4 R S = 2.1% 2 In our paper we show: R S = 0.2% 1) What is the optimal c ( x ) 0 0 20 40 60 80 100 2) How to learn c ( x ) discriminatively coverage [%]

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend