Model-based Machine Learning Chris Bishop Microsoft Research - - PowerPoint PPT Presentation

model based machine learning
SMART_READER_LITE
LIVE PREVIEW

Model-based Machine Learning Chris Bishop Microsoft Research - - PowerPoint PPT Presentation

Model-based Machine Learning Chris Bishop Microsoft Research Cambridge Royal Society, March 2012 Traditional machine learning Logistic regression Neural networks K-means, mixture of Gaussians PCA, kernel PCA, ICA, FA Support vector machines


slide-1
SLIDE 1

Model-based Machine Learning

Chris Bishop Microsoft Research Cambridge Royal Society, March 2012

slide-2
SLIDE 2
slide-3
SLIDE 3

Traditional machine learning

Logistic regression Neural networks K-means, mixture of Gaussians PCA, kernel PCA, ICA, FA Support vector machines Deep belief networks Decision trees and random forests … many others …

slide-4
SLIDE 4

Model-based machine learning

Goal: a single modelling framework which supports a wide range of models Traditional: “how do I map my problem onto a standard algorithm”? Model-based: “what is the model that represents my problem”?

slide-5
SLIDE 5

Realisation of model-based ML

Bayesian framework Probabilistic graphical models Efficient deterministic inference

slide-6
SLIDE 6

Movie recommender demo

slide-7
SLIDE 7

Probabilistic graphical models

Maths (M) Algebra (A) Geometry (G) P(M, G, A) = P(M) P(G|M) P(A|M) Graph structure captures domain knowledge

slide-8
SLIDE 8

Efficient inference

slide-9
SLIDE 9

Local message-passing

?

Maths (M) Algebra (A) Geometry (G)

slide-10
SLIDE 10

What if distributions are intractable?

True distribution Monte Carlo Variational Message Passing Loopy belief propagation Expectation propagation

slide-11
SLIDE 11

Algorithms  Models

  • M. E. Tipping and C. M. Bishop (1997)
  • C. M. Bishop (1999)
slide-12
SLIDE 12
slide-13
SLIDE 13

Childhood Asthma

slide-14
SLIDE 14

Allergic Sensitisation Model

slide-15
SLIDE 15

Comparison with traditional ML

Separation of model and training algorithm

Auto-generated inference algorithm

Easy extension to more complex situations

Modify model, use the same inference algorithms Flexible as requirements change

Compact code

Easy to write and maintain Transparent functionality

Many traditional methods are special cases

One simple framework for newcomers to the field

slide-16
SLIDE 16

“Big data”

Computational size vs. statistical size

?

length temperature

slide-17
SLIDE 17

Noisy ranking

Conventional approach to ranking: “Elo”

single strength value for each player cannot handle teams, or more than 2 players

slide-18
SLIDE 18

Bayesian Ranking: TrueSkillTM

y12

1 2 s1 s2

  • R. Herbrich, T. Minka, and T. Graepel; NIPS (2006)
slide-19
SLIDE 19

s1 s2 s3 s4 t1

y12

t2 t3

y23

Multi-player multi-team model

slide-20
SLIDE 20

y12

1 2 s1 s2

y12

1 2 s1 s2

^ ^ ^ ^ ^

slide-21
SLIDE 21

TrueSkillTM

  • Sept. 2005;

10s of millions of users; millions of matches per day

slide-22
SLIDE 22

Convergence

5 10 15 20 25 30 35 40

Level

100 200 300 400

Number of Games

char (Elo) SQLWildman (Elo) char (TrueSkill™) SQLWildman (TrueSkill™)

slide-23
SLIDE 23

Infer.NET

1. Specify your machine learning problem as a probabilistic model in a .NET program (typically 10-20 lines of code). 2. Use Infer.NET to compile the model into optimized runtime code. 3. Run the code to make inferences

  • n your data automatically.

research.microsoft.com/infernet

slide-24
SLIDE 24

research.microsoft.com/~cmbishop

slide-25
SLIDE 25