Never-Ending Language Learning Tom Mitchell, William Cohen, and Many - - PowerPoint PPT Presentation

never ending language learning
SMART_READER_LITE
LIVE PREVIEW

Never-Ending Language Learning Tom Mitchell, William Cohen, and Many - - PowerPoint PPT Presentation

Never-Ending Language Learning Tom Mitchell, William Cohen, and Many Collaborators Carnegie Mellon University We will never really understand learning until we build machines that learn many different things, from years of diverse


slide-1
SLIDE 1

Tom Mitchell, William Cohen, and Many Collaborators Carnegie Mellon University

Never-Ending Language Learning

slide-2
SLIDE 2

We will never really understand learning until we build machines that

  • learn many different things,
  • from years of diverse experience,
  • in a staged, curricular fashion,
  • and become better learners over time.
slide-3
SLIDE 3

Tenet 2: Natural language understanding requires a belief system A natural language understanding system should react to text by saying either:

  • I understand, and already knew that
  • I understand, and didn’t know, but accept it
  • I understand, and disagree because …
slide-4
SLIDE 4

NELL: Never-Ending Language Learner

Inputs:

  • initial ontology (categories and relations)
  • dozen examples of each ontology predicate
  • the web
  • occasional interaction with human trainers

The task:

  • run 24x7, forever
  • each day:

1. extract more facts from the web to populate the ontology 2. learn to read (perform #1) better than yesterday

slide-5
SLIDE 5

NELL today

Running 24x7, since January, 12, 2010 Result:

  • knowledge base with 90 million candidate beliefs
  • learning to read
  • learning to reason
  • extending ontology
slide-6
SLIDE 6

Globe and Mail Stanley Cup hockey NHL Toronto CFRB Wilson play hired won Maple Leafs home town city paper league Sundin Milson writer radio Air Canada Centre team stadium Canada city stadium politician country Miller airport member Toskala Pearson Skydome Connaught Sunnybrook hospital city company skates helmet uses equipment won Red Wings Detroit hometown GM city company competes with Toyota plays in league Prius Corrola created Hino acquired automobile economic sector city stadium

NELL knowledge fragment

climbing football uses equipment

* including only correct beliefs

slide-7
SLIDE 7

NELL Is Improving Over Time (Jan 2010 to Nov 2014)

number of NELL beliefs vs. time

all beliefs high conf. beliefs 10’s of millions millions

reading accuracy vs. time (average over 31 predicates)

precision@10 mean avg. precision top 1000

human feedback vs. time (average 2.4 feedbacks per predicate per month)

slide-8
SLIDE 8

NELL Today

  • eg. “diabetes”, “Avandia”, “tea”, “IBM”, “love” “baseball” “San Juan”

“BacteriaCausesCondition” “kitchenItem” “ClothingGoesWithClothing” …

slide-9
SLIDE 9

Portuguese NELL

[Estevam Hruschka, 2014]

slide-10
SLIDE 10

How does NELL work?

slide-11
SLIDE 11

Semi-Supervised Bootstrap Learning

Paris Pittsburgh Seattle Montpelier mayor of arg1 live in arg1 San Francisco Berlin denial arg1 is home of traits such as arg1 it’s underconstrained!! anxiety selfishness London

Learn which noun phrases are cities:

slide-12
SLIDE 12

hard (underconstrained) semi-supervised learning problem

Key Idea 1: Coupled semi-supervised training

  • f many functions

much easier (more constrained) semi-supervised learning problem

person

noun phrase

slide-13
SLIDE 13

NP:

person

Type 1 Coupling: Co-Training, Multi-View Learning

Supervised training of 1 function: Minimize:

slide-14
SLIDE 14

NP:

person

Type 1 Coupling: Co-Training, Multi-View Learning

Coupled training of 2 functions: Minimize:

slide-15
SLIDE 15

NP:

person

Type 1 Coupling: Co-Training, Multi-View Learning

[Blum & Mitchell; 98] [Dasgupta et al; 01 ] [Ganchev et al., 08] [Sridharan & Kakade, 08] [Wang & Zhou, ICML10]

slide-16
SLIDE 16

NELL: Learned reading strategies

Mountain: "volcanic crater of _" "volcanic eruptions like _" "volcanic peak of _" "volcanic region of _" "volcano , called _" "volcano called _" "volcano is called _" "volcano known as _" "volcano Mt _" "volcano named _" "volcanoes , including _" "volcanoes , like _" "volcanoes , such as _" "volcanoes include _" "volcanoes including _" "volcanoes such as _" "We 've climbed _" "weather atop _" "weather station atop _" "week hiking in _" "weekend trip through _" "West face of _" "West ridge of _" "west to beyond _" "white ledge in _" "white summit of _" "whole earth , is _" "wilderness area surrounding _" "wilderness areas around _" "wind rent _" "winter ascent of _" "winter ascents in _" "winter ascents of _" "winter expedition to _" "wooded foothills of _" "world famous view of _" "world famous views of _" "you 're popping by _" "you 've just climbed _" "you just climbed _" "you’ve climbed _" "_ ' crater" "_ ' eruption" "_ ' foothills" "_ ' glaciers" "_ ' new dome" "_ 's Base Camp" "_ 's drug guide" "_ 's east rift zone" "_ 's main summit" "_ 's North Face" "_ 's North Peak" "_ 's North Ridge" "_ 's northern slopes" "_ 's southeast ridge" "_ 's summit caldera" "_ 's West Face" "_ 's West Ridge" "_ 's west ridge" "_ (D,DDD ft" ” "_ climbing permits" "_ climbing safari" "_ consult el diablo" "_ cooking planks" "_ dominates the sky line" "_ dominates the western skyline" "_ dominating the scenery”

slide-17
SLIDE 17

NP:

person

Type 1 Coupling: Co-Training, Multi-View Learning

[Blum & Mitchell; 98] [Dasgupta et al; 01 ] [Ganchev et al., 08] [Sridharan & Kakade, 08] [Wang & Zhou, ICML10]

slide-18
SLIDE 18

team person

NP:

athlete coach sport

NP text context distribution NP morphology NP HTML contexts

Multi-view, Multi-Task Coupling

[Blum & Mitchell; 98] [Dasgupta et al; 01 ] [Ganchev et al., 08] [Sridharan & Kakade, 08] [Wang & Zhou, ICML10] athlete(NP) à à person(NP) athlete(NP) à à NOT sport(NP) NOT athlete(NP) ß ß sport(NP) [Taskar et al., 2009] [Carlson et al., 2009]

slide-19
SLIDE 19

coachesTeam(c,t) playsForTeam(a,t) teamPlaysSport(t,s) playsSport(a,s) NP1 NP2

Type 3 Coupling: Relation Argument Types

slide-20
SLIDE 20

team coachesTeam(c,t) playsForTeam(a,t) teamPlaysSport(t,s) playsSport(a,s) person NP1 athlete coach sport team person NP2 athlete coach sport

playsSport(NP1,NP2) à à athlete(NP1), sport(NP2)

Type 3 Coupling: Relation Argument Types

  • ver 2500 coupled

functions in NELL

slide-21
SLIDE 21

Pure EM Approach to Coupled Training

E: estimate labels for each function of each unlabeled example M: retrain all functions, using these probabilistic labels Scaling problem:

  • E step: 25M NP’s, 1014 NP pairs to label
  • M step: 50M text contexts to consider for each function à

1010 parameters to retrain

  • even more URL-HTML contexts…
slide-22
SLIDE 22

NELL’s Approximation to EM

E’ step:

  • Re-estimate the knowledge base:

– but consider only a growing subset of the latent variable assignments – category variables: up to 250 new NP’s per category per iteration – relation variables: add only if confident and args of correct type – this set of explicit latent assignments *IS* the knowledge base

M’ step:

  • Each view-based learner retrains itself from the updated KB
  • “context” methods create growing subsets of contexts
slide-23
SLIDE 23

Continually Learning Reading Components

Initial NELL Architecture

Knowledge Base (latent variables) Text Context patterns (CPL) HTML-URL context patterns (SEAL) Morphology classifier (CML) Beliefs Candidate Beliefs Knowledge Integrator Human advice

slide-24
SLIDE 24

If coupled learning is the key, how can we get new coupling constraints?

slide-25
SLIDE 25

Key Idea 2: Discover New Coupling Constraints

  • learn horn clause rules/constraints:

– learned by data mining the knowledge base – connect previously uncoupled relation predicates – infer new unread beliefs – modified version of FOIL [Quinlan]

0.93 athletePlaysSport(?x,?y) ß athletePlaysForTeam(?x,?z) teamPlaysSport(?z,?y)

slide-26
SLIDE 26

team coachesTeam(c,t) playsForTeam(a,t) teamPlaysSport(t,s) playsSport(a,s) person NP1 athlete coach sport team person NP2 athlete coach sport

Learned Probabilistic Horn Clause Rules

0.93 playsSport(?x,?y) ß playsForTeam(?x,?z), teamPlaysSport(?z,?y)

slide-27
SLIDE 27

If:

x1 competes with (x1,x2) x2 economic sector (x2, x3) x3

Then: economic sector (x1, x3)

economic sector

Infer New Beliefs

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-28
SLIDE 28

If:

x1 competes with (x1,x2) x2 economic sector (x2, x3) x3

Then: economic sector (x1, x3)

economic sector

Inference by Random Walks

PRA: [Lao, Mitchell, Cohen, EMNLP 2011]

PRA:

  • 1. restrict

precondition to a chain.

  • 2. inference

by random walks

slide-29
SLIDE 29

Inference by KB Random Walks

[Lao, Mitchell, Cohen, EMNLP 2011]

KB: Random walk path type: logistic function for R(x,y) where ith feature = probability of arriving at node y starting at node x, and taking a random walk along path of type i Pr( R(x,y) ):

x competes with ? economic sector y

slide-30
SLIDE 30

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.32 Pittsburgh Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = ?

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-31
SLIDE 31

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.32 Pittsburgh Pennsylvania Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = ?

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-32
SLIDE 32

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.32 Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = ?

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-33
SLIDE 33

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.32 Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = ?

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-34
SLIDE 34

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.8 0.32 Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = ? Pr(U.S. | Pittsburgh, TypedPath)

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-35
SLIDE 35

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.8 0.32 AtLocation-1, AtLocation, CityLocatedInCountry 0.20 Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = ?

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-36
SLIDE 36

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.8 0.32 AtLocation-1, AtLocation, CityLocatedInCountry 0.20 Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight Delta PPG CityLocatedInCountry(Pittsburgh) = ?

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-37
SLIDE 37

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.8 0.32 AtLocation-1, AtLocation, CityLocatedInCountry 0.20 Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight Delta PPG

AtLocation

Atlanta Dallas Tokyo CityLocatedInCountry(Pittsburgh) = ?

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-38
SLIDE 38

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.8 0.32 AtLocation-1, AtLocation, CityLocatedInCountry 0.6 0.20 Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight Delta PPG

AtLocation

Atlanta Dallas Tokyo Japan CityLocatedInCountry(Pittsburgh) = ?

CityLocatedInCountry

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-39
SLIDE 39

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.8 0.32 AtLocation-1, AtLocation, CityLocatedInCountry 0.6 0.20 … … … Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = U.S. p=0.58 Delta PPG

AtLocation

Atlanta Dallas Tokyo Japan CityLocatedInCountry(Pittsburgh) = ?

CityLocatedInCountry

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-40
SLIDE 40

Feature = Typed Path CityInState, CityInstate-1, CityLocatedInCountry 0.8 0.32 AtLocation-1, AtLocation, CityLocatedInCountry 0.6 0.20 … … … Pittsburgh Pennsylvania Philadelphia Harisburg

…(14)

U.S. Feature Value Logistic Regresssion Weight CityLocatedInCountry(Pittsburgh) = U.S. p=0.58 Delta PPG

AtLocation

Atlanta Dallas Tokyo Japan CityLocatedInCountry(Pittsburgh) = ?

CityLocatedInCountry

  • 1. Tractable

(bounded length)

  • 2. Anytime
  • 3. Accuracy increases as

KB grows

  • 4. combines probabilities

from different horn clauses

[Lao, Mitchell, Cohen, EMNLP 2011]

slide-41
SLIDE 41

Random walk inference: learned rules

CityLocatedInCountry(city, country):

8.04 cityliesonriver, cityliesonriver-1, citylocatedincountry 5.42 hasofficeincity-1, hasofficeincity, citylocatedincountry 4.98 cityalsoknownas, cityalsoknownas, citylocatedincountry 2.85 citycapitalofcountry,citylocatedincountry-1,citylocatedincountry 2.29 agentactsinlocation-1, agentactsinlocation, citylocatedincountry 1.22 statehascapital-1, statelocatedincountry 0.66 citycapitalofcountry . . .

7 of the 2985 learned rules for CityLocatedInCountry

slide-42
SLIDE 42

Opportunity: Can infer more if we start with more densely connected knowledge graph à as NELL learns, it will become more dense à augment knowledge graph with a second graph of corpus statistics: <subject, verb, object> triples

[Gardner et al, 2014]

slide-43
SLIDE 43

can refer to hometown c:penguins c:pittsburgh river flows through c:monongahela

“Pgh” “Pittsburgh” “Monongahela” “Mon river” “Penguins” “Pens”

can refer to can refer to

NELL: concepts and “noun phrases”

[Gardner et al, 2014]

slide-44
SLIDE 44

can refer to hometown team:penguins city:pittsburgh river flows through river:monongahela “sits astride” “overlooks” “enters” “runs through”

“Pgh” “Pittsburgh” “Monongahela” “Mon river” “Penguins” “Pens”

“remain in” “began in” “supports” “reminded” can refer to can refer to

NELL: concepts and “noun phrases”

SVO triples from 500 M dependency parsed web pages (thank you Chris Re!)

[Gardner et al, 2014]

slide-45
SLIDE 45

can refer to hometown c:penguins c:pittsburgh river flows through c:monongahela “sits astride” “overlooks” “enters” “runs through”

“Pgh” “Pittsburgh” “Monongahela” “Mon river” “Penguins” “Pens”

“remain in” “began in” “supports” “reminded” can refer to can refer to

NELL: concepts and “noun phrases”

SVO triples from 500 M dependency parsed web pages (thank you Chris Re!)

  • Circumvents NELL’s fixed vocabulary of relations!
  • Sadly, adding these does not help: too sparse
  • But clustering verb phrases based on latent

embedding (NNMF), produces significant improvement

  • {“lies on”, “runs through”, “flows through”, …}
  • Precision/recall over 15 NELL relations:

KB only: 0.80 / 0.33 KB + SVOlatent: 0.87 / 0.42

[Gardner et al., 2014]

[Gardner et al, 2014]

slide-46
SLIDE 46

Key Idea 3: Automatically extend ontology

slide-47
SLIDE 47

Ontology Extension (1)

Goal:

  • Add new relations to ontology

Approach:

  • For each pair of categories C1, C2,
  • cluster pairs of known instances, in terms of

text contexts that connect them

[Mohamed et al., EMNLP 2011]

slide-48
SLIDE 48

Example Discovered Relations

Category Pair Frequent Instance Pairs Text Contexts Suggested Name MusicInstrument Musician sitar, George Harrison tenor sax, Stan Getz trombone, Tommy Dorsey vibes, Lionel Hampton ARG1 master ARG2 ARG1 virtuoso ARG2 ARG1 legend ARG2 ARG2 plays ARG1 Master Disease Disease pinched nerve, herniated disk tennis elbow, tendonitis blepharospasm, dystonia ARG1 is due to ARG2 ARG1 is caused by ARG2 IsDueTo CellType Chemical epithelial cells, surfactant neurons, serotonin mast cells, histomine ARG1 that release ARG2 ARG2 releasing ARG1 ThatRelease Mammals Plant koala bears, eucalyptus sheep, grasses goats, saplings ARG1 eat ARG2 ARG2 eating ARG1 Eat River City Seine, Paris Nile, Cairo Tiber river, Rome ARG1 in heart of ARG2 ARG1 which flows through ARG2 InHeartOf

[Mohamed et al. EMNLP 2011]

slide-49
SLIDE 49

NELL: sample of self-added relations

  • athleteWonAward
  • animalEatsFood
  • languageTaughtInCity
  • clothingMadeFromPlant
  • beverageServedWithFood
  • fishServedWithFood
  • athleteBeatAthlete
  • athleteInjuredBodyPart
  • arthropodFeedsOnInsect
  • animalEatsVegetable
  • plantRepresentsEmotion
  • foodDecreasesRiskOfDisease
  • clothingGoesWithClothing
  • bacteriaCausesPhysCondition
  • buildingMadeOfMaterial
  • emotionAssociatedWithDisease
  • foodCanCauseDisease
  • agriculturalProductAttractsInsect
  • arteryArisesFromArtery
  • countryHasSportsFans
  • bakedGoodServedWithBeverage
  • beverageContainsProtein
  • animalCanDevelopDisease
  • beverageMadeFromBeverage
slide-50
SLIDE 50

Ontology Extension (2)

Goal:

  • Add new subcategories

Approach:

  • For each category C,
  • train NELL to read the relation

SubsetOfC: C à C

[Burr Settles]

*no new software here, just add this relation to

  • ntology
slide-51
SLIDE 51

NELL: subcategories discovered by reading

Animal:

  • Pets

– Hamsters, Ferrets, Birds, Dog, Cats, Rabbits, Snakes, Parrots, Kittens, …

  • Predators

– Bears, Foxes, Wolves, Coyotes, Snakes, Racoons, Eagles, Lions, Leopards, Hawks, Humans, …

Learned reading patterns for

"arg1 and other medium sized arg2" "arg1 and other jungle arg2” "arg1 and

  • ther magnificent arg2" "arg1 and other

pesky arg2" "arg1 and other mammals and arg2" "arg1 and other Ice Age arg2" "arg1 or other biting arg2" "arg1 and other marsh arg2" "arg1 and other migrant arg2” "arg1 and other monogastric arg2" "arg1 and other mythical arg2" "arg1 and other nesting arg2" "arg1 and other night arg2" "arg1

AnimalSubset(arg1,arg2)

slide-52
SLIDE 52

NELL: subcategories discovered by reading

Animal:

  • Pets

– Hamsters, Ferrets, Birds, Dog, Cats, Rabbits, Snakes, Parrots, Kittens, …

  • Predators

– Bears, Foxes, Wolves, Coyotes, Snakes, Racoons, Eagles, Lions, Leopards, Hawks, Humans, …

Chemical:

  • Fossil fuels

– Carbon, Natural gas, Coal, Diesel, Monoxide, Gases, …

  • Gases

– Helium, Carbon dioxide, Methane, Oxygen, Propane, Ozone, Radon…

Learned reading patterns:

"arg1 and other medium sized arg2" "arg1 and other jungle arg2” "arg1 and

  • ther magnificent arg2" "arg1 and other

pesky arg2" "arg1 and other mammals and arg2" "arg1 and other Ice Age arg2" "arg1 or other biting arg2" "arg1 and other marsh arg2" "arg1 and other migrant arg2” "arg1 and other monogastric arg2" "arg1 and other mythical arg2" "arg1 and other nesting

Learned reading patterns:

"arg1 and other hydrocarbon arg2” "arg1 and other aqueous arg2” "arg1 and other hazardous air arg2" "arg1 and oxygen are arg2” "arg1 and such synthetic arg2” "arg1 as a lifting arg2" "arg1 as a tracer arg2" "arg1 as the carrier arg2” "arg1 as the inert arg2" "arg1 as the primary cleaning arg2” "arg1 and other noxious arg2" "arg1 and other trace arg2" "arg1 as the reagent arg2" "arg1 as the tracer

slide-53
SLIDE 53

NELL Architecture

Knowledge Base (latent variables) Text Context patterns (CPL) Orthographic classifier (CML) Beliefs Candidate Beliefs Evidence Integrator Human advice Actively search for web text (OpenEval) Infer new beliefs from

  • ld

(PRA) Image classifier (NEIL) Ontology extender (OntExt) URL specific HTML patterns (SEAL)

slide-54
SLIDE 54

Key Idea 4: Cumulative, Staged Learning

  • 1. Classify noun phrases (NP’s) by category
  • 2. Classify NP pairs by relation
  • 3. Discover rules to predict new relation instances
  • 4. Learn which NP’s (co)refer to which latent concepts
  • 5. Discover new relations to extend ontology
  • 6. Learn to infer relation instances via targeted random walks
  • 7. Vision: connect NELL and NEIL
  • 8. Learn to microread single sentences
  • 9. Learn to assign temporal scope to beliefs
  • 10. Goal-driven reading: predict, then read to corroborate/correct
  • 11. Make NELL a conversational agent on Twitter
  • 12. Add a robot body to NELL

Learning X improves ability to learn Y

NELL is here

slide-55
SLIDE 55

Consistency Correctness Self reflection

slide-56
SLIDE 56

The core problem:

  • Agents can measure internal consistency,

but not correctness Challenge:

  • Under what conditions does consistency à correctness?
slide-57
SLIDE 57

The core problem:

  • Agents can measure internal consistency,

but not correctness Challenge:

  • Under what conditions does consistency à correctness?
  • Can an autonomous agent determine its accuracy from
  • bserved consistency?
slide-58
SLIDE 58

Problem setting:

  • have N different estimates of target function
  • agreement between fi, fj :

[Platanios, Blum, Mitchell, UAI 2014]

slide-59
SLIDE 59

Problem setting:

  • have N different estimates of target function
  • agreement between fi, fj :

Key insight: errors and agreement rates are related

[Platanios, Blum, Mitchell, UAI 2014]

Pr[neither makes error] + Pr[both make error]

  • prob. fi and fi

agree

  • prob. fi

error

  • prob. fj

error

  • prob. fi and fj

both make error

slide-60
SLIDE 60

Estimating Error from Unlabeled Data

  • 1. IF f1 , f2 , f3 make indep. errors, and accuracies > 0.5

THEN à Measure errors from unlabeled data:

  • use unlabeled data to estimate a12, a13, a23
  • solve three equations for three unknowns e1, e2, e3
slide-61
SLIDE 61

Estimating Error from Unlabeled Data

  • 1. IF f1 , f2 , f3 make indep. errors, accuracies > 0.5

THEN à

  • 2. but if errors not independent
slide-62
SLIDE 62

Estimating Error from Unlabeled Data

  • 1. IF f1 , f2 , f3 make indep. errors, accuracies > 0.5

THEN à

  • 2. but if errors not independent
slide-63
SLIDE 63

True error (red), estimated error (blue)

NELL classifiers:

[Platanios, Blum, Mitchell, UAI 2014]

slide-64
SLIDE 64

True error (red), estimated error (blue)

NELL classifiers: Brain image fMRI classifiers:

[Platanios, Blum, Mitchell, UAI 2014]

slide-65
SLIDE 65

Summary

  • 1. Use coupled training for semi-supervised learning
  • 2. Datamine the KB to learn probabilistic inference rules
  • 3. Automatically extend ontology
  • 4. Use staged learning curriculum

New directions:

  • Self-reflection, self-estimates of accuracy (A. Platanios)
  • Incorporate vision with NEIL (Abhinav Gupta)
  • Microreading (Jayant Krishnamurthy, Ndapa Nakashole)
  • Aggressive ontology expansion (Derry Wijaya)
  • Portuguese NELL (Estevam Hrushka)
  • never-ending learning phones? robots? traffic lights?
slide-66
SLIDE 66

thank you

and thanks to: Darpa, Google, NSF, Yahoo!, Microsoft, Fulbright, Intel follow NELL on Twitter: @CMUNELL browse/download NELL’s KB at http://rtw.ml.cmu.edu