Cross-Domain Scruffy Inference Kenneth C. Arnold Henry Lieberman - - PowerPoint PPT Presentation

cross domain scruffy inference
SMART_READER_LITE
LIVE PREVIEW

Cross-Domain Scruffy Inference Kenneth C. Arnold Henry Lieberman - - PowerPoint PPT Presentation

Cross-Domain Scruffy Inference Kenneth C. Arnold Henry Lieberman MIT Mind Machine Project Software Agents Group, MIT Media Laboratory AAAI Fall Symposium on Common Sense Knowledge November 2010 Kenneth C. Arnold, Henry Lieberman (MIT)


slide-1
SLIDE 1

Cross-Domain Scruffy Inference

Kenneth C. Arnold Henry Lieberman

MIT Mind Machine Project Software Agents Group, MIT Media Laboratory

AAAI Fall Symposium on Common Sense Knowledge November 2010

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 1 / 16

slide-2
SLIDE 2

Vision

Informal (“scruffy”) inductive reasoning over non-formalized knowledge Use multiple knowledge bases without tedious alignment.

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 2 / 16

slide-3
SLIDE 3

We’ve Done This. . .

ConceptNet and WordNet (Havasi et al. 2009) Topics and Opinions in Text (Speer et al. 2010) Code and Descriptions of Purpose (Arnold and Lieberman 2010) but how does it work?

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 3 / 16

slide-4
SLIDE 4

This Talk

Background Blending is Collective Matrix Factorization. Singular vectors rotate. Other blending layouts work too.

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 4 / 16

slide-5
SLIDE 5

Background

Matrix Representations of Knowledge

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 5 / 16

slide-6
SLIDE 6

Background

Factored Inference

Filling in missing values is inference.

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 6 / 16

slide-7
SLIDE 7

Background

Factored Inference

Represent each concept i and each feature j by k-dimensional vectors ci and fj such that when A(i, j) is known, A(i, j) ≈ ci · fj. If A(i, j) is unknown, infer ci · fj. Equivalently, stack each ci in rows of C, same for F, then A ≈ CF T.

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 7 / 16

slide-8
SLIDE 8

Collective Factorization

Quantifying factorization quality

Quantify the “≈” in A ≈ CF T as a divergence: D(XY T|A) Minimizing loss ensures that the factorization fits the data Many functions possible, e.g., SVD minimizes squared error: Dx2(ˆ A|A) =

  • ij

(aij − ˆ aij)2.

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 8 / 16

slide-9
SLIDE 9

Collective Factorization

Collective Matrix Factorization

An analogy. . . Let people p rate restaurants r, represented by positive or negative values in p × r matrix A. Restaurants also have characteristics c (e.g., “serves vegetarian food”, “takes reservations”, etc.), represented by matrix B. Incorporating characteristics may improve rating prediction. Use the same restaurant vector to factor preferences and characteristics: A ≈ PRT B ≈ RCT

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 9 / 16

slide-10
SLIDE 10

Collective Factorization

Collective Matrix Factorization

A ≈ PRT B ≈ RCT

(A is person by restaurant, B is restaurant by characteristics)

Collective Matrix Factorization (Singh and Gordon 2008) gives a framework for solving this type of problem Spread out the approximation loss: αD(PRT|A) + (1 − α)D(RCT|B) At α = 1, factors as if characteristics were just patterns of ratings. At α = 0, factors as if only qualities, not individual restaurants, mattered for ratings.

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 10 / 16

slide-11
SLIDE 11

Collective Factorization

Blending is a CMF

A ≈ PRT B ≈ RCT

(A is person by restaurant, B is restaurant by characteristics)

Can also solve with Blending: Z =

  • αAT

(1 − α)B

  • ≈ R

P C T If decomposition is SVD, loss is seperable by component: D

  • R

P C T |Z

  • = D(RPT|αAT) + D(RCT|(1 − α)B)

⇒ Blending is a kind of Collective Matrix Factorization

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 11 / 16

slide-12
SLIDE 12

Blended Data Rotates the Factorization

Veering

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 12 / 16

slide-13
SLIDE 13

Blended Data Rotates the Factorization

Blended Data Rotates the Factorization

What happens at an intersection point? Consider you’re blending X and Y. Start with X ≈ ABT; what happens as you add in Y? First add in the new space that only Y covered. Now data is off-axis, so rotate the axes to align with the data.

−1.5 −1.0 −0.5 0.0 0.5 1.0 1.5 −1.0 −0.5 0.0 0.5 1.0

a = 1, θ = 0π a = 1, θ = 0.25π a = 1, θ = 0.5π a = 0, θ = 0.5π

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 13 / 16

slide-14
SLIDE 14

Blended Data Rotates the Factorization

Veering

“Veering” is caused by singular vectors of the blend rotating between corresponding singular vectors of the source matrices.

0.0 0.5 1.0 1.5 2.0 α 0.0 0.5 1.0 1.5 2.0 2.5 singular values

θU = 1.1 θU = 0.26 θU = 1.3 Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 14 / 16

slide-15
SLIDE 15

Layout Tricks

Bridge Blending

English ConceptNet French ConceptNet

English Features French Concepts French Features English Concepts X Y

En ↔ Fr Dictionary

T T

General bridge blend: X Y Z

UXY U0Z VX0 VYZ T = UXYV T

X0

UXYV T

YZ

U0ZV T

X0

U0ZV T

YZ

  • Again, loss factors:

D(ˆ A|A) =D(UXYV T

X0|X) + D(UXYV T YZ|Y)+

D(U0ZV T

X0|0) + D(U0ZV T YZ|Z)

VYZ ties factorization of X and Z together through bridge data Y. Could use weighted loss in empty corner.

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 15 / 16

slide-16
SLIDE 16

Summary

Summary

Blending is a Collective Matrix Factorization “Veering” indicates singular vectors rotating between datasets What’s next?

CMF permits many objective functions, even different ones for different input data. What’s appropriate for commonsense inference? Incremental? Can CMF do things we thought we needed 3rd-order for?

Kenneth C. Arnold, Henry Lieberman (MIT) Cross-Domain Scruffy Inference CSK 2010 16 / 16