SLIDE 29 Introduction Sparse high-dimensional models
Distributional Memory (Baroni and Lenci 2010)
I Tensor of (word, link, word) triples, e.g. (book, obj, read)
I also (sharp, as adj as, knife); (geek, use, computer); . . .
I TypeDM: feature scores = local MI (Evert 2004) based on
number of distinct surface realisations of the link pattern
I 30,686 target terms ◊ 25,336 link types ◊ 30,686 collocates
I W1 ◊ LW2 matricization yields state-of-the-art DSM
I very high-dimensional: 30,686 ◊ 3,127,436 matrix I extremely sparse: 131 million nonzero cells = 0.137%
 Dimensionality reduction to make data set manageable
I e.g. 1.25 M uninformative features with single nonzero entry Stefan Evert (TU Darmstadt) Dimensionality Reduction for DSM wordspace.collocations.de 18 / 50