dsolve morphological segmentation for german using
play

Dsolve Morphological Segmentation for German using Conditional - PowerPoint PPT Presentation

Dsolve Morphological Segmentation for German using Conditional Random Fields Kay-Michael W urzner, Bryan Jurish { wuerzner,jurish } @bbaw.de SFCM Universit at Stuttgart 17th September 2015 Outline p Morphological analysis p Existing


  1. Dsolve – Morphological Segmentation for German using Conditional Random Fields Kay-Michael W¨ urzner, Bryan Jurish { wuerzner,jurish } @bbaw.de SFCM Universit¨ at Stuttgart 17th September 2015

  2. Outline p Morphological analysis p Existing approaches p Morphological segmentation as sequence labeling p Experiments p Discussion & Outlook

  3. Morphological analysis Goal p identification & classification of t operations t operands Deriv . . . forming complex words Operations p compounding Comp er noun suffix p derivation p inflection basket noun ball verb Operands p morphemes ( � deep analysis), or p morphs ( � surface analysis)

  4. Morphological analysis: ambiguity Ministern . . . w.r.t. Identification [ mini adj ][ Stern noun ] [ Minister noun ][ n dat. pl. ] p > 1 segmentation possible ‘mini-star’ ‘ministers’ Sammelei . . . w.r.t. Classification [ sammel verb ][ Ei noun ] [ sammel verb ][ ei noun suffix ] p > 1 category available ‘collector’s egg’ ‘compilation’

  5. Existing approaches: finite-state methods p Finite lexicon & regular rules using (weighted) finite-state transducers (cf. Karttunen & Beesley, 2003) great- <5> 3 � :Sg � :NN grandma 0 1 2 s:Pl 4 p Tropical semiring weights as measure of complexity t word formation processes associated with non-negative costs t prefer minimal-cost (least complex) analyses p German: e.g. SMOR , TAGH (Schmid et al. 2004; Geyken & Hanneforth 2005)

  6. Existing approaches: affix removal p Identify & remove bound morphemes (prefixes, suffixes) (Porter 1980) t assume remaining material is the stem p Usually implemented as series of cascaded rewrite heuristics (Moreira & Huyck 2001) no Word ends Word ends Begin ... in ’s’? in ’a’? yes yes no Plural Feminine Augmentative reduction reduction reduction p No (exhaustive) lexicon necessary p Syllable (CV) structure supports affix removal p Works best for non-compounding languages; t has also been applied to German (Reichel & Weinhammer 2004)

  7. Existing approaches: morphology induction Basic Idea p bootstrap segmentation model from un-annontated raw text p traceable back to Harris’ notion of “Successor Frequency” � � SF( w, i ) = outDegree ptaNode( w 1 · · · w i ) p SF peaks indicate morpheme boundaries Heuristic Approaches (e.g. Goldsmith 2001) p minimum stem length, maximum affix length, minimum # stems / suffix, . . . p tend to under-segment words (poor recall) Stochastic Approaches (e.g. Creutz & Lagus 2002, 2005) p incremental greedy MDL segmentation � hierarchical model p tend to over-segment words (poor precision)

  8. Existing approaches: summary +rules -rules +lexicon finite-state morphology Dsolve -lexicon affix removal stemming morphology induction

  9. Existing approaches: summary +rules -rules +lexicon finite-state morphology Dsolve -lexicon affix removal stemming morphology induction p Lexicon- & grammar-creation � very labor-intensive p Hard to debug, hard to maintain p Efficient implementations available p Very good analysis quality

  10. Existing approaches: summary +rules -rules +lexicon finite-state morphology Dsolve -lexicon affix removal stemming morphology induction p Grammar creation requires much less manual effort than FSM p Hard to debug, tricky to implement efficiently p Ambiguity handling � difficult p Mediocre analysis quality

  11. Existing approaches: summary +rules -rules +lexicon finite-state morphology Dsolve -lexicon affix removal stemming morphology induction p Least labor-intensive (given an induction algorithm) p No direct influence on resulting grammar (only via training-corpus selection) p Inherent ranking of multiple available analyses p Insufficient analysis quality (for production applications)

  12. Segmentation ∼ Labeling: binary classification p Sequence classification t Set of observation symbols O , set of classes C t Map an observation o = o 1 . . . o n onto the most probable string of classes c = c 1 . . . c n using an underlying statistical model p Observations O : surface character alphabet (Klenk & Langer 1989) p Classes C = { 0 , 1 } where  1 if o i is followed by a morph boundary  c i = 0 otherwise  p Example Ge.folg.s.leute.n (“henchmen [ dative ] ”) G e f o l g s l e u t e n 0 1 0 0 0 1 1 0 0 0 0 1 0

  13. Segmentation ∼ Labeling: span-based classes p Span-based annotation (Ruokolainen et al. 2013) p Observations O : surface character alphabet p Classes C = { B, I, E, S } where  if o i is preceded and followed by a morph boundary S      otherwise, if o i is preceded by a morph boundary B  c i = otherwise, if o i is followed by a morph boundary E      otherwise I  p Example � Ge �� folg �� s �� leute �� n � (“henchmen [ dative ] ”) G e f o l g s l e u t e n B E B I I E S B I I I E S

  14. Segmentation ∼ Labeling: typed boundary classes p Classification of morph boundaries p Observations O : surface character alphabet p Classes C = { + , # , ∼ , 0 } where  + if o i is the final character of a prefix      # otherwise, if o i is is the final character of a free morph  c i = ∼ otherwise, if o i +1 is the initial character of a suffix      0 otherwise  p Example Ge+folg ∼ s#leute ∼ n (“henchmen [ dative ] ”) G e f o l g s l e u t e n 0 + 0 0 0 ∼ # 0 0 0 0 ∼ 0

  15. Dsolve p Surface analysis of German words using sequence labeling p Type-sensitive classification scheme p Conditional Random Field model predicts boundary location and type p Features for an input string o = o 1 . . . o n use only observable context: t each position i is assigned a feature function f k j for each substring of o of length m = ( k − j + 1) ≤ N within a context window of N − 1 characters relative to position i t N is the context window size or “order” of the Dsolve model ( �≡ CRF order) f k j ( o, i ) = o i + j · · · o i + k for − N < j ≤ k < N p Trained on modest set of manually annotated data

  16. Experiments Materials p Manual annotation of 15,522 distinct German word-forms t types and locations of word-internal morph boundaries p For reference: canoo.net , Etymologisches W¨ orterbuch des Deutschen Boundary type #/Boundaries #/Words prefix-stem ( + ) 4,078 3,315 stem-stem ( # ) 5,808 5,543 stem-suffix ( ∼ ) 11,182 8,347 21,068 11,967 total p Published under the CC BY-SA 3.0 license: http://kaskade.dwds.de/gramophone/de-dlexdb.data.txt

  17. Experiments Method p Report inter-annotator agreement for a data subset p Compare morph boundary detection of Dsolve CRF approach to t Morfessor FlatCat (Gr¨ onroos et al. 2014) t Span-based morph annotation (Ruokolainen et al. 2013) p Compute results for morph boundary classification p Test model orders 1 ≤ N ≤ 5 using 10-fold cross validation p Report precision ( pr ), recall ( rc ), harmonic average ( F ), and word accuracy ( acc ) Implementation p wapiti for CRF training and application (Lavergne et al. 2010)

  18. Experiments: evaluation measures Given a finite set W of annotated words and a finite set of boundary classes C (with the non-boundary class 0 ∈ C ), we associate with each word w = w 1 w 2 . . . w m ∈ W two partial boundary-placement functions B relevant ,w : N → C \{ 0 } : i �→ c : ⇔ c occurs at position i in w B retrieved ,w : N → C \{ 0 } : i �→ c : ⇔ c predicted at position i in w and define | relevant ∩ retrieved | Precision pr := | retrieved | | relevant ∩ retrieved | Recall rc := | relevant | 2 · pr · rc F-score F := pr+rc |{ w ∈ W | B retrieved ,w = B relevant ,w }| Accuracy acc := , where: | W | relevant := { ( w, i, c ) | ( i �→ c ) ∈ B relevant ,w } retrieved := { ( w, i, c ) | ( i �→ c ) ∈ B retrieved ,w }

  19. Experiments: inter-annotator agreement p Independent 2 nd manual annotation of a data subset ( n = 1000 ) by an expert p Our own annotation serves as the “gold standard” (i.e. relevant ) Boundary Symbol pr% rc% F% acc% + 92.05 97.20 94.56 n/a # 96.01 93.28 94.63 n/a ∼ 93.28 92.66 92.97 n/a TOTAL [+types] 93.74 93.74 93.74 87.40 TOTAL [ − types] 96.20 96.20 96.20 87.40 p Reasonably high agreement with discrepancies particularly w.r.t.: t latinate word formation (e.g. volunt(˜)aristisch , “voluntaristic”) t prefixion ↔ compounding (e.g. *weg+gehen vs. weg#gehen , “to leave”)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend