Combining Global Models for Parsing Universal Dependencies
Team C2L2 —
Tianze Shi, Felix G. Wu, Xilun Chen, Yao Cheng
Cornell University
Parsing Universal Dependencies Team C2L2 Tianze Shi, Felix G. Wu, - - PowerPoint PPT Presentation
Combining Global Models for Parsing Universal Dependencies Team C2L2 Tianze Shi, Felix G. Wu, Xilun Chen, Yao Cheng Cornell University Overview Scope of Our System What we did What we didnt do Word Segmentation Sentence
Combining Global Models for Parsing Universal Dependencies
Team C2L2 —
Tianze Shi, Felix G. Wu, Xilun Chen, Yao Cheng
Cornell University
Overview — Scope of Our System
What we did What we didn’t do
Overview — Highlights
argmax
𝑧∈𝒵
based models
compact features
resource demand
syntactic transfer
Overview — System Pipeline
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
Sentence delimited & tokenized
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
Languages OOV rates ↓ (word) ko – Korean 43.68% la – Latin 41.22% sk – Slovak 36.51% … … Average 14.4%
* Measured on development set
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
p a r s i n g
parsing
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
dependency Universal parsing dependency Universal parsing
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
Bi-LSTM features
Eisner’s Arc-eager Global Arc-hybrid Global Reparsing by Eisner’s (Sagae and Lavie, 2006)
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
* Shi, Huang and Lee (2017, EMNLP)
Global Transition-based Parsing
Kuhlmann, Gómez-Rodríguez and Satta, 2011)
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
Eisner’s Arc-eager Arc-hybrid
Compact (2) Feature Set Scoring function: deep bi-affine
(Dozat and Manning, 2017)
head modifier stack top buffer top stack top buffer top
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
Ensembling
75.00 74.32 74.00 73.75
73 73.5 74 74.5 75
LAS
Full Single Arc-eager Single Arc-hybrid Single Eisner’s
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
head modifier concat( ) Multi-layer perceptron
nsubj
…….
II. Feature Extraction I. UDPipe Pre-process III. Unlabeled Parsing IV. Arc Labeling
Effect of Ensemble
75.00 74.69
73 73.5 74 74.5 75
LAS
Full Single Labeler
Results — Official Ranking
Big Treebanks 2 Small Treebanks 1 PUD Treebanks 2 Surprise Languages 1 Overall 2
Strategies — Small Treebanks
Train on: {fr, fr_partut, fr_sequoia} All tasks Combined model
Finetune on fr_sequoia All tasks Finetune on fr_partut All tasks Finetune on fr All tasks
fr model fr_partut model fr_sequoia model
Task finetune Task finetune Task finetune
Results — Small Treebanks
fr fr_partut fr_sequoia fr 84.09 fr_partut 79.53 fr_sequoia 84.65 Combined 87.57 85.57 82.80 +Finetune 87.87 86.65 86.37
* UAS results on dev set, using gold segmentation
Test Treebank
Train Treebank
Strategies — Surprise Languages
p a r s i n g
parsing
Bi-directional LSTM
parsing
concat( )
UPOS tag Bag of Morphology
Morphology tags
Max pooling
Results — Surprise Languages
Target Source* Ranking Buryat Hindi 2 Upper Sorbian Czech 1 Kurmanji Persian 1 North Sámi Finnish 1 Average 1
*selected via WALS
Implementation
Efficiency
* Not Benchmark Results 16.27 4.64 26.17 8.88 5.96 5 10 15 20 25 30
Stanford (Stanford) C2L2 (Ithaca) IMS (Stuttgart) HIT-SCIR (Harbin) LATTICE (Paris) Runtime (Hours) *
LAS CPUs RAM 76.30 4 16 75.00 2 8 74.42 12 64 72.11 1 8 70.93 8 32
Combining Global Models for Parsing Universal Dependencies
Team C2L2 — Tianze Shi, Felix G. Wu, Xilun Chen, Yao Cheng
argmax
𝑧∈𝒵
based models
fine-tuning
https://github.com/CoNLL-UD-2017/C2L2