Approximating Learning Curves for Active-Learning-Driven Annotation
Katrin T
- manek and Udo Hahn
Approximating Learning Curves for Active-Learning-Driven Annotation - - PowerPoint PPT Presentation
Approximating Learning Curves for Active-Learning-Driven Annotation Katrin T omanek and Udo Hahn Jena University Language and Information Engineering (JULIE) Lab Agenda Introduction to Active Learning Stopping Conditions
corpus selection annotation passive annotation scenario (aka Random Sampling)
labeled unlabeled
annotation „intelligent“ example selection corpus selection annotation passive annotation scenario (aka Random Sampling) active annotation scenario (aka Active Learning)
labeled unlabeled labeled unlabeled
labeled examples
C1 C2 Cn ... committee: (1) train on subsets (bagging) (2) predict labels (4) select by dis- agreement (3) calculate dis- agreement
u1: P1 P2 ... Pn
D1(P1,P2,Pn)
...
uk: P1 P2 ... Pn
Dk(P1,P2,Pn)
... U
AL pool: unlabeled examples
(5) annotate
learning curves
60K tokens
> 50% reduction
130K tokens F=0.83 learning curves
learning curves
bad
Could gain a lot by further annotation
bad bad
Could gain a lot by further annotation Further annotation will not increase classifier performance significantly
bad bad better
Could gain a lot by further annotation Further annotation will not increase classifier performance significantly
➔ not applicable in practice as gold standard not available
– Estimate the (progression of) learning curve without need
– Based on agreement among committee members – Does not require extra labeling effort – Agreement curve approximates progression of learning
➔ We can tell relative position in annotation process from it:
– Steep slope ? – Convergence ?
– Agreement among committee:
➔ When agreement among committee members
– On separate validation set
– Otherwise agreement curve often not reliable
each AL iteration:
– Approximation of learning curve usually works well in
simulation scenarios, because... » few hard cases left in later AL iterations (perfect agreement)
– But fails in real-world annotation scenarios, because...
» in practice AL will always find tricky cases...
– News-paper, MUC entities (PERS, LOC, ORG) – AL pool: ~ 14,000 sentences – Gold Standard: ~ 3,500 sentences
learning curves agreement curve
➔ Method to monitor progress of annotation needed
– Works well: good approximation of learning curve – No extra annotation effort: does not require labeled gold
http://www.julielab.de/