Introduction Detecting Errors in Effects of Annotation Errors - - PowerPoint PPT Presentation

introduction
SMART_READER_LITE
LIVE PREVIEW

Introduction Detecting Errors in Effects of Annotation Errors - - PowerPoint PPT Presentation

Detecting Errors in Introduction Detecting Errors in Effects of Annotation Errors Detecting Errors in Corpus Annotation Corpus Annotation Corpus Annotation Detmar Meurers Detmar Meurers Detmar Meurers University of T ubingen


slide-1
SLIDE 1 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Detecting Errors in Corpus Annotation

On Variation Detection (http://decca.osu.edu) Detmar Meurers University of T¨ ubingen CLARA Thematic Training Course on Methods and Technologies for Consolidating and Harmonising Treebank Annotation UFAL, Charles University, Prague December 13–16, 2010 1 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Introduction

Corpora with “gold standard” annotation are used ◮ as training and testing material for NLP algorithms/tools ◮ for searching for linguistically relevant patterns Such annotation generally results from a semi-automatic markup process, which can include errors through ◮ automatic processes ◮ human annotation or post-editing 2 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Effects of Annotation Errors

◮ Less reliable training of NLP technology ◮ van Halteren et al. (2001): a tagger trained on WSJ (Marcus et al. 1993) performs significantly worse than
  • ne trained on LOB (Johansson 1986)
◮ Less reliable evaluation of NLP technology ◮ van Halteren (2000): 13.6%–20.5% of cases where WPDV tagger disagrees with BNC-sampler annotation, cause is error in BNC-sampler (0.3% error, Leech 1997). Error rates for other corpora much higher. ◮ Padro & Marquez (1998): because of errors in the testing data, cannot tell which of two taggers is better ◮ Low precision and recall of queries for already rare linguistic phenomena ◮ Meurers (2005): low precision of queries for verbal complex patterns since certain finite and non-finite verb forms are not reliably distinguished by German taggers 3 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Effects of Annotation Errors

Searching for linguistic phenomena: The role of precision ◮ By precision of search we are referring to: ◮ Of the results to the query, how many represent the learner language patterns searched for? ◮ False positives can result in two ways: ◮ Expression used in query also characterizes patterns
  • ther than the ones we are interested in.
◮ Some of the annotations the query refers to are incorrect. ◮ Requirements on precision of search ◮ for qualitative analysis: Needs to be high enough to find relevant examples among the false positives. ◮ for quantitative analysis: For reliable results, very high precision is required, in particular where specific rare language phenomena are concerned. ◮ As known from Zipf’s curse, most things occur rarely . . . 4 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Effects of Annotation Errors

Searching for linguistic phenomena: The role of recall ◮ By recall of search we are referring to: ◮ How many of the intended examples that in principle are in the corpus are in fact found by the query? ◮ Requirements on recall of search ◮ for qualitative analysis: Any results found useful, but danger of partial blindness where subcases are not captured by query approximating target phenomenon. ◮ for quantitative analysis: Maximizing recall is crucial for reliable quantitative results. ⇒ Where a query characterizing a target phenomenon is expressed in terms of annotation, high annotation quality is important, and essential for quantitative analysis. 5 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

How to obtain high quality annotation

◮ Annotate corpus independently several times, then test interannotator agreement (Brants & Skut 1998; Artstein & Poesio 2009) ◮ Interannotator agreement: Can the distinctions made in the annotation scheme can be applied consistently based on the information available in the corpus? ◮ Define adequate annotation scheme, with explicit documentation and a list of problematic cases to achieve maximal agreement (Voutilainen & J¨ arvinen 1995; Sampson & Babarczy 2003). ◮ keep only distinctions which can be reliably and consistently identified and annotated uniquely ◮ appendix of difficult cases and how to resolve them crucial 6 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Our research questions

◮ How about automatic methods for error detection? ◮ Detection can feed into repair as second stage of correction (cf. also Oliva 2001; Blaheta 2002). ◮ What can be done for annotation of language in general? ◮ Can errors be found in common “gold standard” corpora regarding their ◮ part-of-speech annotation (Dickinson & Meurers 2003a) ◮ syntactic annotation (Dickinson & Meurers 2003b; Boyd, Dickinson & Meurers 2007) ◮ discontinuous syntactic annotation (Dickinson & Meurers 2004) ◮ dependency annotation (Boyd, Dickinson & Meurers 2008) including spoken language corpora (Dickinson & Meurers 2005a). ⇒ Detection of annotation errors through automatic analysis of comparable data recurring in the corpus ◮ DECCA NSF project (http://decca.osu.edu) ◮ Dickinson (2005) 7 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Variation Detection for POS Annotation

(Dickinson & Meurers 2003a) ◮ POS tagging reduces the set of lexically possible tags to the correct tag for a specific corpus occurrence. ◮ A word occurring multiple times in a corpus can occur with more than one annotation. ◮ Variation: material occurs multiple times in corpus with different annotations ◮ Variation can result from ◮ genuine ambiguity ◮ inconsistent, erroneous tagging ◮ How can one find such variation and decide whether it’s an ambiguity or error? 8 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Classifying variation

◮ The key to classifying variation lies in the context: ◮ The more similar the context of the occurrences, the more likely the variation is an error. ◮ A simple way of making “similarity of context” concrete is to say it consists of ◮ words ◮ which immediately surround the variation, and ◮ require identity of contexts. ⇒ Extract all n-grams containing at least one token that is annotated differently in another occurrence of that n-gram in the corpus. ◮ variation nucleus: recurring unit with different annotation ◮ variation n-gram: variation nucleus with identical context 9 / 91
slide-2
SLIDE 2 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Computing variation n-grams

◮ Example from WSJ: Variation 12-gram with off (1) to ward off a hostile takeover attempt by two European shipping concerns ◮ once annotated as a preposition (IN), and ◮ once as a particle (RP). ◮ Note: Such a 12-gram contains two variation 11-grams: (2) to ward ward
  • ff
  • ff
a a hostile hostile takeover takeover attempt attempt by by two two Eur. Eur. shipping shipping concerns → Calculate variation n-grams based on variation n−1-grams to obtain an algorithm efficient enough for large corpora. ◮ Essentially an instance of the a priori algorithm (Agrawal & Srikant 1994). 10 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Computing variation n-grams

Algorithm
  • 1. Calculate the set of variation unigrams in the corpus
and store them.
  • 2. Extend the n-grams by one word to either side. For
each resulting (n + 1)-gram ◮ check whether it has another instance in the corpus and ◮ store it in case there is a variation in the way the
  • ccurrences are tagged.
  • 3. Repeat step 2 until we reach an n for which no variation
n-grams are in corpus. Running this algorithm on the Penn Treebank 3 version of the WSJ, retrieves variation n-grams up to length 224. 11 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Computing variation n-grams

Example: WSJ in Penn Treebank 3 ◮ general corpus information: ◮ 1,289,201 tokens ◮ 51,457 types ◮ 23,497 of types appear only once (= 1.8% of tokens) ◮ 98.2% of tokens appear more than once ◮ variation nuclei: ◮ 7,033 types ◮ 711,994 tokens = 55.2% of all corpus tokens ◮ variation n-grams: ◮ longest: 224 ◮ 2,495 distinct variation nuclei for 6 ≤ n ≤ 224 ◮ 16,319 distinct variation nuclei for 3 ≤ n ≤ 224 ◮ each corpus position counting only in longest n-gram 12 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristics for classifying variation

  • I. The length of the context
Idea: The longer the n-gram, the more likely the variation is an error. Example: In a variation 184-gram, the nucleus lending varies between adjective (jj) and common noun (nn). ←− − − − − − − − − − lending − − − − − − − − − −→ 109 identical words jj/nn 74 identical words Here, nn is the correct annotation of this n-gram. Note: Heuristics independent of corpus, tagset, or language. 13 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristics for classifying variation

  • II. Distrust the fringe
Idea: Morphological and syntactic properties are governed
  • locally. The further the variation nucleus is away from the
edge of the n-gram, the more likely it is an error. Example: A variation 37-gram with the nucleus joined
  • ccurring as first word:
(3) a. John P . Karalis . . .
  • b. John P
. Karalis has . . . joined the Phoenix , Ariz. , law firm of Brown & Bain . Mr. Karalis , 51 , will specialize in corporate law and international law at the 110-lawyer firm . Before joining Apple in 1986 , The context preceding the 37-gram shows: ◮ In a. the verb must be tagged as past tense (vbd), ◮ in b. as past participle (vbn). 14 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Why does the non-fringe heuristic work?

◮ Non-fringe heuristic: one element of recurring context around a recurring nucleus is generally sufficient to determine that a variation in an annotation is erroneous. ◮ Is this an artifact of the WSJ annotation or is there independent motivation for such a general heuristic? ◮ Interestingly, recent research on language acquisition by Toby Mintz (USC) has addressed a related question: ◮ How do humans discover and learn categories of words? His results show that humans seem to make use of exactly such non-fringe patterns (frames) to learn categories! 15 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Independent evidence from language acquisition

◮ Mintz (2002) shows that lexical co-occurrence information of an element surrounded by a frame (i.e., X Y) leads to categorization in adults. ◮ Mintz (2003): frequent frames supply robust category information, consistent across child language corpora. ◮ Example for a frame from CHILDES (MacWhinney 2000): ◮ you put it ◮ you want it ◮ you see it → you it ◮ Cross-linguistic viability of frame concept confirmed for French (Chemla et al. 2009) and Mandarin (Xiao et al. 2006). 16 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Independent evidence from language acquisition

◮ Chemla et al. (2009) show that humans categorize words most reliably when surrounded by a frame. The other same size contexts are much worse: 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 French English Accuracy X__Y XY__ __XY ⇒ The non-fringe heuristic used for annotation error detection relies on the basic human cognitive abilities that led to the linguistic categories in the first place. 17 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results for the WSJ

◮ Of the 2,495 distinct variation nuclei (types) 6 ≤ n ≤ 224: ◮ 2,436 are errors (97.64%) ◮ Correcting the instances of these variation nuclei by hand yielded 4417 token corrections. ◮ 59 are genuine ambiguities ◮ 32 were 6-grams, 10 were 7-grams, 4 were 8-grams, . . . → relevance of heuristic to prefer long context ◮ 57 appeared first/last → relevance of heuristic to distrust the fringe ◮ 31 are the first word of the n-gram, varying between two specific tags: past tense verb (vbd) and past participle (vbn). ◮ Of 7141 distinct non-fringe variation n-gram types 3 ≤ n ≤ 224, based on sampling we found that ◮ 6626 are errors (92.8%) → each at least one correction ◮ given 3% estimated POS error rate in the WSJ, the method has a POS error recall of at least 17% 18 / 91
slide-3
SLIDE 3 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Feedback for revising annotation scheme

For 140 of the 2436 erroneous variation nuclei, the variation was clearly incorrect, but which tag is the correct one is unclear from the guidelines (Santorini 1990). Example: Salomon Brothers Inc Brothers is tagged ◮ 27 times as proper noun (nnp) ◮ 22 as plural proper noun (nnps). ⇒ Variation n-gram error detection helps identify error-prone distinctions, which need to be documented more explicitly or possibly eliminated, e.g.: ◮ proper vs. common nouns ◮ certain types of noun-adjective homographs 19 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Related work on POS error detection

◮ Work with another focus, which could be combined with
  • ur consistency-checking approach:
◮ Deriving and searching for bigrams of tags which should never be allowed (Kvˇ etˇ
  • n & Oliva 2002). →
Inconsistencies are mostly possible bigrams. ◮ Sparse Markov transducers used to detect anomalies, i.e., rare local tag patterns (Eskin 2000). → Inconsistencies are mostly recurrent, not rare. ◮ Using parsing failures to detect ill-formed annotation serving as parser input (Hirakawa et al. 2000; M¨ uller & Ule 2002). → Language specific resources. ◮ Searching and correcting with hand-written rules (Oliva 2001; Blaheta 2002) ◮ Related to consistency of annotation: ◮ Comparing tagger output with gold standard (van Halteren 2000; Abney et al. 1999). Taggers detect consistent behavior in order to replicate it. 20 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Summary for POS error detection

◮ We discussed a detection methods for POS annotation errors in gold-standard corpora: ◮ detect variation within comparable contexts ◮ classify such variation as error or ambiguity using general heuristics ◮ Idea relies on multiple corpus occurrences of a particular word with different annotations → particularly useful for hand-corrected gold-standard corpora ◮ Evaluation showed the method detects errors in the WSJ with ◮ 92.8% precision ◮ 17% estimated recall ◮ Qualitative inspection of the detected variation can provide valuable feedback for annotation scheme (re)design and documentation. 21 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Variation Detection for Syntactic Annotation

(Dickinson & Meurers 2003b, 2004; Boyd, Dickinson & Meurers 2007) ◮ Let’s try to apply variation detection to the syntactic annotation in treebanks! ◮ How can two syntactically annotated sentences be compared for this? ◮ Variation detection is closely related to interannotator agreement testing for multiply annotated corpus. ◮ How are multiple annotations of the same sentences compared for testing interannotator agreement? ◮ Calder (1997) and Brants & Skut (1998) present algorithm for detecting differences in annotation. ◮ algorithm is annotation-driven, asymmetric, and sentence-based ⇒ We are looking for a data-driven, symmetric, string-based approach. 22 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Defining variation nuclei for syntactic annotation

How can we obtain a data-driven definition of a variation nucleus as the unit of data on which the comparison of syntactic annotation can be based? Problem: No one-to-one mapping between word and label, as with part of speech. Idea: Decompose variation nucleus detection into series of runs for all relevant string lengths, more specifically ◮ define one-to-one mapping between string of a given length and the label for that string ◮ perform runs for strings from length 1 to longest constituent in corpus 23 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Defining variation nuclei for syntactic annotation

How to compare annotation for syntactic variation nuclei ◮ To obtain a uniform mapping from strings to labels ◮ assign all non-constituent occurrences of a string the special label nil. ◮ Only compare categories assigned to the entire nucleus. ◮ This intentionally ignores the internal structure, ◮ which is taken into account when shorter strings are checked. 24 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Examples from the WSJ corpus

◮ Variation between two syntactic category labels: (4) maturity labeled as next Tuesday NP twice PP
  • nce
◮ Variation between constituent and non-constituent: market NN received VBD its PRP$ biggest JJS jolt NN last JJ month NN from IN Campeau NNP Corp. NNP , , NP NP TMP NP NP market NN received VBD its PRP$ biggest JJS jolt NN last JJ month NN from IN Campeau NNP Corp. NNP , , NP NP 25 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Computing the variation nuclei of a treebank

A simple way to calculate all variation nuclei:
  • 1. step through corpus:
◮ store all stretches of length i with category label or nil
  • 2. eliminate the non-varying stretches
Problem: Inefficient generate-and-test method considering all stretches of strings starting at any position in the corpus. Insight: ◮ The way we have set things up, variation involves at least one constituent occurrence of a nucleus. → Only strings analyzed as constituent somewhere in corpus need to be compared to annotation of other
  • ccurrences of that string.
26 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Computing variation n-grams for a treebank

Algorithm For each constituent length i (1 ≤ i ≤ |longest-constituent|):
  • 1. Compute the set of nuclei:
a) Find all constituents of length i: store them with their category label b) For each type of string stored as constituent of length i, add nil for each non-constituent occurrence
  • 2. Compute variation nuclei set as:
◮ all nuclei from step 1 with more than one label
  • 3. Generate variation n-grams for these variation nuclei,
just as defined for part of speech annotation 27 / 91
slide-4
SLIDE 4 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

A case study: Applying the method to the WSJ

◮ Two types of syntactic information in the PennTreebank3 (Marcus, Santorini, Marcinkiewicz & Taylor 1999): ◮ syntactic category generally determined by ◮ lexical material in the covered string and ◮ the way this material is combined ◮ syntactic function (also) determined by ◮ material outside constituent ◮ We focus on the syntactic category. ◮ Technical realization: ◮ TIGERRegistry (Lezius et al. 2002) converts, e.g., temporal noun phrase (NP-TMP) to noun phrase node (NP) under temporal (TMP) edge ◮ variation n-gram test based on node labels only 28 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Dealing with unary trees

◮ unary branch causes same string to be annotated by two distinct categories ◮ would be detected as variation in annotation ◮ eliminate unary branches and relabel with mother/daughter category label, adding 70 new labels to original 27. ◮ Example: NP | QP | 10 million

NP/QP | 10 million 29 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Constituent lengths in the WSJ

1 10 100 1000 10000 181932 25 50 75 100 125 150 175 200 225 250 275 number of occurrences size of constituent × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × × ◮ Possible syntactic variation nucleus lengths: 1 ≤ n ≤ 271 ◮ Largest repeating string with variation in annotation: length 46 30 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Error detection results

◮ Total: 6277 distinct, non-fringe variation nuclei ◮ distinct: each corpus position is only taken into account for longest variation n-gram it occurs in ◮ non-fringe: nucleus is surrounded by at least one word
  • f identical context
◮ We inspecting 100 randomly sampled examples: ◮ 71% errors, with 95% confidence interval for point estimate of .71 being (.6211, .7989) → between 3898 and 5014 erroneous variation nuclei, each corresponding to at least one token error ◮ What are the reasons for the misclassified ambiguities? 31 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Misclassified Ambiguities I: Null elements

◮ 10 of the 29 ambiguous nuclei in sample are null elements varying between two different categories. ◮ WSJ annotators inserted markers for arguments and adjuncts realized non-locally, or unstated units of measurement (cf. Bies et al. 1995, p. 59). ◮ Example: *EXP* (expletive) annotated as S or SBAR (5) . . . it [S *EXP* ] may be a wise business investment * [S to help * keep those jobs and sales taxes within city limits] . (6) . . . it [SBAR *EXP*] may be impossible [SBAR for the broker to carry out the order] because . . . ⇒ Ambiguity arises where null items occur in place of element non-locally realized. 32 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Misclassified Ambiguities I: Null elements

Effect of eliminating variation detection for null elements ◮ remove null elements from set of variation nuclei of length 1 ◮ resulting number of non-fringe distinct variation nuclei: 5584 ◮ 78.9% of sample are errors, with 95% conf. interval (.7046, .8732): ◮ between 3934 and 4875 erroneous variation nuclei, each corresponding to at least one error instance 33 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Misclassified Ambiguities II: Coordination

◮ 6 of the 29 ambiguities deal with coordinate structures. ◮ Annotation scheme distinguishes simple (i.e., non-modified) and complex coordinate elements. ◮ Even if an element is simple, it is annotated like a complex element when conjoined with one. 34 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Coordinate structure example

interest in a flat coordinate structure: The DT amount NN covers VBZ taxes NNS , , interest NN and CC penalties NNS
  • wed
VBN NP NP interest in a complex coordinate structure: a DT lot NN
  • f
IN back JJ taxes NNS , , interest NN and CC civil JJ fraud NN penalties NNS NP NP NP NP NP ⇒ Annotation scheme makes a distinction externally motivated 35 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Related work on syntactic error detection

◮ CCGbank (Hockenmaier & Steedman 2005): derived from Penn Treebank, fixing some errors: ◮ e.g.: “Under ADVP , if the adverb has only one child, and it is tagged as NNP , change it to RB.” ◮ Blaheta (2002): discusses types of errors and some rules to identify them ◮ e.g.: “If an IN is occurring somewhere other than under a PP , it is likely to be a mistag.” ◮ Ule & Simov (2004) search for unexpected rules, using information about a node and its mother ◮ Discrepancies between mother and daughter annotation can point to errors 36 / 91
slide-5
SLIDE 5 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Summary for constituency error detection

◮ We showed how one can extend the POS-error detection approach to syntactic annotation. ◮ Illustrated with a case study based on WSJ treebank that the method is successful (71% precision) in detecting inconsistencies in syntactic category annotation. ◮ Approach supports two aspects of treebank improvement: ◮ makes it possible to find and correct erroneous variation in corpus annotation ◮ provides feedback for development of empirically adequate standards for syntactic annotation, identifying distinctions difficult to maintain over entire corpus 37 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Discontinuous constituents

◮ Discontinuous constituents (or equivalents) have been proposed in a wide range of syntactic frameworks, e.g.: ◮ Tree Adjoining Grammar (Kroch & Joshi 1987; Rambow & Joshi 1994) ◮ Categorial Grammar (Dowty 1996; Hepple 1994; Morrill 1995) ◮ linearization-based Head-Driven Phrase Structure Grammar (Reape 1993; Kathol 1995; Richter & Sailer 2001; M¨ uller 1999; Penn 1999; Donohue & Sag 1999; Bonami et al. 1999) ◮ non-projective Dependency Grammar (Br¨
  • ker 1998; Pl´
atek et al. 2001) ◮ approaches positing tangled trees (McCawley 1982; Huck 1985; Ojeda 1987; Blevins 1990) ◮ They are also used in German treebanks (NEGRA, Skut et al. 1997, 1998; TIGER, Brants et al. 2002) 38 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Some examples for discontinuous constituents

◮ An English extraposition example: (7) The man came into the room who everybody loved. ◮ An English particle verb example: (8) a. I called up John.
  • b. I called John up.
◮ German extraposition example (Brants et al. 2002): (9) Ein a Mann man kommt comes , , der who lacht laughs ‘A man who laughs comes.’ Here Ein Mann der lacht is an NP constituent. 39 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Treebanks and discontinuous constituents

◮ Treebanks which have been developed for languages with relatively free constituent order often represent discontinuous constituents (one way or another). ◮ For German, we take a closer look at: ◮ NEGRA Treebank (Skut et al. 1997, 1998) ◮ written language: Frankfurter Rundschau, a national newspaper ◮ 20,000 sentences (350,000 tokens) ◮ flat structures as encoding of argument structure ◮ TIGER Treebank (Brants et al. 2002) ◮ Extension of the NEGRA project ◮ > 35,000 sentences (700,000 tokens) 40 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

The NEGRA/TIGER Treebanks

◮ annotation consists of tree structures with node and edge labels ◮ tree structure: ◮ encodes argument structure ◮ properties: ◮ crossing branches used extensively ◮ no empty terminal nodes ◮ each daughter has one mother (but some secondary edges) ◮ node and edge labels encode: ◮ phrase level: syntactic categories ◮ lexical level: STTS part-of-speech (Schiller, Teufel & Thielen 1995) 41 / 91

An extraposition example (NEGRA corpus)

Schade ADJD , $, da"s KOUS kein PIAT Arzt NN anwesend ADJD ist VAFIN , $, der PRELS sich PRF auskennt VVFIN . $. SB OA HD S NK NK RC NP CP SB PD HD S PD SB 1 2 3 4 5 6 7 8 9 10 11 500 501 502 503 S Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Error detection for discontinuous constituents

◮ The variation n-gram method relies on the assumption that a continuous string can be mapped to a category. ◮ Extend it to account for the fact that ◮ the variation nuclei, and ◮ their contexts are no longer required to be continuous strings, and ◮ adapt the variation classification heuristics accordingly. 43 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Adapting the algorithm to discontinuity

Error detection for syntactic annotation is broken down into runs for all constituent lengths (1 ≤ i ≤ |longest-constituent|): ◮ Constituent size includes only the tokens that are a part
  • f the constituent, not possibly intervening material.
  • 1. Compute the set of nuclei:
a) Find all constituents of size i: store them with their category label b) For each type of string stored as constituent of length i, add each non-constituent occurrence with label NIL
  • 2. The variation nuclei set is the set of all nuclei with more
than one label
  • 3. Generate variation n-grams for these variation nuclei.
44 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Notes on Variation Nuclei

Discontinuous non-constituent occurrences To find all strings that match a constituent in the corpus, we need to take discontinuous strings into account. The strings to be found may not be constituents: (10) in
  • n
diesem this Punkt point seien are sich self Bonn Bonn und and London London nicht not einig agreed . . ‘Bonn and London do not agree on this point.’ (11) in
  • n
diesem this Punkt point seien are sich self Bonn Bonn und and London London
  • ffensichtlich
clearly nicht einig not agreed . . Here, sich einig is an AP in (10), but a discontinuous non-constituent (= NIL) in (11). 45 / 91
slide-6
SLIDE 6 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Notes on Variation Nuclei

Limiting the occurrence of non-constituents Constituents can overlap with non-constituent occurrences: (12) Ohne without diese these Ausgaben, expenses so so die the Weltbank, world bank seien are die Menschen the people totes dead Kapital capital ‘according to the worldbank, without these expenses the people are dead capital’ ◮ The string die Menschen occurs twice: ◮ once as a continuous constituent ◮ once as a discontinuous non-constituent ⇒ If a constituent overlaps with a non-constituent string, ignore the non-constituent string. 46 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Computing variation nuclei efficiently

Use tries for storage Task: Find all potentially discontinuous strings that match a string occurring as a constituent in the corpus. Determine a tractable domain for the search: Syntactic annotation: only consider strings within a sentence. How to do the search: ◮ Inefficient generate-and-test method:
  • 1. Generate every (potentially discontinuous) substring of
a sentence (= 2n−1 cases for sentence length n)
  • 2. Test to see which ones match a constituent.
◮ Incremental method using a trie as a guide:
  • 1. Store all constituents in a trie with words at nodes
  • 2. Incrementally match every (potentially discontinuous)
substring of a sentence with a path in the trie. ⇒ Incremental matching significantly reduces search space. 47 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Which contexts for discontinuous constituents

Idea: the more similar the context, the more likely variation in the annotation of a nucleus is an error. ◮ Previously: expanded context to left and right ◮ Now: also expand into internal context, i.e., material contained within span of discontinuous constituent but not part of constituent itself How to do it: ◮ Incrementally add context adjacent to the nucleus. ◮ Why? The most local context helps the most with disambiguation. ⇒ Require surrounding context for every terminal element
  • f the nucleus in order for it to be non-fringe.
48 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results on the TIGER corpus

The Setup ◮ Used TIGER treebank (Brants et al. 2002), a German newspaper corpus with 712.332 tokens in 40.020 sentences ◮ Evaluation of whether a detected variation points to an error was carried out by George Smith and Robert Langner of the TIGER project. 49 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results on the TIGER corpus

Baseline, without context: ◮ Method detects 10.964 variation nuclei. ◮ 13% pointed to at least one token error in sample of 100. (95% conf. interv.: 702 (6.4%) – 2.148 (19.6%) are errors) Using word contexts (non-fringe nuclei): ◮ Resulted in 500 shortest non-fringe variation nuclei, shortest non-fringe = rely solely on non-fringe heuristic ◮ 80% pointed to at least one token error in sample of 100. (95% conf. interv.: 361 (72.2%) – 439 (87.8%) are errors) ◮ Precision comparable to regular syntactic annotation (71% in Dickinson & Meurers 2003b). 50 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Increasing recall

The variation n-gram method for detecting annotation errors ◮ Finds recurring data and compares analyses in different corpus instances ◮ Uses shared context as a heuristic to determine when analyses should be annotated identically Two ways to increase recall: ◮ Redefine variation nuclei, to extend the set of what counts as recurring data for which annotation is compared. ◮ Redefine context and heuristics, to obtain more variation n-grams predicted to be errors. 51 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Approach explored

Using part-of-speech nuclei to increase recall ◮ To increase the number of errors found, relax the requirements of what constitutes comparable strings ◮ Redefine variation nuclei: POS instead of words Example (WSJ corpus, PennTreebank3 tagset, 45 tags): (13) a. Boeing on Friday said 0 it received [NP an/DT order/NN] *ICH* from Martinair Holl
  • b. it received [NP a/DT contract/NN *ICH*] from Timken Co.
◮ But more general recurring units (POS tags) may negatively impact the precision of error detection. ◮ To use a more general representation, we also need more constraints on the disambiguating contexts. 52 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Approach explored

Identifying reliable contexts to maintain high precision ◮ Example illustrating problem that shortest non-fringe heuristic does not ensure sufficient context: (14) a. crippled * by a bitter , decade-long strike that *T* began [PP in/IN 1967/CD] and cut circulation in half
  • b. its problems began [PP in/IN [NP 1987/CD and early
1988]] when its . . . Here, the variation in structure is a correct ambiguity. ◮ What treebank information can accurately distinguish erroneous variation from legitimate ambiguity? → We explore three new heuristics, based on annotation: ◮ Heuristic 1: Shared complete bracketing ◮ Heuristic 2: Shared partial bracketing ◮ Heuristic 3: Shared vertical context 53 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 1: Shared complete bracketing

Target 1: Variation with bracketing agreement, i.e., between nuclei which are constituents (XP vs. YP) ◮ Both annotations agree on the bracketing → Significantly more likely that variation in label is an error Ex.: RB JJ varies between NP and wrong ADJP in 4-gram: (15) a. This was [NP too/RB much/JJ] for James Oakes , the court ’s chief judge .
  • b. Avondale was notified * by Louisiana officials in 1986 that
it was [ADJP potentially/RB responsible/JJ] for a cleanup at an oil-recycling plant . Heuristic 1: Shared complete bracketing is comparable context 54 / 91
slide-7
SLIDE 7 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 2: Shared partial bracketing

Target 2: Variation between constituent and non-constituent (16) a. crippled * by a bitter , decade-long strike that *T* began [PP in/IN 1967/CD] and cut circulation in half
  • b. its problems began [PP in/IN [NP 1987/CD and early
1988]] when its . . . ◮ Legitimate attachment difference here because ◮ “in 1967” forms a complete VP with began, but “in 1987” does not → one word of surrounding context is not sufficient to distinguish the two cases ◮ Can we define a heuristic to reduce risk of attachment ambiguities? 55 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 2: Shared partial bracketing

Heuristic 2: Require one extra word on side(s) without shared bracket. Example: Erroneous variation for the variation nucleus VBG JJ NNS (17) he stayed inside the Capitol * [VP [VP monitoring/VBG tax-and-budget/JJ talks/NNS] instead of flying to San Francisco . . . ] (18) one of the first bids under new takeover rules aimed * at * [VP encouraging/VBG open/JJ bids/NNS instead of gradual accumulation of large stakes] . ◮ The constituent and the nil string share a left (vp) bracket, but not a right one. ◮ Requiring extra word of context on right side (of) supports that these cases are indeed comparable. 56 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 3: Shared vertical context

Target 3: Variation in bracketing of nucleus, but shared bracketing in n-gram Example: erroneous variation for nucleus RB JJR IN CD: (19) a. will be diluted * to [NP [QP slightly/RB less/JJR than/IN 50/CD] %] after . . .
  • b. will fall
to [NP slightly/RB more/JJR than/IN 11/CD %] from slightly more than 14 % . ◮ Considering n-gram with additional token (%) to right of nucleus provides shared complete bracketing (NP). Heuristic 3: Shared bracketing of n-gram resulting from adding a word of context to left and/or right of nucleus. 57 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results for POS nuclei

After generalizing the nuclei from words to POS, we obtain ◮ 50,396 variation nuclei for the WSJ ◮ 16,598 of which remain after removing nuclei which are single null elements (cf. Dickinson & Meurers 2003b) ◮ Significantly higher than the 3,619 comparable cases with word nuclei To gauge performance of POS nuclei: ◮ Sampled 100 cases from 16,598 to examine by hand ◮ 28% point to an error ◮ 4,647 estimated cases of errors, which is a significant improvement in recall over 2,745 for word nuclei 58 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results with heuristics

◮ Heuristics select 6,343 variation nuclei of 16,598: ◮ Heuristic 1 (Shared complete bracketing): 1,339 ◮ Heuristic 2 (Shared partial bracketing): 3,731 ◮ Heuristic 3 (Shared vertical context): 1,273 ◮ Inspected random sample of 100 cases to judge precision: Overall 68.69% Heuristic 1 61% Heuristic 2 61% Heuristic 3 85% ◮ Estimate 4,357 errors from 6,343 cases, 59% increase in recall over estimated 2,745 errors with word nuclei ◮ 73 cases not covered by heuristics: 8.22% precision ⇒ New heuristics cover most cases, approaching high precision of word nucleus method while increasing recall. 59 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Limitations of POS nuclei

◮ Generalizing from word to POS nuclei is not always successful, i.e., POS class not fine grained enough. ◮ Example: variation trigram “remains JJ for” (20) a. a virus that *T* [VP remains [ADJP active/JJ] [PP for a few days]]
  • b. remains [ADJP responsible/JJ for the individual
policy services department] ◮ Depends upon particular adjective in determining how the for phrase attaches ◮ One could explore refining or lexicalizing some part-of-speech classes to account for such differences. 60 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Alternatives ways to increase recall

◮ Use more general types of context (e.g., POS tags, Dickinson 2005; Dickinson & Meurers 2005b) ◮ 8,715 shortest non-fringe variation nuclei, with an estimated 53% error detection precision ◮ Could be combined with the POS nucleus approach using the new heuristics. ◮ Immediate dominance variation method (Dickinson & Meurers 2005c) based on RHSs of treebank rules ◮ Overlaps with shared complete bracketing cases when RHS is complete sequence of POS tags ◮ Mainly only handles errors stemming from variation in labeling and not bracketing errors → Separate slides on exploring endocentricity 61 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Summary for increasing recall of constituency error detection

◮ Increased error detection recall for syntactic annotation by generalizing nature of comparable recurring unit ◮ Generalized variation nuclei of variation n-gram method to POS tags instead of using identical surface forms ◮ Determined additional contextual heuristics for errors 62 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Variation Detection for Dependency Annotation

(Boyd, Dickinson & Meurers 2008) ◮ A range of high-quality dependency treebanks for a variety of different languages are available, e.g.: ◮ Prague Dependency Treebank (PDT) of Czech (Hajiˇ c et al. 2001) ◮ Alpino Dependency Treebank of Dutch (van der Beek et al. 2001) ◮ Talbanken05 corpus of Swedish (Nivre et al. 2006) ◮ Arboretum treebank for Danish (Bick 2003) ◮ Danish Dependency Treebank (Kromann et al. 2004) ◮ Multi-lingual dependency parsing highlighted by 2006 CoNLL-X Shared Task ◮ As far as we are aware, little work has been done on automatically detecting errors in dependency treebanks. 63 / 91
slide-8
SLIDE 8 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Dependency annotation

Some characteristics ◮ Dependency annotation ◮ captures grammatical relations between words ◮ can relate non-adjacent elements ◮ may include non-projectivity, i.e., dependency arcs can cross ◮ Example from Talbanken05 corpus (Nivre et al. 2006): (21) DT SS DT OO Deras utbildning tar 345 dagar Their education takes 345 days 64 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Corpora used for dependency error detection

◮ Explore approach on the basis of three diverse dependency annotation schemes for three languages: ◮ Talbanken05 corpus of Swedish (Nivre et al. 2006) ◮ approx. 320,000 tokens ◮ distinguishes 69 dependency relations ◮ Prague Dependency Treebank (PDT 2.0) of Czech (Hajiˇ c et al. 2003), Analytical layer (surface syntax): ◮ 1.5 million tokens (88,000 sentences) ◮ distinguishes 28 dependency relations ◮ Tiger Dependency Bank (TigerDB) of German (Forst et al. 2004) ◮ semi-automatically derived from the Tiger Treebank (Brants et al. 2002), a corpus of German newspaper text taken from the Frankfurter Rundschau. ◮ 36,326 tokens (1,868 sentences) ◮ distinguishes 53 dependency relations, following English PARC 700 Dependency Bank (King et al. 2003), including sublexical and abstract nodes 65 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Adapting the method to dependency annotation

◮ What is involved in applying the variation n-gram method to dependency annotation? ◮ Mapping from a pair of words to their dependency relation label, we have variation nuclei of size 2. ◮ We encode the head information into the label: ◮ R means the head is on the right ◮ L for the left ◮ Example from Talbanken05 corpus: (21) DT SS DT OO Deras utbildning tar 345 dagar Their education takes 345 days ◮ utbildning tar: SS-R ◮ tar dagar: OO-L 66 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Applying the variation n-gram method

With the dependency annotated data encoded in this way, there are three different possibilities for errors: ◮ Errors in labeling: SUBJ vs. OBJ ◮ Errors in what the head is: OBJ-L vs. OBJ-R ◮ Errors in dependency identification: OBJ vs. NIL What needs to be added to the basic picture? Take the nature of dependency annotation into account in ◮ defining the set of variations that need to be considered ◮ determining a notion of context sufficient to identify the variations which are errors → heuristics 67 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Errors in dependency identification

◮ The existence or absence of a dependency is captured by variation with the special label NIL ◮ E.g. DT vs. NIL in the following Talbanken05 example: (22) a. RA DT DT PA HD HD Backberger s¨ ager i sin artikel ’ Den heliga familjen . . . Backberger says in her article ’ The sacred family . . . b. RA DT DT PA HD HD Backberger skriver i sin artikel ’ Den heliga familjen . . . Backberger writes in her article ’ The sacred family . . . ◮ Search for NIL pairs restricted to within same sentence 68 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Accounting for the nature of dependencies

◮ Variation n-gram approach for constituency useful starting point, but how do we adapt it to the nature of dependencies? ◮ There are formal and linguistic issues in adapting the method from constituency to dependency annotation, including: ◮ Overlap ◮ Contiguity 69 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Constituency and Dependency

Overlap ◮ Two phrases in a constituency-based representation never share any of their daughters, unless one is properly included in the other. ◮ In a dependency representation, the same head may participate in multiple dependency relations, causing dependency pairs to overlap in one token. ◮ Example: tar is the head of two dependencies in (21) (21) DT SS DT OO Deras utbildning tar 345 dagar Their education takes 345 days 70 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Constituency and Dependency

Dealing with overlap (23) DO IO She showed the department chair the beautiful old chair . ⇒ Set up variation detection to compare sets of all dependencies between the words in the nucleus. ◮ e.g., for the case above: < showed, chair, {DO, IO} > 71 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Constituency and Dependency

Contiguity ◮ Within traditional constituency frameworks, the sisters in a local tree are contiguous, i.e., their terminal yield is a continuous string. ◮ For dependency annotation, a dependency graph will
  • ften relate non-contiguous elements.
(24) SS PL AA PA Handeln ger tillbaka med dem Commerce gives back with them ⇒ Ensure that intervening material is considered as context. 72 / 91
slide-9
SLIDE 9 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Single-head constraint for dependencies

◮ Dependency annotation schemes differ in whether they assume a single-head constraint, i.e., whether every word is a dependent of exactly one head. ◮ E.g., the TigerDB does not satisfy this constraint: (25) SB MO DET OA OC INF Wer aber soll den Schiedsrichter spielen ? Who but shall the referee play ? SB ◮ Variation detection checks each mappings from nucleus to its annotation independently, so no single-head assumption is needed. 73 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Indirect annotation

◮ Variation n-gram approach is strictly data driven ◮ In mapping from words to dependency labels, each dependency relation label is considered independent of the others. ◮ Locality assumption similar to the well-known independence assumption for local trees in PCFGs. ◮ In some dependency treebanks, no such locality requirement is enforced: some labels are based upon annotation decisions elsewhere in the graph. ◮ Examples for such indirect dependency encoding: prepositions, complementizers, coordination in the PDT (analytical layer). 74 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Indirect annotation

Example: Coordination (26) a. Atr Sb Pred AuxP Adv Nejlevnˇ ejˇ s´ ı telefony jsou v Brit´ anii cheapest telephones are in Britain b. AuxP Adv Pred Sb Co Coord Sb Co Na pokoj´ ıch jsou telefony a faxy in rooms are telephones and fax machines 75 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Indirect annotation

Example: Prepositions (27) a. AuxP Atr utk´ an´ ı v Brnˇ e game in Brno Noun Prep Noun b. AuxP Adv zadrˇ zen v Brnˇ e detained in Brno Verb Prep Noun 76 / 91

Indirect annotation

Example: Indirection can cross significant contexts (28) a. Atr Atr AuxP Atr Atr Co Coord AuxP Atr Co Oblastn´ ı sdruˇ zen´ ı ODS na severn´ ı Moravˇ e a ve Slezsku regional branches of ODS in Northern Moravia and in Silesia Adj Noun Noun Prep Adj Noun Conj Prep Noun b. AuxP Atr Adv Co Coord AuxP Adv Co na severn´ ı Moravˇ e a ve Slezsku sp´ ach´ ano in Northern Moravia and in Silesia committed Prep Adj Noun Conj Prep Noun Verb Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Indirect annotation

Possible recoding of some cases as local to head Original: (29) a. AuxP Atr utk´ an´ ı v Brnˇ e game in Brno Noun Prep Noun b. AuxP Adv zadrˇ zen v Brnˇ e detained in Brno Verb Prep Noun Recoded as: (30) a. Atr AuxP utk´ an´ ı v Brnˇ e game in Brno Noun Prep Noun b. Adv AuxP zadrˇ zen v Brnˇ e detained in Brno Verb Prep Noun 78 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Adapting the variation nuclei algorithm

  • 1. Compute the set of nuclei:
a) Store all dependency pairs with dependency label. ◮ The dependency relations annotated in the corpus are handled as nuclei of size two and mapped to their label plus a marker of the head (L/R). ◮ The labels of overlapping type-identical nuclei are collapsed into a set of labels. b) For each distinct pair of words stored as dependency, search for non-dependency occurrences of words and add the nuclei with label NIL. ◮ A trie data structure is used to store all potential nuclei and to guide the search for NIL nuclei. ◮ Search is limited to pairs within same sentence. ◮ NIL nuclei which overlap with a genuine dependency are not considered.
  • 2. Compute the set of variation nuclei by determining
which of the stored nuclei have more than one label. 79 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Disambiguating contexts for dependencies

Question: How can we detect which variations are errors, given that dependencies are generally non-adjacent? ◮ How much context do we need to determine if a variation is an error? So far, we have relied on immediately surrounding context, which is cognitively plausible. Do dependencies need ◮ more context? Many dependencies are non-adjacent. ◮ less context? More information is now encoded in the nucleus itself (head information). 80 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 1: NIL internal context heuristic

◮ The general non-fringe heuristic is applicable: an element in a dependency nucleus is non-fringe iff it is next to a context word or the other word in the nucleus ◮ Are there less strict, reliable dependency annotation context heuristics than the non-fringe heuristic? ◮ Heuristic 1: NIL internal context heuristic: ◮ only require identity of internal context when the variation involves nil ◮ Heuristic 2: Dependency context heuristic 81 / 91
slide-10
SLIDE 10 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 1: NIL internal context heuristic

Example for case predicted to be an error (31) a. SB OP OBJ OC INF “ Wirtschaftspolitik l¨ aßt auf sich warten ” economic policy lets
  • n
itself wait b. DET SB OP OBJ OC INF Die Wirtschaftspolitik l¨ aßt auf sich warten . the economic policy lets
  • n
itself wait ‘Economic policy is a long time coming.’ 82 / 91

Heuristic 1: NIL internal context heuristic

Example for case predicted not to be an error (32) a. MO DET MO OBJ in den Vereinigten Staaten in the United States b. MO DET MO NUMBER OBJ OP DET MO OBJ in den vergangenen zehn Jahren an die Vereinigten Staaten in the past ten years to the United States Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 2: Dependency context heuristic

◮ If the head of a variation nucleus is being used in the same function in all instances, the variation in the labeling of the nucleus is more likely to be an error. ◮ Conversely, when the head is used differently, it is more likely a genuine ambiguity. 84 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Heuristic 2: Dependency context heuristic

Example: head with two different functions → not an error (33) a. DT PA ++ DT CC i den ena eller andra formen in the
  • ne
  • r
  • ther
form b. DT DT ++ CC PA i den ena eller båda f¨ ardriktningarna in the
  • ne
  • r
  • ther
directions 85 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results

Talbanken 05 ◮ 197,123 tokens in 11,431 sentences in sections P and G ◮ 210 different variation nuclei using non-fringe heuristic ◮ 92.9% precision (195 error nuclei) (thanks to Joakim Nivre, Mattias Nilsson and Eva Pettersson for the evaluation) ◮ 274 error instances: ◮ 145 labelling confusion ◮ 129 dependency identification ◮ observations: ◮ common problems: determiner (DT), preposition (PA) ◮ more errors with adverbials (73) than arguments (31) 86 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results

PDT 2.0 ◮ 38,482 sentences (670,544 tokens) in full/amw section ◮ 553 different variation nuclei using non-fringe heuristic ◮ 426 cases after removing errors involving punctuation ◮ 354 cases after recoding indirect AuxP and AuxC deps. ◮ 59.7% precision (205 error nuclei) (thanks to Jirka Hana and Jan ˇ Stˇ ep´ anek for evaluation) ◮ 251 error instances ◮ 152 labeling confusion (60.6%) ◮ 99 dependency identification ◮ observations: ◮ 49% of false positives due to other indirect annotation scheme decisions (coordination) ◮ common problem with AdvAtr vs. AtrAdv, preference for adverbial of predicate vs. attribute of lower node 87 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Results

TigerDB ◮ Only used sentences with lexically-rooted dependency structures, ignoring abstract and sublexical nodes. ◮ 1,567 sentences (29,373 tokens) ◮ 276 variation nuclei, NIL internal context heuristic ◮ 48.1% precision (133 error nuclei) ◮ 149 error instances ◮ 46 labeling errors ◮ 103 dependency identification ◮ observations: ◮ consistent tokenization is a problem for multi-word expressions and proper names, e.g., Den Haag (The Hague), zur Zeit (at that time) ◮ prepositional argument vs. modifier distinction difficult, e.g., Bedarf an X (demand for X). ◮ false positives due to ambiguous tokens, for which POS disambiguation would help 88 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Outlook: Increasing recall

The issue ◮ word-word dependencies are highly specific. ◮ How can they be generalized to increase the number of recurring dependency pairs limiting the recall of the variation detection method? Specific lexical properties of head important, e.g.: ◮ Lexical information is known to improve PCFGs through head-lexicalization (e.g., Collins 1996) To characterize the dependent, POS class may be sufficient (cf. subcategorization frame in lexicalized theories of grammar). ⇒ Generalize from word-word to word-POS dependencies ◮ For nuclei not annotated as a dependency (NIL), use head-dependent orientation of string we compare it to. 89 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Outlook: Increasing recall

Tagset dependency ◮ Use of word-POS dependencies is dependent on the granularity of the POS tagset used. ◮ Talbanken05 corpus has 40 coarse-grained POS tags ◮ PDT 2.0 distinguishes 4290 POS tags (Hajiˇ c 2004) ◮ For positional tagsets, one can decide which positions
  • f the tagset to use, e.g.,
◮ including case information is likely to increase precision ◮ distinguishing comparative and superlative adjectives could decrease recall 90 / 91
slide-11
SLIDE 11 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

Summary

◮ We motivated the need for error detection in annotated corpora, and introduced the variation n-gram approach as an automatic error detection method. ◮ Research on category learning in humans provides independent evidence for the notion of context used. ◮ The method successfully detects errors in ◮ part of speech ◮ constituency, ◮ discontinuous constituency, ◮ and dependency annotation ◮ We showed that the method can provide significant feedback on annotation scheme distinctions which ◮ are not sufficiently documented, ◮ rely on representational choices not locally motivated, ◮ or cannot reliably be made based on the evidence found in the corpus, 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary

References

Abeill´ e, A. (ed.) (2003). Treebanks: Building and using syntactically annotated
  • corpora. Dordrecht: Kluwer Academic Publishers.
http://treebank.linguist.jussieu.fr/toc.html. Abney, S., R. E. Schapire & Y. Singer (1999). Boosting Applied to Tagging and PP
  • Attachment. In P
. Fung & J. Zhou (eds.), Proceedings of the 1999 Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora. pp. 38–45. Agrawal, R. & R. Srikant (1994). Fast Algorithms for Mining Association Rules in Large Databases. In J. B. Bocca, M. Jarke & C. Zaniolo (eds.), VLDB 1994. Morgan Kaufmann, pp. 487–499. Artstein, R. & M. Poesio (2009). Survey Article: Inter-Coder Agreement for Computational Linguistics. Computational Linguistics 34(4), 1–42. URL http://www.mitpressjournals.org/doi/abs/10.1162/coli.07-034-R2. Bick, E. (2003). Arboretum, a Hybrid Treebank for Danish. In Proceedings of TLT 2003 (2nd Workshop on Treebanks and Linguistic Theory. V¨ axj¨
  • , Sweden, pp.
9–20. Bies, A., M. Ferguson, K. Katz & R. MacIntyre (1995). Bracketing Guidelines for Treebank II Style Penn Treebank Project. University of Pennsylvania. ftp://ftp.cis.upenn.edu/pub/treebank/doc/manual/root.ps.gz. Blaheta, D. (2002). Handling noisy training and testing data. In Proceedings of the 7th conference on Empirical Methods in Natural Language Processing. pp. 111–116. http://www.cs.brown.edu/∼dpb/papers/dpb-emnlp02.html. 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary Blevins, J. (1990). Syntactic Complexity: Evidence for Discontinuity and
  • Multidomination. Ph.D. thesis, University of Massachusetts, Amherst, MA.
Bonami, O., D. Godard & J.-M. Marandin (1999). Constituency and word order in French subject inversion. In G. Bouma, E. W. Hinrichs, G.-J. M. Kruijff & R. T. Oehrle (eds.), Constraints and Resources in Natural Language Syntax and Semantics, Stanford, CA: CSLI Publications, Studies in Constraint-Based Lexicalism, pp. 21–40. Boyd, A., M. Dickinson & D. Meurers (2007). Increasing the Recall of Corpus Annotation Error Detection. In Proceedings of the Sixth Workshop on Treebanks and Linguistic Theories (TLT-07). Bergen, Norway. URL http://purl.org/dm/papers/boyd-et-al-07b.html. Boyd, A., M. Dickinson & D. Meurers (2008). On Detecting Errors in Dependency
  • Treebanks. Research on Language and Computation 6(2), 113–137. URL
http://purl.org/dm/papers/boyd-et-al-08.html. Brants, S., S. Dipper, S. Hansen, W. Lezius & G. Smith (2002). The TIGER
  • Treebank. In Proceedings of the Workshop on Treebanks and Linguistic
  • Theories. Sozopol, Bulgaria. www.bultreebank.org/proceedings/paper03.pdf.
Brants, T. & W. Skut (1998). Automation of Treebank Annotation. In Proceedings of New Methods in Language Processing (NeMLaP-98). Syndey. http://www.coli.uni-sb.de/∼thorsten/publications/Brants-Skut-NeMLaP98.ps.gz Br¨
  • ker, N. (1998). Separating Surface Order and Syntactic Relations in a
Dependency Grammar. In Proceedings of the 17th International Conference
  • n Computational Linguistics (COLING) and the 36th Annual meeting of the
ACL (ACL). Montreal. Calder, J. (1997). On aligning trees. In Proceedings of the Second Conference of Empirical Methods in Natural Language Processing. Brown University. http://xxx.lanl.gov/abs/cmp-lg/9707016. 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary Chemla, E., T. H. Mintz, S. Bernal & A. Christophe (2009). Categorizing words using ‘frequent frames’: what cross-linguistic analyses reveal about distributional acquisition strategies. Developmental Science 12(3). URL http://dx.doi.org/10.1111/j.1467-7687.2009.00825.x. Collins, M. J. (1996). A New Statistical Parser Based on Bigram Lexical
  • Dependencies. In A. Joshi & M. Palmer (eds.), Proceedings of the
Thirty-Fourth Annual Meeting of the Association for Computational Linguistics. San Francisco: Morgan Kaufmann Publishers, pp. 184–191. URL citeseer.nj.nec.com/collins96new.html. Dickinson, M. (2005). Error detection and correction in annotated corpora. Ph.D. thesis, The Ohio State University. Dickinson, M. & W. D. Meurers (2003a). Detecting Errors in Part-of-Speech
  • Annotation. In Proceedings of the 10th Conference of the European Chapter of
the Association for Computational Linguistics (EACL-03). Budapest, Hungary,
  • pp. 107–114. http://ling.osu.edu/∼dm/papers/dickinson-meurers-03.html.
Dickinson, M. & W. D. Meurers (2003b). Detecting Inconsistencies in Treebanks. In Proceedings of the Second Workshop on Treebanks and Linguistic Theories (TLT-03). V¨ axj¨
  • , Sweden, pp. 45–56.
http://ling.osu.edu/∼dm/papers/dickinson-meurers-tlt03.html. Dickinson, M. & W. D. Meurers (2004). Error detection with discontinuous
  • constituents. In P
. Rodrigues, D. Cavar & J. Herring (eds.), Proceedings of the First Midwest Computational Linguistics Colloquium. Bloomington, Indiana. Dickinson, M. & W. D. Meurers (2005a). Detecting Annotation Errors in Spoken Language Corpora. In The Special Session on treebanks for spoken language and discourse at NODALIDA-05. Joensuu, Finland. URL http://ling.osu.edu/∼dickinso/papers/dickinson-meurers-nodalida05.html. 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary Dickinson, M. & W. D. Meurers (2005b). Detecting Errors in Discontinuous Structural Annotation. In Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL ’05). pp. 322–329. http://www.aclweb.org/anthology/P/P05/P05-1040. Dickinson, M. & W. D. Meurers (2005c). Prune Diseased Branches to Get Healthy Trees! How to Find Erroneous Local Trees in a Treebank and Why It Matters. In Proceedings of the Fourth Workshop on Treebanks and Linguistic Theories (TLT 2005). Barcelona, Spain. URL http://ling.osu.edu/∼dm/papers/dickinson-meurers-tlt05.html. Donohue, C. & I. A. Sag (1999). Domains in Warlpiri. In Abstracts of the Sixth Int. Conference on HPSG. Edinburgh: University of Edinburgh, pp. 101–106. http://www-csli.stanford.edu/∼sag/papers/warlpiri.ps. Dowty, D. R. (1996). Towards a Minimalist Theory of Syntactic Structure. In H. Bunt & A. van Horck (eds.), Discontinuous Constituency, Berlin and New York, NY: Mouton de Gruyter, vol. 6 of Natural Language Processing. Eskin, E. (2000). Automatic Corpus Correction with Anomaly Detection. In Proceedings of the First Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-00). Seattle, Washington. http://www.cs.columbia.edu/∼eeskin/papers/treebank-anomaly-naacl00.ps. Forst, M., N. Bertomeu, B. Crysmann, F. Fouvry, S. Hansen-Schirra & V. Kordoni (2004). Towards a Dependency-Based Gold Standard for German Parsers. The TIGER Dependency Bank. In S. Hansen-Schirra, S. Oepen & H. Uszkoreit (eds.), 5th International Workshop on Linguistically Interpreted Corpora (LINC-04) at COLING. Geneva, Switzerland: COLING, pp. 31–38. URL http://aclweb.org/anthology/W04-1905. Hajiˇ c, J. (2004). Disambiguation of Rich Inflection (Computational Morphology of Czech). Karolinum, Charles Univeristy Press, Prague, Czech Republic. 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary Hajiˇ c, J., A. B¨
  • hmov´
a, E. Hajiˇ cov´ a & B. Vidov´ a-Hladk´ a (2003). The Prague Dependency Treebank: A Three-Level Annotation Scenario. In Abeill´ e (2003),
  • chap. 7, pp. 103–127. URL
http://ufal.mff.cuni.cz/pdt2.0/publications/HajicHajicovaAl2000.pdf. http://treebank.linguist.jussieu.fr/toc.html. Hajiˇ c, J., B. Hladk´ a & P . Pajas (2001). The Prague Dependency Treebank: Annotation Structure and Support. In IRCS Workshop on Linguistic Databases. Hepple, M. (1994). Discontinuity and the Lambek Calculus. In Proceedings of the 15th Conference on Computational Linguistics (COLING-94). Kyoto. ftp://ftp.dcs.shef.ac.uk/home/hepple/papers/coling94.ps. Hirakawa, H., K. Ono & Y. Yoshimura (2000). Automatic Refinement of a POS Tagger Using a Reliable Parser and Plain Text Corpora. In Proceedings of the 18th International Conference on Computational Linguistics (COLING). Saarbr¨ ucken, Germany: ICCL. Hockenmaier, J. & M. Steedman (2005). CCGbank: User’s Manual. Tech. Rep. MS-CIS-05-09, Department of Computer Science and Information Science, University of Pennsylvania, Philadelphia. Huck, G. (1985). Exclusivity and discontinuity in phrase structure grammar. In West Coast Conference on Formal Linguistics (WCCFL). Stanford University, CSLI Publications, vol. 4, pp. 92–98. Huck, G. & A. Ojeda (eds.) (1987). Discontinuous Constituency. No. 20 in Syntax and Semantics. New York, et al.: Academic Press. Johansson, S. (1986). The Tagged LOB Corpus: Users’ Manual. Norwegian Computing Centre for the Humanities, Bergen. Kathol, A. (1995). Linearization-Based German Syntax. Ph.D. thesis, Ohio State University, Columbus, OH. Revised version published by Oxford University Press, 2000. 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary King, T. H., R. Crouch, S. Riezler, M. Dalrymple & R. M. Kaplan (2003). The PARC 700 Dependency Bank. In Proceedings of the 4th International Workshop on Linguistically Interpreted Corpora, held at the 10th Conference of the European Chapter of the Association for Computational Linguistics (EACL-03). Budapest. URL http://www2.parc.com/isl/groups/nltt/fsbank/. Kroch, A. S. & A. K. Joshi (1987). Analyzing Extraposition in a Tree Adjoining
  • Grammar. In Huck & Ojeda (1987).
Kromann, M. T., L. Mikkelsen & S. K. Lynge (2004). Danish Dependency Treebank: Annotation Guide. http://www.id.cbs.dk/∼mtk/treebank/guideT.html. Kvˇ etˇ
  • n, P
. & K. Oliva (2002). Achieving an Almost Correct PoS-Tagged Corpus. In P . Sojka, I. Kopeˇ cek & K. Pala (eds.), Text, Speech and Dialogue 5th International Conference, TSD 2002, Brno, Czech Republic, September 9-12,
  • 2002. Heidelberg: Springer, no. 2448 in Lecture Notes in Artificial Intelligence
(LNAI), pp. 19–26. Leech, G. (1997). A Brief Users’ Guide to the Grammatical Tagging of the British National Corpus. UCREL, Lancaster University, Lancaster. http://www.hcu.ox.ac.uk/BNC/what/gramtag.html. Lezius, W., H. Biesinger & C. Gerstenberger (2002). TIGERRegistry Manual. IMS, University of Stuttgart. MacWhinney, B. (2000). The CHILDES project: Tools for analyzing talk. Mahwah, NJ: Lawrence Erlbaum Associates, third edition ed. Marcus, M., B. Santorini & M. A. Marcinkiewicz (1993). Building a large annotated corpus of English: The Penn Treebank. Computational Linguistics 19(2), 313–330. ftp://ftp.cis.upenn.edu/pub/treebank/doc/cl93.ps.gz. Marcus, M., B. Santorini, M. A. Marcinkiewicz & A. Taylor (1999). Treebank-3
  • Corpus. Linguistic Data Consortium. Philadelphia.
http://www.ldc.upenn.edu/Catalog/CatalogEntry.jsp?catalogId=LDC99T42. 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary McCawley, J. D. (1982). Parentheticals and discontinuous constituent structure. Linguistic Inquiry 13(1), 91–106. Meurers, W. D. (2005). On the use of electronic corpora for theoretical linguistics. Case studies from the syntax of German. Lingua 115(11), 1619–1639. http://ling.osu.edu/∼dm/papers/meurers-03.html. Mintz, T. H. (2002). Category induction from distributional cues in an artificial
  • language. Memory & Cognition 30, 678–686.
Mintz, T. H. (2003). Frequent frames as a cue for grammatical categories in child directed speech. Cognition 90, 91–117. Morrill, G. V. (1995). Discontinuity in categorial grammar. Linguistics and Philosophy 18, 175–219. M¨ uller, F. H. & T. Ule (2002). Annotating topological fields and chunks – and revising POS tags at the same time. In Proceedings of COLING. http://ling.osu.edu/∼dm/02/spring/795K/mueller-ule.ps. M¨ uller, S. (1999). Deutsche Syntax deklarativ. Head-Driven Phrase Structure Grammar f¨ ur das Deutsche. No. 394 in Linguistische Arbeiten. T¨ ubingen: Max Niemeyer Verlag. Nivre, J., J. Nilsson & J. Hall (2006). Talbanken05: A Swedish Treebank with Phrase Structure and Dependency Annotation. In Proceedings of the fifth international conference on Language Resources and Evaluation (LREC2006). Genoa, Italy. Ojeda, A. (1987). Discontinuity, multidominances and unbounded dependency in Generalized Phrase Structure Grammar. In Huck & Ojeda (1987). Oliva, K. (2001). The Possibilities of Automatic Detection/Correction of Errors in Tagged Corpora: A Pilot Study on a German Corpus. In V. Matouˇ sek, 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary P . Mautner, R. Mouˇ cek & K. Tauˇ ser (eds.), Text, Speech and Dialogue. 4th International Conference, TSD 2001, Zelezna Ruda, Czech Republic, September 11-13, 2001, Proceedings. Springer, vol. 2166 of Lecture Notes in Computer Science, pp. 39–46. Padro, L. & L. Marquez (1998). On the Evaluation and Comparison of Taggers: the Effect of Noise in Testing Corpora. In COLING-ACL. pp. 997–1002. URL citeseer.ist.psu.edu/padro98evaluation.html. Penn, G. (1999). Linearization and WH-extraction in HPSG: Evidence from Serbo-Croatian. In R. D. Borsley & A. Przepi´
  • rkowski (eds.), Slavic in HPSG,
Stanford, CA: CSLI Publications, pp. 149–182. Pl´ atek, M., T. Holan, V. Kuboˇ n & K. Oliva (2001). Word-Order Relaxations and Restrictions within a Dependency Grammar. In G. Satta (ed.), Proceedings of the Seventh International Workshop on Parsing Technologies (IWPT). Beijing: Tsinghua University Press, pp. 237–240. Rambow, O. & A. Joshi (1994). A Formal Look at Dependency Grammars and Phrase-Structure Grammars, with Special Consideration of Word-Order
  • Phenomena. In L. Wanner (ed.), Current Issues in Meaning-Text-Theory,
London: Pinter. http://arxiv.org/abs/cmp-lg/9410007. Reape, M. (1993). A Formal Theory of Word Order: A Case Study in West
  • Germanic. Ph.D. thesis, University of Edinburgh, Edinburgh.
Richter, F. & M. Sailer (2001). On the Left Periphery of German Finite Sentences. In W. D. Meurers & T. Kiss (eds.), Constraint-Based Approaches to Germanic Syntax, Stanford, CA: CSLI Publications, Studies in Constraint-Based Lexicalism, pp. 257–300. Sampson, G. & A. Babarczy (2003). Limits to annotation precision. In Proceedings
  • f the 4th International Workshop on Linguistically Interpreted Corpora
(LINC-03). pp. 61–68. http://www.grsampson.net/Alta.html. 91 / 91
slide-12
SLIDE 12 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary Santorini, B. (1990). Part-Of-Speech Tagging Guidelines for the Penn Treebank Project (3rd Revision, 2nd printing). Ms., UPenn. Schiller, A., S. Teufel & C. Thielen (1995). Guidlines f¨ ur das Taggen deutscher Textcorpora mit STTS. Tech. rep., IMS-CL, Univ. Stuttgart and SfS, Univ. T¨
  • ubingen. http://www.cogsci.ed.ac.uk/∼simone/stts guide.ps.gz.
Skut, W., T. Brants, B. Krenn & H. Uszkoreit (1998). A Linguistically Interpreted Corpus of German Newspaper Text. In Proceedings of the ESSLLI Workshop
  • n Recent Advances in Corpus Annotation. Saarbr¨
ucken, Germany. http: //www.coli.uni-sb.de/∼thorsten/publications/Skut-ea-ESSLLI-Corpus98.ps.gz Skut, W., B. Krenn, T. Brants & H. Uszkoreit (1997). An Annotation Scheme for Free Word Order Languages. In Proceedings of the 5th Conference on Applied Natural Language Processing (ANLP). Washington, D.C. http://www.coli.uni-sb.de/∼thorsten/publications/Skut-ea-ANLP97.ps.gz. Ule, T. & K. Simov (2004). Unexpected Productions May Well be Errors. In Proceedings of Fourth International Conference on Language Resources and Evaluation (LREC 2004). Lisbon, Portugal. http://www.sfs.uni-tuebingen.de/∼ule/Paper/us04lrec.pdf. van der Beek, L., G. Bouma, R. Malouf & G. van Noord (2001). The Alpino Dependency Treebank. In Computational Linguistics in the Netherlands (CLIN) 2001, Amsterdam: Rodopi. van Halteren, H. (2000). The Detection of Inconsistency in Manually Tagged Text. In A. Abeill´ e, T. Brants & H. Uszkoreit (eds.), Proceedings of the Second Workshop on Linguistically Interpreted Corpora (LINC-00). Luxembourg. van Halteren, H., W. Daelemans & J. Zavrel (2001). Improving Accuracy in Word Class Tagging through the Combination of Machine Learning Systems. Computational Linguistics 27(2), 199–229. 91 / 91 Detecting Errors in Corpus Annotation Detmar Meurers University of T¨ ubingen Introduction Effects of Annotation Errors How to obtain high quality Part of Speech Variation detection Computing variation n-grams Independent evidence from language acquisition Results for the WSJ Annotation scheme feedback Constituency Variation detection Computing variation n-grams WSJ case study Results Discontinuity Computing var. n-grams Results for TIGER Increasing recall Results Dependency Variation detection Nature of dependencies Indirect annotation Increasing recall Summary Voutilainen, A. & T. J¨ arvinen (1995). Specifying a shallow grammatical representation for parsing purposes. In Proceedings of the 7th Conference of the EACL. Dublin, Ireland. http://www.aclweb.org/anthology/E95-1029. Xiao, L., X. Cai & T. Lee (2006). The development of the verb category and verb argument structures in Mandarin-speaking children before two years of age. Paper presented at The Seventh Tokyo Conference on Psycholinguistics. Keio University. 91 / 91