readability assessment for sentences
play

Readability Assessment for Sentences Introduction Motivation, - PowerPoint PPT Presentation

Readability Assessment for Sentences Sowmya Vajjala Readability Assessment for Sentences Introduction Motivation, Methods and Evaluation Background Our Approach Corpora Features Modeling Sowmya Vajjala Relative Comparison Ranking


  1. Readability Assessment for Sentences Sowmya Vajjala Readability Assessment for Sentences Introduction Motivation, Methods and Evaluation Background Our Approach Corpora Features Modeling Sowmya Vajjala Relative Comparison Ranking (with Detmar Meurers) Conclusions Center for Language Technology University of Gothenburg, Sweden 20 November 2014 1 / 29

  2. Readability What is readability analysis? Assessment for Sentences Sowmya Vajjala Introduction We want to measure how difficult it is to read a text, Background Our Approach ◮ based on properties of the text , using criteria which are Corpora ◮ data-induced: using corpora with graded texts Features Modeling ◮ theory-driven: constructs known to reflect complexity Relative Comparison Ranking Conclusions ◮ given a purpose , e.g., ◮ humans: to support reading and comprehension ◮ read texts at a specific level of language proficiency. ◮ carry out specific tasks (e.g., answer questions) etc., ◮ machines: evaluation of generation systems ◮ sometimes personalized to a user, through information- ◮ obtained directly (e.g., questionnaire), or ◮ indirectly (e.g, inferred from nature of a search query) 2 / 29

  3. Readability Why do we need readability for sentences? Assessment for Sentences some application scenarios Sowmya Vajjala Introduction Background Our Approach Corpora ◮ selecting appropriate sentences for language learners Features Modeling in CALL. (Segler 2007; Pil´ an et al. 2013, 2014) Relative Comparison Ranking ◮ understanding the difficulty of survey questions Conclusions (Lenzner 2013) ◮ predicting sentence fluency in Machine Translation (Chae & Nenkova 2009) ◮ for text simplification (Vajjala & Meurers 2014a; Dell’Orletta et al. 2014) 3 / 29

  4. Readability Why do WE need it? Assessment for Sentences Sowmya Vajjala Introduction Background Our Approach Corpora Features Modeling Relative Comparison ◮ identifying target sentences for text simplification. Ranking Conclusions ◮ evaluating text simplification approaches. 4 / 29

  5. Readability Our Approach: Overview Assessment for Sentences Sowmya Vajjala Introduction Background ◮ Corpus : publicly accessible, sentence level corpora Our Approach (texts not prepared by us) Corpora Features Modeling ◮ Features : from Vajjala & Meurers (2014b), that work Relative Comparison Ranking well at a text level. Conclusions ◮ Modeling : 1. binary classification (easy vs difficult) 2. apply document level regression model on sentences. 3. pair-wise ranking ◮ Evaluation : within and cross corpus evaluations with multiple real-life datasets 5 / 29

  6. Readability Corpora-1: Wikipedia-SimpleWikipedia Assessment for Sentences Sowmya Vajjala Introduction Background ◮ Zhu et al. (2010) created a publicly available, sentence Our Approach Corpora aligned corpus from Wikipedia and Simple Wikipedia. Features Modeling Relative Comparison ◮ ∼ 80,000 pairs of sentences in simplified and Ranking Conclusions unsimplified versions. ◮ Example pair: 1. Wiki: Chinese styles vary greatly from era to era and are traditionally named after the ruling dynasty. 2. Simple Wiki: There are many Chinese artistic styles, which are usually named after the ruling dynasty. 6 / 29

  7. Readability Corpora-2: OneStopEnglish.com Assessment for Sentences Sowmya Vajjala Introduction Background Our Approach ◮ OneStopEnglish (OSE) is an English teachers resource Corpora website published by the Macmillan Education Group. Features Modeling Relative Comparison ◮ They publish Weekly News Lessons which consist of Ranking news articles sourced from The Guardian . Conclusions ◮ The articles are rewritten by teaching experts for English language learners at three reading levels (elementary, intermediate, advanced) ◮ We obtained permission to collect articles and compiled a corpus of 76 article triplets (228 in total) 7 / 29

  8. Readability Corpora-2: OneStopEnglish.com Assessment for Sentences sentence aligned corpus creation Sowmya Vajjala Introduction Background Our Approach ◮ creation process: Corpora Features 1. parse the pdf files and extract text content. Modeling Relative Comparison Ranking 2. split all texts into sentences. Conclusions 3. compare sentences between versions and match them by their cosine similarity (Nelken & Shieber 2006) . ◮ two versions of the corpus: 1. OSE2Corpus: ∼ 3000 sentence pairs. 2. OSE3Corpus: ∼ 850 sentence triplets. * contact me if anyone wants to use this corpus. 8 / 29

  9. Readability OSE Corpus: Example Assessment for Sentences Sowmya Vajjala Introduction Background Our Approach Corpora Features adv: In Beijing, mourners and admirers made their way to lay Modeling flowers and light candles at the Apple Store. Relative Comparison Ranking Conclusions inter: In Beijing, mourners and admirers came to lay flowers and light candles at the Apple Store. ele: In Beijing, people went to the Apple Store with flowers and candles. 9 / 29

  10. Readability Features-1 Assessment for Sentences From Vajjala & Meurers (2014b) Sowmya Vajjala Introduction ◮ Lexical Features Background Our Approach ◮ lexical richness features from Second Language Corpora Acquisition (SLA) research Features Modeling ◮ e.g., Type-Token ratio, noun variation, . . . Relative Comparison Ranking ◮ POS density features Conclusions ◮ e.g., # nouns/# words, # adverbs/# words, . . . ◮ traditional features and formulae ◮ e.g., # sentence length in words . . . ◮ Syntactic Features ◮ syntactic complexity features from SLA research. ◮ e.g., # dep. clauses/clause, average clause length, . . . ◮ other parse tree features ◮ e.g., # NPs per sentence, avg. parse tree height, . . . 10 / 29

  11. Readability Features -2 Assessment for Sentences Sowmya Vajjala Introduction Background Our Approach ◮ Morphological properties of words Corpora Features ◮ e.g., Does the word contain a stem along with an affix? Modeling Relative Comparison abundant=abound+ant Ranking ◮ Age of Acquisition (AoA) Conclusions ◮ average age-of-acquisition of words in a text ◮ Other Psycholinguistic features ◮ e.g., word abstractness ◮ Avg. number of senses per word (obtained from WordNet) 11 / 29

  12. Readability Sentence Readability: Binary Classification Assessment for Sentences Vajjala & Meurers (2014a) Sowmya Vajjala Introduction Background Our Approach Corpora Features ◮ We started with training a sentence-level readability Modeling Relative Comparison model on Wikipedia corpus: Ranking ◮ Binary classification: simple – hard Conclusions ◮ 65–68% accuracy, depending on training set size. ◮ increasing training sample size from 10K to 80K samples did not improve the accuracy much! ◮ As regression: r = 0 . 4 ◮ Why is it so bad? 12 / 29

  13. Readability What is the problem? Assessment for Sentences Sowmya Vajjala ◮ What happens if we just apply a document level Introduction readability model on this corpus? Background ◮ Model (Vajjala & Meurers 2014b): outputs readability Our Approach Corpora score on a scale of 1-5, 5 being difficult. Features Modeling Relative Comparison 50 Ranking Wiki Simple Wiki Conclusions Percentage of the total sentences at that level 45 40 35 30 25 20 15 10 5 1 1.5 2 2.5 3 3.5 4 4.5 5 Reading level 13 / 29

  14. Readability What can we infer? Assessment for Sentences Sowmya Vajjala Introduction Background Our Approach Corpora ◮ There are all sorts of sentences in both versions. Features Modeling ◮ Wikipedia has more sentences at higher reading levels Relative Comparison Ranking than Simple Wikipedia. Conclusions - Is this the reason binary classification failed? ◮ one idea: A simple sentence is only simpler than its unsimplified version. It can also still be hard. ⇒ Simplification could be relative, not absolute. 14 / 29

  15. Readability Is Simplification Relative? Assessment for Sentences How can we study this? Sowmya Vajjala Introduction Background ◮ One approach: Our Approach Corpora ◮ compute reading levels of normal (N) and simplified (S) Features Modeling sentences using our document level readability model. Relative Comparison Ranking ◮ evaluate simplification classification using the Conclusions percentages of S < N, S = N and S > N ◮ the higher the percentage for S < N, the better the model is, at evaluating sentence level readability. ◮ Why?: Simplified versions are expected to be at a lower reading level than Normal versions!! ◮ How big must | S − N | be to interpret it as a categorical difference in reading level? → We call this the d-value . 15 / 29

  16. Readability What exactly is d-value ? Assessment for Sentences Sowmya Vajjala Introduction Background ◮ It is a measure of how fine-grained the model is in Our Approach Corpora identifying reading-level differences between sentences. Features Modeling ◮ For example, let us say d = 0.3. Relative Comparison Ranking ◮ Now, when N = 3.4, S = 3.2, | S − N | = 0 . 2 , < d Conclusions ⇒ S=N. ◮ If N=3.5, S=3.1, | S − N | = 0 . 4 , > d ⇒ S < N. ◮ What is good for us?: the model should be able to identify as many pairs as possible as S < N. ◮ S = N is probably okay, but S > N is bad. 16 / 29

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend