natural language processing for analyzing disaster
play

Natural Language Processing for Analyzing Disaster Recovery Trends - PowerPoint PPT Presentation

Natural Language Processing for Analyzing Disaster Recovery Trends Expressed in Large Text Corpora Lucy H. Lin , Scott B. Miles, Noah A. Smith University of Washington 19 October 2018 Introduction Problem : empirical data describing disaster


  1. Natural Language Processing for Analyzing Disaster Recovery Trends Expressed in Large Text Corpora Lucy H. Lin , Scott B. Miles, Noah A. Smith University of Washington 19 October 2018

  2. Introduction Problem : empirical data describing disaster recovery is scarce Available data source: text corpora (e.g., news articles) Manual analysis of large text corpora is slow… ⇒ constrains pre-/post-event recovery planning ⇒ + natural language processing

  3. Introduction anger among homeowners.” freq Proposition query: Frequency across time: aggregate “Those in charge of recovery are making moves to appease the growing “Unfamiliar bureaucratic systems are causing the majority of the stress.” Matched sentences: “Dealing with authorities is causing stress and anxiety.” query corpus 2 3 4 5 1 1 1 1 1 1 0 0 0 0 0 2 2 2 2 2

  4. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  5. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  6. 2010–2011 Canterbury earthquake disaster: timeline September 2010 : • Epicenter: 35km west of Christchurch • Moderate damage February 2011 : • Epicenter: 10km southeast of Christchurch • Extremely high ground acceleration • 185 deaths, thousands of felt aftershocks

  7. 2010–2011 Canterbury earthquake disaster: impacts Damages : • Estimated $40 billion • Housing: 100k houses in need of repairs • Water, utilities, road infrastructure: extensive damage Recovery groups : • Government: CERA, SCIRT (sunset after 5 years) • Community: Regenerate Christchurch Recovery still ongoing: public development projects, residential rezoning

  8. 2010–2011 Canterbury earthquake disaster: text data Corpus : 982 NZ news articles (2010–2015) post-earthquakes • stuff.co.nz , nzherald.co.nz • Community wellbeing • Infrastructure • Decision-making e.g.: “The council should have consulted residents before making decisions.” Proposition queries : 20 queries, covering

  9. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  10. Semantic matching Goal : find sentences with similar meaning to the query. • Needs to be more powerful than word/phrase-level matching. • Related to information retrieval, but want all matches .

  11. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  12. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  13. Semantic matching: fast filter . . stress average Goal : quickly filter out unlikely matches. Dealing with . . . . . . anxiety cosine . similarity . bureaucratic Word vector based comparison between two sentences: average . Unfamiliar .   ⊏ ⊐  ⊏ ⊐  corpus sent. ⊏ ⊐         ⊏ ⊐   ⊏ ⊐  ⊏ ⊐    query sent. ⊏ ⊐       ⊏ ⊐

  14. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  15. Semantic matching: syntax-based model Finer-grained matching : take word order/syntax into account. Intuition : transformation between sentences is indicative of their relationship.

  16. Semantic matching: syntax-based model causing +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) ? query root stress is unfamiliar authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  17. Semantic matching: syntax-based model root +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing are systems ? query stress unfamiliar causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  18. Semantic matching: syntax-based model root +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing are dealing ? query stress unfamiliar causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  19. Semantic matching: syntax-based model root +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing is dealing ? query stress unfamiliar causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  20. Semantic matching: syntax-based model query +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing is authorities with dealing root unfamiliar stress causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  21. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  22. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  23. User evaluation Questions : • How good are the sentences matched by our method? • Do potential users think this kind of tool will be helpful? User study : 20 emergency managers

  24. User evaluation: output quality Rated output from 20 proposition queries: • Different method variants • Different parts of method: • Not selected by filter • Selected by filter, but not part of final output • Top-scoring output from filter • Method output (from syntax-based model) • 1-5 scale (Krippendorf’s α = 0 . 784)

  25. User evaluation: example Query: There is a shortage of construction workers. “The quarterly report for Canterbury included analysis on ( 1 : completely unrelated to the query) Greater Christchurch Value of Work projections.”

  26. User evaluation: example Query: There is a shortage of construction workers. December.” ( 3 : related to but does not adequately express the query) “The construction sector’s workload was expected to peak in

  27. User evaluation: example Query: There is a shortage of construction workers. and was likely to remain that way.” ( 5 : expresses the query in its entirety) “Greater Christchurch’s labour supply for the rebuild was tight

  28. User evaluation: results 1 2 3 Average score Best performing system Not selected by filter Selected by filter (unmatched) 3 . 22 3 . 1 2 . 5 2 . 03 1 . 5 1 . 06

  29. User evaluation: results 1 2 3 Average score Best performing system Not selected by filter Selected by filter (unmatched) Highest-scoring by filter 3 . 22 3 . 1 2 . 5 2 . 03 1 . 5 1 . 06

  30. User evaluation: results 1 2 3 Average score Best performing system Not selected by filter Selected by filter (unmatched) Highest-scoring by filter Matched by method 3 . 22 3 . 1 2 . 5 2 . 03 1 . 5 1 . 06

  31. User evaluation: is this interesting? Other feedback : news/other text corpora • 17/20 respondents interested in measuring ideas in

  32. User evaluation: round two Follow-up study : • 7 return participants • Replicated findings of first user study • Participant-supplied queries (18)

  33. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  34. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly.

  35. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly. “Orion Energy CEO Roger Sutton says most of the west of Christchurch now has fully restored power.”

  36. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly. “He had no water but power had been restored in his area.”

  37. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly. “TV3 reports that power has now been restored to 60 per cent of Christchurch.”

  38. Recovery trends: example #1 6 restore power and the situation could remain for the next few “It had been unable to access the electricity network to The power system was fully restored quickly. Frequency 12 9 3 2010 0 2015 2014 2013 2012 2011 days.”

  39. Recovery trends: example #2 2010 2011 2012 2013 2014 2015 0 2 4 6 Frequency Dealing w/authorities is causing stress and anxiety.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend