Systems & Applications: Introduction Ling 573 NLP Systems and - - PowerPoint PPT Presentation

systems applications introduction
SMART_READER_LITE
LIVE PREVIEW

Systems & Applications: Introduction Ling 573 NLP Systems and - - PowerPoint PPT Presentation

Systems & Applications: Introduction Ling 573 NLP Systems and Applications April 1, 2014 Roadmap Motivation 573 Structure Question-Answering Shared Tasks Motivation Information retrieval is very powerful


slide-1
SLIDE 1

Systems & Applications: Introduction

Ling 573 NLP Systems and Applications April 1, 2014

slide-2
SLIDE 2

Roadmap

— Motivation — 573 Structure — Question-Answering — Shared Tasks

slide-3
SLIDE 3

Motivation

— Information retrieval is very powerful

— Search engines index and search enormous doc sets — Retrieve billions of documents in tenths of seconds

slide-4
SLIDE 4

Motivation

— Information retrieval is very powerful

— Search engines index and search enormous doc sets — Retrieve billions of documents in tenths of seconds

— But still limited!

slide-5
SLIDE 5

Motivation

— Information retrieval is very powerful

— Search engines index and search enormous doc sets — Retrieve billions of documents in tenths of seconds

— But still limited!

— Technically – keyword search (mostly)

slide-6
SLIDE 6

Motivation

— Information retrieval is very powerful

— Search engines index and search enormous doc sets — Retrieve billions of documents in tenths of seconds

— But still limited!

— Technically – keyword search (mostly) — Conceptually

— User seeks information

— Sometimes a web site or document

slide-7
SLIDE 7

Motivation

— Information retrieval is very powerful

— Search engines index and search enormous doc sets — Retrieve billions of documents in tenths of seconds

— But still limited!

— Technically – keyword search (mostly) — Conceptually

— User seeks information

— Sometimes a web site or document — Very often, the answer to a question

slide-8
SLIDE 8

Why Question-Answering?

— People ask questions on the web

slide-9
SLIDE 9

Why Question-Answering?

— People ask questions on the web

— Web logs:

— Which English translation of the bible is used in official

Catholic liturgies?

— Who invented surf music? — What are the seven wonders of the world?

slide-10
SLIDE 10

Why Question-Answering?

— People ask questions on the web

— Web logs:

— Which English translation of the bible is used in official

Catholic liturgies?

— Who invented surf music? — What are the seven wonders of the world?

— 12-15% of queries

slide-11
SLIDE 11

Why Question-Answering?

— People ask questions on the web

— Web logs:

— Which English translation of the bible is used in official

Catholic liturgies?

— Who invented surf music? — What are the seven wonders of the world?

— 12-15% of queries

— Search sites (e.g., Google) beginning to include

— Canonical factoids, esp. Wikipedia infobox data

— Dates, conversions, birthdates

slide-12
SLIDE 12

Why Question Answering?

— Answer sites proliferate:

slide-13
SLIDE 13

Why Question Answering?

— Answer sites proliferate:

— Top hit for ‘questions’ :

slide-14
SLIDE 14

Why Question Answering?

— Answer sites proliferate:

— Top hit for ‘questions’ : Ask.com

slide-15
SLIDE 15

Why Question Answering?

— Answer sites proliferate:

— Top hit for ‘questions’ : Ask.com

— Also: Yahoo! Answers, wiki answers, Facebook,…

— Collect and distribute human answers

slide-16
SLIDE 16

Why Question Answering?

— Answer sites proliferate:

— Top hit for ‘questions’ : Ask.com

— Also: Yahoo! Answers, wiki answers, Facebook,…

— Collect and distribute human answers — Do I Need a Visa to Go to Japan?

slide-17
SLIDE 17

Why Question Answering?

— Answer sites proliferate:

— Top hit for ‘questions’ : Ask.com

— Also: Yahoo! Answers, wiki answers, Facebook,…

— Collect and distribute human answers — Do I Need a Visa to Go to Japan?

— eHow.com — Rules regarding travel between the United States and Japan

are governed by both countries. Entry requirements for Japan are contingent on the purpose and length of a traveler's visit.

— Passport Requirements

— Japan requires all U.S. citizens provide a valid passport and a

return on "onward" ticket for entry into the country. Additionally, the United States requires a passport for all citizens wishing to enter or re-enter the country.

slide-18
SLIDE 18

Search Engines & QA

— Who was the prime minister of Australia during the

Great Depression?

slide-19
SLIDE 19

Search Engines & QA

— Who was the prime minister of Australia during the

Great Depression? — Rank 1 snippet:

— The conservative Prime Minister of Australia, Stanley Bruce

slide-20
SLIDE 20

Search Engines & QA

— Who was the prime minister of Australia during the

Great Depression? — Rank 1 snippet:

— The conservative Prime Minister of Australia, Stanley Bruce

— Wrong!

— Voted out just before the Depression

slide-21
SLIDE 21

Perspectives on QA

— TREC QA track (1999---)

— Initially pure factoid questions, with fixed length answers

— Based on large collection of fixed documents (news) — Increasing complexity: definitions, biographical info, etc

— Single response

slide-22
SLIDE 22

Perspectives on QA

— TREC QA track (~1999---)

— Initially pure factoid questions, with fixed length answers

— Based on large collection of fixed documents (news) — Increasing complexity: definitions, biographical info, etc

— Single response

— Reading comprehension (Hirschman et al, 2000---)

— Think SAT/GRE

— Short text or article (usually middle school level) — Answer questions based on text

— Also, ‘machine reading’

slide-23
SLIDE 23

Perspectives on QA

— TREC QA track (~1999---)

— Initially pure factoid questions, with fixed length answers

— Based on large collection of fixed documents (news) — Increasing complexity: definitions, biographical info, etc

— Single response

— Reading comprehension (Hirschman et al, 2000---)

— Think SAT/GRE

— Short text or article (usually middle school level) — Answer questions based on text

— Also, ‘machine reading’

— And, of course, Jeopardy! and Watson

slide-24
SLIDE 24

Natural Language Processing and QA

— Rich testbed for NLP techniques:

slide-25
SLIDE 25

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval

slide-26
SLIDE 26

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition

slide-27
SLIDE 27

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition — Tagging

slide-28
SLIDE 28

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition — Tagging — Information extraction

slide-29
SLIDE 29

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition — Tagging — Information extraction — Word sense disambiguation

slide-30
SLIDE 30

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition — Tagging — Information extraction — Word sense disambiguation — Parsing

slide-31
SLIDE 31

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition — Tagging — Information extraction — Word sense disambiguation — Parsing — Semantics, etc..

slide-32
SLIDE 32

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition — Tagging — Information extraction — Word sense disambiguation — Parsing — Semantics, etc.. — Co-reference

slide-33
SLIDE 33

Natural Language Processing and QA

— Rich testbed for NLP techniques:

— Information retrieval — Named Entity Recognition — Tagging — Information extraction — Word sense disambiguation — Parsing — Semantics, etc.. — Co-reference

— Deep/shallow techniques; machine learning

slide-34
SLIDE 34

573 Structure

— Implementation:

slide-35
SLIDE 35

573 Structure

— Implementation:

— Create a factoid QA system

slide-36
SLIDE 36

573 Structure

— Implementation:

— Create a factoid QA system

— Extend existing software components — Develop, evaluate on standard data set

slide-37
SLIDE 37

573 Structure

— Implementation:

— Create a factoid QA system

— Extend existing software components — Develop, evaluate on standard data set

— Presentation:

slide-38
SLIDE 38

573 Structure

— Implementation:

— Create a factoid QA system

— Extend existing software components — Develop, evaluate on standard data set

— Presentation:

— Write a technical report — Present plan, system, results in class

slide-39
SLIDE 39

573 Structure

— Implementation:

— Create a factoid QA system

— Extend existing software components — Develop, evaluate on standard data set

— Presentation:

— Write a technical report — Present plan, system, results in class — Give/receive feedback

slide-40
SLIDE 40

Implementation: Deliverables

— Complex system:

— Break into (relatively) manageable components — Incremental progress, deadlines

slide-41
SLIDE 41

Implementation: Deliverables

— Complex system:

— Break into (relatively) manageable components — Incremental progress, deadlines

— Key components:

— D1: Setup

slide-42
SLIDE 42

Implementation: Deliverables

— Complex system:

— Break into (relatively) manageable components — Incremental progress, deadlines

— Key components:

— D1: Setup — D2: Baseline system, Passage retrieval

slide-43
SLIDE 43

Implementation: Deliverables

— Complex system:

— Break into (relatively) manageable components — Incremental progress, deadlines

— Key components:

— D1: Setup — D2: Baseline system, Passage retrieval — D3: Question processing, classification

slide-44
SLIDE 44

Implementation: Deliverables

— Complex system:

— Break into (relatively) manageable components — Incremental progress, deadlines

— Key components:

— D1: Setup — D2: Baseline system, Passage retrieval — D3: Question processing, classification — D4: Answer processing, final results

slide-45
SLIDE 45

Implementation: Deliverables

— Complex system:

— Break into (relatively) manageable components — Incremental progress, deadlines

— Key components:

— D1: Setup — D2: Baseline system, Passage retrieval — D3: Question processing, classification — D4: Answer processing, final results

— Deadlines:

— Little slack in schedule; please keep to time — Timing: ~12 hours week; sometimes higher

slide-46
SLIDE 46

Presentation

— Technical report:

— Follow organization for scientific paper

— Formatting and Content

slide-47
SLIDE 47

Presentation

— Technical report:

— Follow organization for scientific paper

— Formatting and Content

— Presentations:

— 10-15 minute oral presentation for deliverables

slide-48
SLIDE 48

Presentation

— Technical report:

— Follow organization for scientific paper

— Formatting and Content

— Presentations:

— 10-15 minute oral presentation for deliverables — Explain goals, methodology, success, issues

slide-49
SLIDE 49

Presentation

— Technical report:

— Follow organization for scientific paper

— Formatting and Content

— Presentations:

— 10-15 minute oral presentation for deliverables — Explain goals, methodology, success, issues — Critique each others’ work

slide-50
SLIDE 50

Presentation

— Technical report:

— Follow organization for scientific paper

— Formatting and Content

— Presentations:

— 10-15 minute oral presentation for deliverables — Explain goals, methodology, success, issues — Critique each others’ work — Attend ALL presentations

slide-51
SLIDE 51

Working in Teams

— Why teams?

slide-52
SLIDE 52

Working in Teams

— Why teams?

— Too much work for a single person — Representative of professional environment

slide-53
SLIDE 53

Working in Teams

— Why teams?

— Too much work for a single person — Representative of professional environment

— Team organization:

— Form groups of 3 (possibly 2) people

slide-54
SLIDE 54

Working in Teams

— Why teams?

— Too much work for a single person — Representative of professional environment

— Team organization:

— Form groups of 3 (possibly 2) people — Arrange coordination — Distribute work equitably

slide-55
SLIDE 55

Working in Teams

— Why teams?

— Too much work for a single person — Representative of professional environment

— Team organization:

— Form groups of 3 (possibly 2) people — Arrange coordination — Distribute work equitably

— All team members receive the same grade

— End-of-course evaluation

slide-56
SLIDE 56

First Task

— Form teams:

— Email Glenn gslaydeni@uw.edu with the team list

slide-57
SLIDE 57

Resources

— Readings:

— Current research papers in question-answering

slide-58
SLIDE 58

Resources

— Readings:

— Current research papers in question-answering — Jurafsky & Martin/Manning & Schutze text

— Background, reference, refresher

slide-59
SLIDE 59

Resources

— Readings:

— Current research papers in question-answering — Jurafsky & Martin/Manning & Schutze text

— Background, reference, refresher

— Software:

slide-60
SLIDE 60

Resources

— Readings:

— Current research papers in question-answering — Jurafsky & Martin/Manning & Schutze text

— Background, reference, refresher

— Software:

— Build on existing system components, toolkits

— NLP

, machine learning, etc

— Corpora, etc

slide-61
SLIDE 61

Resources: Patas

— System should run on patas

— Existing infrastructure

— Software systems — Corpora — Repositories

slide-62
SLIDE 62

Shared Task Evaluations

— Goals:

— Lofty:

slide-63
SLIDE 63

Shared Task Evaluations

— Goals:

— Lofty:

— Focus research community on key challenges

— ‘Grand challenges’

slide-64
SLIDE 64

Shared Task Evaluations

— Goals:

— Lofty:

— Focus research community on key challenges

— ‘Grand challenges’

— Support the creation of large-scale community resources

— Corpora: News, Recordings, Video — Annotation: Expert questions, labeled answers,..

slide-65
SLIDE 65

Shared Task Evaluations

— Goals:

— Lofty:

— Focus research community on key challenges

— ‘Grand challenges’

— Support the creation of large-scale community resources

— Corpora: News, Recordings, Video — Annotation: Expert questions, labeled answers,..

— Develop methodologies to evaluate state-of-the-art

— Retrieval, Machine Translation, etc

slide-66
SLIDE 66

Shared Task Evaluations

— Goals:

— Lofty:

— Focus research community on key challenges

— ‘Grand challenges’

— Support the creation of large-scale community resources

— Corpora: News, Recordings, Video — Annotation: Expert questions, labeled answers,..

— Develop methodologies to evaluate state-of-the-art

— Retrieval, Machine Translation, etc

— Facilitate technology/knowledge transfer b/t industry/acad.

slide-67
SLIDE 67

Shared Task Evaluation

— Goals:

— Pragmatic:

slide-68
SLIDE 68

Shared Task Evaluation

— Goals:

— Pragmatic:

— Head-to-head comparison of systems/techniques

— Same data, same task, same conditions, same timing

slide-69
SLIDE 69

Shared Task Evaluation

— Goals:

— Pragmatic:

— Head-to-head comparison of systems/techniques

— Same data, same task, same conditions, same timing

— Centralizes funding, effort

slide-70
SLIDE 70

Shared Task Evaluation

— Goals:

— Pragmatic:

— Head-to-head comparison of systems/techniques

— Same data, same task, same conditions, same timing

— Centralizes funding, effort — Requires disclosure of techniques in exchange for data

— Base:

slide-71
SLIDE 71

Shared Task Evaluation

— Goals:

— Pragmatic:

— Head-to-head comparison of systems/techniques

— Same data, same task, same conditions, same timing

— Centralizes funding, effort — Requires disclosure of techniques in exchange for data

— Base:

— Bragging rights

slide-72
SLIDE 72

Shared Task Evaluation

— Goals:

— Pragmatic:

— Head-to-head comparison of systems/techniques

— Same data, same task, same conditions, same timing

— Centralizes funding, effort — Requires disclosure of techniques in exchange for data

— Base:

— Bragging rights — Government research funding decisions

slide-73
SLIDE 73

Shared Tasks: Perspective

— Late ‘80s-90s:

slide-74
SLIDE 74

Shared Tasks: Perspective

— Late ‘80s-90s:

— ATIS: spoken dialog systems — MUC: Message Understanding: information extraction

slide-75
SLIDE 75

Shared Tasks: Perspective

— Late ‘80s-90s:

— ATIS: spoken dialog systems — MUC: Message Understanding: information extraction

— TREC (Text Retrieval Conference)

— Arguably largest ( often >100 participating teams) — Longest running (1992-current) — Information retrieval (and related technologies)

— Actually hasn’t had ‘ad-hoc’ since ~2000, though

— Organized by NIST

slide-76
SLIDE 76

TREC Tracks

— Track: Basic task organization

slide-77
SLIDE 77

TREC Tracks

— Track: Basic task organization — Previous tracks:

— Ad-hoc – Basic retrieval from fixed document set

slide-78
SLIDE 78

TREC Tracks

— Track: Basic task organization — Previous tracks:

— Ad-hoc – Basic retrieval from fixed document set — Cross-language – Query in one language, docs in other

— English, French, Spanish, Italian, German, Chinese, Arabic

slide-79
SLIDE 79

TREC Tracks

— Track: Basic task organization — Previous tracks:

— Ad-hoc – Basic retrieval from fixed document set — Cross-language – Query in one language, docs in other

— English, French, Spanish, Italian, German, Chinese, Arabic

— Genomics

slide-80
SLIDE 80

TREC Tracks

— Track: Basic task organization — Previous tracks:

— Ad-hoc – Basic retrieval from fixed document set — Cross-language – Query in one language, docs in other

— English, French, Spanish, Italian, German, Chinese, Arabic

— Genomics — Spoken Document Retrieval

slide-81
SLIDE 81

TREC Tracks

— Track: Basic task organization — Previous tracks:

— Ad-hoc – Basic retrieval from fixed document set — Cross-language – Query in one language, docs in other

— English, French, Spanish, Italian, German, Chinese, Arabic

— Genomics — Spoken Document Retrieval — Video search

slide-82
SLIDE 82

TREC Tracks

— Track: Basic task organization — Previous tracks:

— Ad-hoc – Basic retrieval from fixed document set — Cross-language – Query in one language, docs in other

— English, French, Spanish, Italian, German, Chinese, Arabic

— Genomics — Spoken Document Retrieval — Video search — Question Answering

slide-83
SLIDE 83

Current TREC tracks

— TREC 2014:

— Contextual Suggestion — Clinical Decision Support Track — Federated Web Search — Knowledge-base Acceleration — Microblog — Session — Temporal Summarization — Web

slide-84
SLIDE 84

Other Shared Tasks

— International:

— CLEF (Europe); NTCIR (Japan); FIRE (India)

slide-85
SLIDE 85

Other Shared Tasks

— International:

— CLEF (Europe); NTCIR (Japan); FIRE (India)

— Other NIST:

— DUC (Document Summarization) — Machine Translation — Topic Detection & Tracking

slide-86
SLIDE 86

Other Shared Tasks

— International:

— CLEF (Europe); NTCIR (Japan); FIRE (India)

— Other NIST:

— DUC (Document Summarization) — Machine Translation — Topic Detection & Tracking

— Various:

— CoNLL (NE, parsing,..); SENSEVAL: WSD; PASCAL

(morphology); BioNLP (biological entities, relations)

slide-87
SLIDE 87

Other Shared Tasks

— International:

— CLEF (Europe); NTCIR (Japan); FIRE (India)

— Other NIST:

— DUC (Document Summarization) — Machine Translation — Topic Detection & Tracking

— Various:

— CoNLL (NE, parsing,..); SENSEVAL: WSD; PASCAL

(morphology); BioNLP (biological entities, relations)

— Mediaeval (multi-media information access)

slide-88
SLIDE 88

TREC Question-Answering

— Several years (1999-2007)

— Started with pure factoid questions from news

sources

slide-89
SLIDE 89

TREC Question-Answering

— Several years (1999-2007)

— Started with pure factoid questions from news

sources

— Extended to lists, relationship

slide-90
SLIDE 90

TREC Question-Answering

— Several years (1999-2007)

— Started with pure factoid questions from news

sources

— Extended to lists, relationship — Extended to blog data — Employed question series

slide-91
SLIDE 91

TREC Question-Answering

— Several years (1999-2007)

— Started with pure factoid questions from news

sources

— Extended to lists, relationship — Extended to blog data — Employed question series — Added temporal constraints — ‘Complex, interactive’ evaluation

slide-92
SLIDE 92

TREC Question-Answering

— Several years (1999-2007)

— Started with pure factoid questions from news sources — Extended to lists, relationship — Extended to blog data — Employed question series — Added temporal constraints — ‘Complex, interactive’ evaluation

— Combined with summarization to form TAC (2008---)

— Text Analytics Conference

— Opinion Q/A, Knowledge-based population, Scientific

Summarization

slide-93
SLIDE 93

TREC Question-Answering

— Provides:

— Lists of questions — Document collections (licensed via LDC) — Ranked document results — Evaluation tools: Answer verification patterns — Derived resources:

— E.g. Roth and Li’s question categories, training/test

— Reams of related publications

slide-94
SLIDE 94

Questions

— <top>

— <num> Number: 894 — <desc> Description: How far is it from Denver to

Aspen?

— </top>

slide-95
SLIDE 95

Questions

— <top>

— <num> Number: 894 — <desc> Description: How far is it from Denver to

Aspen?

— </top> — <top>

— <num> Number: 895 — <desc> Description: What county is Modesto,

California in?

— </top>

slide-96
SLIDE 96

Documents

—

<DOC><DOCNO> APW20000817.0002 </DOCNO>

—

<DOCTYPE> NEWS STORY </DOCTYPE><DATE_TIME> 2000-08-17 00:05 </ DATE_TIME>

—

<BODY> <HEADLINE> 19 charged with drug trafficking </HEADLINE>

—

<TEXT><P>

—

UTICA, N.Y . (AP) - Nineteen people involved in a drug trafficking ring in the Utica area were arrested early Wednesday, police said.

—

</P><P>

—

Those arrested are linked to 22 others picked up in May and comprise ''a major cocaine, crack cocaine and marijuana distribution organization,'' according to the U.S. Department of Justice.

—

</P>

slide-97
SLIDE 97

Answer Keys

— 1394: French — 1395: Nicole Kidman — 1396: Vesuvius — 1397: 62,046 — 1398: 1867 — 1399: Brigadoon

slide-98
SLIDE 98

Reminder

— Team up!