Recognizing Mentions of Adverse Drug Reaction in Social Media - - PowerPoint PPT Presentation

recognizing mentions of adverse drug reaction in social
SMART_READER_LITE
LIVE PREVIEW

Recognizing Mentions of Adverse Drug Reaction in Social Media - - PowerPoint PPT Presentation

Recognizing Mentions of Adverse Drug Reaction in Social Media Gabriel Stanovsky, Daniel Gruhl, Pablo N. Mendes Bar-Ilan University, IBM Research, Lattice Data Inc. April 2017 In this talk 1. Problem: Identifying adverse drug reactions in social


slide-1
SLIDE 1

Recognizing Mentions of Adverse Drug Reaction in Social Media

Gabriel Stanovsky, Daniel Gruhl, Pablo N. Mendes

Bar-Ilan University, IBM Research, Lattice Data Inc.

April 2017

slide-2
SLIDE 2

In this talk

  • 1. Problem: Identifying adverse drug reactions in social media

◮ “I stopped taking Ambien after three weeks, it gave me a

terrible headache”

slide-3
SLIDE 3

In this talk

  • 1. Problem: Identifying adverse drug reactions in social media

◮ “I stopped taking Ambien after three weeks, it gave me a

terrible headache”

  • 2. Approach

◮ LSTM transducer for BIO tagging ◮ + Signal from knowledge graph embeddings

slide-4
SLIDE 4

In this talk

  • 1. Problem: Identifying adverse drug reactions in social media

◮ “I stopped taking Ambien after three weeks, it gave me a

terrible headache”

  • 2. Approach

◮ LSTM transducer for BIO tagging ◮ + Signal from knowledge graph embeddings

  • 3. Active learning

◮ Simulates a low resource scenario

slide-5
SLIDE 5

Task Definition

Adverse Drug Reaction (ADR)

Unwanted reaction clearly associated with the intake of a drug

◮ We focus on automatic ADR identification on social media

slide-6
SLIDE 6

Motivation - ADR on Social Media

  • 1. Associate unknown side-effects with a given drug
  • 2. Monitor drug reactions over time
  • 3. Respond to patients’ complaints
slide-7
SLIDE 7

CADEC Corpus (Karimi et al., 2015)

ADR annotation in forum posts (Ask-A-Patient)

◮ Train: 5723 sentences ◮ Test: 1874 sentences

slide-8
SLIDE 8

Challenges

slide-9
SLIDE 9

Challenges

◮ Context dependent

“Ambien gave me a terrible headache” “Ambien made my headache go away”

slide-10
SLIDE 10

Challenges

◮ Context dependent

“Ambien gave me a terrible headache” “Ambien made my headache go away”

◮ Colloquial

“hard time getting some Z’s”

slide-11
SLIDE 11

Challenges

◮ Context dependent

“Ambien gave me a terrible headache” “Ambien made my headache go away”

◮ Colloquial

“hard time getting some Z’s”

◮ Non-grammatical

“Short term more loss”

slide-12
SLIDE 12

Challenges

◮ Context dependent

“Ambien gave me a terrible headache” “Ambien made my headache go away”

◮ Colloquial

“hard time getting some Z’s”

◮ Non-grammatical

“Short term more loss”

◮ Coordination

“abdominal gas, cramps and pain”

slide-13
SLIDE 13

Approach: LSTM with knowledge graph embeddings

slide-14
SLIDE 14

Task Formulation

Assign a Beginning, Inside, or Outside label for each word

Example

“[I]O [stopped]O [taking]O [Ambien]O [after]O [three]O [weeks]O – [it]O [gave]O [me]O [a]O [terrible]ADR-B [headache]ADR-I”

slide-15
SLIDE 15

Model

◮ bi-RNN transducer model

◮ Outputs a BIO tag for each word ◮ Takes into account context from both past and future words

slide-16
SLIDE 16

Integrating External Knowledge

◮ DBPedia: Knowledge graph based on Wikipedia

◮ (Ambien, type, Drug) ◮ (Ambien, contains, hydroxypropyl)

slide-17
SLIDE 17

Integrating External Knowledge

◮ DBPedia: Knowledge graph based on Wikipedia

◮ (Ambien, type, Drug) ◮ (Ambien, contains, hydroxypropyl)

◮ Knowledge graph embedding

◮ Dense representation of entities ◮ Desirably:

Related entities in DBPedia ⇐ ⇒ Closer in KB-embedding

slide-18
SLIDE 18

Integrating External Knowledge

◮ DBPedia: Knowledge graph based on Wikipedia

◮ (Ambien, type, Drug) ◮ (Ambien, contains, hydroxypropyl)

◮ Knowledge graph embedding

◮ Dense representation of entities ◮ Desirably:

Related entities in DBPedia ⇐ ⇒ Closer in KB-embedding

◮ We experiment with a simple approach:

◮ Add verbatim concept embeddings to word feats

slide-19
SLIDE 19

Prediction Example

slide-20
SLIDE 20

Evaluation

P R F1 ADR Oracle 55.2 100 71.1

◮ ADR Orcale - Marks gold ADR’s regardless of context

◮ Context matters → Oracle errs on 45% of cases

slide-21
SLIDE 21

Evaluation

Emb. % OOV P R F1 ADR Oracle 55.2 100 71.1 LSTM Random 69.6 74.6 71.9 LSTM Google 12.5 85.3 86.2 85.7 LSTM Blekko 7.0 90.5 90.1 90.3

◮ ADR Orcale - Marks gold ADR’s regardless of context

◮ Context matters → Oracle errs on 45% of cases

◮ External knowledge improves performance:

◮ Blekko > Google > Random Init.

slide-22
SLIDE 22

Evaluation

Emb. % OOV P R F1 ADR Oracle 55.2 100 71.1 LSTM Random 69.6 74.6 71.9 LSTM Google 12.5 85.3 86.2 85.7 LSTM Blekko 7.0 90.5 90.1 90.3 LSTM + DBPedia Blekko 7.0 92.2 94.5 93.4

◮ ADR Orcale - Marks gold ADR’s regardless of context

◮ Context matters → Oracle errs on 45% of cases

◮ External knowledge improves performance:

◮ Blekko > Google > Random Init. ◮ DBPedia provides embeddings for 232 (4%) of the words

slide-23
SLIDE 23

Active Learning: Concept identification for low-resource tasks

slide-24
SLIDE 24

Annotation Flow

Concept Expansion

Bootstrap lexicon

Train & Predict

RNN transducer

Silver Active Learning

Uncertainty sampling

Adjudicate Gold

slide-25
SLIDE 25

Annotation Flow

Concept Expansion

Bootstrap lexicon

Train & Predict

RNN transducer

Silver Active Learning

Uncertainty sampling

Adjudicate Gold

slide-26
SLIDE 26

Annotation Flow

Concept Expansion

Bootstrap lexicon

Train & Predict

RNN transducer

Silver Active Learning

Uncertainty sampling

Adjudicate Gold

slide-27
SLIDE 27

Annotation Flow

Concept Expansion

Bootstrap lexicon

Train & Predict

RNN transducer

Silver Active Learning

Uncertainty sampling

Adjudicate Gold

slide-28
SLIDE 28

Training from Rascal

200 400 600 800 1000 0.2 0.4 0.6 0.8 1

# Annotated Sentences F1

active learning random sampling

◮ Performance after 1hr annotation: 74.2 F1 (88.8 P, 63.8 R) ◮ Uncertainty sampling boosts improvement rate

slide-29
SLIDE 29

Wrap-Up

slide-30
SLIDE 30

Future Work

◮ Use more annotations from CADEC

◮ E.g., symptoms and drugs

◮ Use coreference / entity linking to find DBPedia concepts

slide-31
SLIDE 31

Conclusions

◮ LSTMs can predict ADR on social media ◮ Novel use of knowledge base embeddings with LSTMs ◮ Active learning can help ADR identification in low-resource

domains

slide-32
SLIDE 32

Conclusions

◮ LSTMs can predict ADR on social media ◮ Novel use of knowledge base embeddings with LSTMs ◮ Active learning can help ADR identification in low-resource

domains

Thanks for listening! Questions?