Interpretable Multimodal Deep Learning for Objective Diagnosis, - - PowerPoint PPT Presentation

interpretable multimodal deep learning for objective
SMART_READER_LITE
LIVE PREVIEW

Interpretable Multimodal Deep Learning for Objective Diagnosis, - - PowerPoint PPT Presentation

Interpretable Multimodal Deep Learning for Objective Diagnosis, Prognosis and Biomarker Discovery Faisal Mahmood, PhD Postdoctoral Fellow Department of Biomedical Engineering Johns Hopkins University faisalm@jhu.edu | http://faisal.ai March


slide-1
SLIDE 1

March 20, 2019 1

Interpretable Multimodal Deep Learning for Objective Diagnosis, Prognosis and Biomarker Discovery

Faisal Mahmood, PhD Postdoctoral Fellow Department of Biomedical Engineering Johns Hopkins University faisalm@jhu.edu | http://faisal.ai

faisalm@jhu.edu | faisal.ai

slide-2
SLIDE 2

Deep Learning for Medical Imaging – Major Challenges

  • 1. Limited Annotated Data
  • Under representation of rare conditions.
  • Limited experts available for annotation.
  • Privacy Issues
  • Cost

Faisal Mahmood, Nicholas J. Durr et al."Unsupervised Reverse Domain Adaptation for Synthetic Medical Images via Adversarial Training." IEEE Transactions on Medical Imaging (2018).

mmmmmmmmmm Pathology 65%

70%

Johns Hopkins Electronic Medical Records

Radiology 15%

... Unlabeled Unstructured Transfer Learning Pre- Trained Does not capture clinical diversity.

slide-3
SLIDE 3
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.

Deep Learning for Medical Imaging – Major Challenges Train Test

How can we train AI systems robust to variability in the data?

slide-4
SLIDE 4
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.

Deep Learning for Medical Imaging – Major Challenges Train Test

(F. Mahmood et al., 2018)

slide-5
SLIDE 5
  • 3. Structured Prediction
  • Global vs Local features.

Deep Learning for Medical Imaging – Major Challenges

Modify

AI System

Input Output Ground Truth

Compare

  • F. Mahmood, 2018 – (EN.580.142.13)
slide-6
SLIDE 6
  • 3. Structured Prediction
  • Global vs Local features.

Deep Learning for Medical Imaging – Major Challenges Ai System

Input Output Ground Truth

Compare

Per-pixel classification or regression is unstructured. Each pixel is considered conditionally independent.

How can we develop conditionally dependent deep learning models?

slide-7
SLIDE 7
  • 4. Incorporating Multimodal Information
  • Subjective diagnosis is multimodal.

Deep Learning for Medical Imaging – Major Challenges

Imaging Data

Histopathology Immunohistochemistry Radiology

Endoscopy

Multi-omics Data

Genomics Proteomics

Transcriptomic s miRNAomics

Metabolomics

Patient Specific Data

Patient History

Familial History Clinical Phenotyping Pharmacologic Data

… …

How Can we Fuse Unregistered, Uncorrelated and Noisy Data? How can we separate patient specific and general information?

slide-8
SLIDE 8

Deep Learning for Medical Imaging – Major Challenges

  • 1. Limited Annotated Data
  • Under representation of rare conditions.
  • Limited experts available for annotation.
  • Privacy Issues, Cost
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.
  • 3. Structured Prediction
  • Global vs Local features.
  • 4. Incorporating Multimodal Information
  • Subjective diagnosis is multimodal.
slide-9
SLIDE 9

Computational Pathology Computational Endoscopy

slide-10
SLIDE 10

Computational Pathology Computational Endoscopy

slide-11
SLIDE 11

Application: Depth Estimation for Endoscopy Purpose: Predict Topography from Monocular Images Colonoscopy Gives 2D Images Topography Matters

Faisal Mahmood, Nicholas J. Durr et al." Deep learning and conditional random fields-based depth estimation and topographical reconstruction from conventional endoscopy" Medical Image Analysis (2018).

Endoscopic Depth and Topography

slide-12
SLIDE 12

Endoscopic Depth and Topography

Application: Depth Estimation for Endoscopy Purpose: Predict Topography from Monocular Images Colonoscopy Gives 2D Images Topography Matters 60% of colorectal cancer cases detected after optical colonoscopy are associated with missed lesions.

How do gastroenterologists predict the presence of a polyp? Predict the size of the perforations. Predict surface topography.

Faisal Mahmood, Nicholas J. Durr et al." Deep learning and conditional random fields-based depth estimation and topographical reconstruction from conventional endoscopy" Medical Image Analysis (2018).

slide-13
SLIDE 13

Depth Estimation from Monocular Endoscopy Images

No Ground Truth Depth Data:

  • Limited real estate on an endoscope.
  • Regulatory approvals required to add depth sensor.

8.8mm

Solution: Generate Synthetic Endoscopy Data

Faisal Mahmood, Nicholas J. Durr et al." Deep learning and conditional random fields-based depth estimation and topographical reconstruction from conventional endoscopy" Medical Image Analysis (2018).

slide-14
SLIDE 14

Generating Synthetic Endoscopy Data with GT Depth

(F. Mahmood et al., 2018)

slide-15
SLIDE 15

Generating Synthetic Endoscopy Data with GT Depth

(F. Mahmood et al., 2018)

slide-16
SLIDE 16

Deep Learning for Medical Imaging – Major Challenges

  • 1. Limited Annotated Data
  • Under representation of rare conditions.
  • Limited experts available for annotation.
  • Privacy Issues
  • Cost
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.
  • 3. Structured Prediction
  • Global vs Local features.
  • 4. Incorporating Multimodal Information
  • Subjective diagnosis is multimodal.
slide-17
SLIDE 17

Training with Endoscopy Synthetic Data

Problem: Standard Deep Learning Networks are not sufficiently context aware.

Solution: Add non-local information using a joint CNN-Graphical Model Setup.

(F. Mahmood et al., 2018)

slide-18
SLIDE 18

Solution: Joint CNN-CRF Model

(F. Mahmood et al., 2018)

slide-19
SLIDE 19

Typical Deep Learning Flow

Solution: Joint CNN-CRF Model

Adds Non-local Information

(F. Mahmood et al., 2018)

slide-20
SLIDE 20

Deep Learning for Medical Imaging – Major Challenges

  • 1. Limited Annotated Data
  • Under representation of rare conditions.
  • Limited experts available for annotation.
  • Privacy Issues
  • Cost
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.
  • 3. Structured Prediction
  • Global vs Local features.
  • 4. Incorporating Multimodal Information
  • Subjective diagnosis is multimodal.
slide-21
SLIDE 21

Adapting Synthetic Networks to Real Data

Train Test Problem: Network trained on synthetic data does not work with real data. Solution: Adversarial Reverse Domain Adaptation.

(Mahmood et al., 2018)

slide-22
SLIDE 22

Adversarial Reverse Domain Adaptation

Typical Flow: Transfer Synthetic Data to Real-like Domain. Proposed Flow: Transfer Real Data to Synthetic-like Domain. Remove Patient Specific Features while preserving features necessary for depth estimation.

  • F. Mahmood, Nicholas J. Durr et al."Unsupervised Reverse Domain Adaptation for Synthetic Medical Images via Adversarial Training."

IEEE Transactions on Medical Imaging (2018).

slide-23
SLIDE 23

Adversarial Reverse Domain Adaptation

Faisal Mahmood, Nicholas J. Durr et al."Unsupervised Reverse Domain Adaptation for Synthetic Medical Images via Adversarial Training." IEEE Transactions on Medical Imaging (2018).

Patient Specific Details Removed

Shape, Shading, Intensity Preserved

Endoscopy Images

Synthetic-like

Representation (Mahmood et al., 2018)

slide-24
SLIDE 24

Endoscopy Depth Estimation

Colonoscopy Video Depth Estimate Colonoscopy Video Depth Estimate

(Mahmood et al., 2018)

slide-25
SLIDE 25

Validation – Endoscopy Depth Estimation

(Mahmood et al., 2018)

slide-26
SLIDE 26

Estimated Depth to Topography

Faisal Mahmood, Nicholas J. Durr et al."Unsupervised Reverse Domain Adaptation for Synthetic Medical Images via Adversarial Training." IEEE Transactions on Medical Imaging (2018).

slide-27
SLIDE 27

Polyp Charcterization

Hyperplastic Serrated Adenoma

Can we predict the type of polyp without a biopsy only from RGB Image using limited data?

slide-28
SLIDE 28

Polyp Charcterization

Hyperplastic Serrated Adenoma

  • 76 Videos
  • All videos labeled by 4 Senior Gastroenterologists & 3 Fellows
  • Average GI Accuracy: Senior: 63.4% Fellow: 53.7%
slide-29
SLIDE 29

Deep Learning for Medical Imaging – Major Challenges

  • 1. Limited Annotated Data
  • Under representation of rare conditions.
  • Limited experts available for annotation.
  • Privacy Issues
  • Cost
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.
  • 3. Structured Prediction
  • Global vs Local features.
  • 4. Incorporating Multimodal Information
  • Subjective diagnosis is multimodal.
slide-30
SLIDE 30

Multimodal Data Fusion

Faisal Mahmood, Nicholas J. Durr et al."Unsupervised Reverse Domain Adaptation for Synthetic Medical Images via Adversarial Training." IEEE Transactions on Medical Imaging (2018).

Fusion Network

Depth is an additional predicted modality. Can we fuse depth and RGB to get better polyp classification results? Does depth fusion help train a network that requires less data?

slide-31
SLIDE 31

RGB-D Classification via Depth Fusion

(Mahmood et al., 2018)

Data Fusion in Feature Space is better than Concatenation.

slide-32
SLIDE 32

Error

Multimodal Densenet

(F. Mahmood et al., 2018)

slide-33
SLIDE 33

RGB-D Classification RGB vs RGB-D

Error

Average Accuracy = 93% using RGB-D

slide-34
SLIDE 34

20 40 60 80 100 120 140 160 180 200

epochs

0.1 0.2 0.3 0.4 0.5 0.6 0.7

test error DenseNet DenseNet Fit MultiDense MultiDense Fit

RGB vs RGB-D Classification

(F. Mahmood et al., 2018)

RGB+Depth Classificatio n RGB Classificatio n

slide-35
SLIDE 35

Gradient Class Activation Maps

RGB Classification RGB-D Classification Adenoma

(F. Mahmood et al., 2018)

slide-36
SLIDE 36

RGB Classification RGB-D Classification Input

Gradient Class Activation Maps

(F. Mahmood et al., 2018)

Using just 76 polyp videos with fused depth it is possible to build a classifier with AUC > 0.9

slide-37
SLIDE 37

Computational Pathology Computational Endoscopy

slide-38
SLIDE 38

Automated Breast Cancer Grading

Nuclear Atypia Tubule/Gland Formation

1 2 3 1 2 3 1 2 3

Mitotic Activity

slide-39
SLIDE 39

Typical AI for Pathology Flow

Whole Slide Image

≈ 1 Billion Pixels!

Region of Interest Ground Truth from Different Classes

This Needs a Lot of Labeled Data!

Ductal Carcinoma

Invasive Ductal Carcinoma Other Diagnostic Classes

Convolutional Neural Network

  • F. Mahmood, 2018 – (EN.580.142.13)

Grade Ground Truth Grade Interobserver & Intraobserver Variability?

slide-40
SLIDE 40

Automated Breast Cancer Grading

  • A. Nuclear Atypia
  • B. Tubule/Gland Formation

1 2 3 1 2 3 1 2 3

  • C. Mitotic Activity

Nuclei Segmentation Mitosis Detection Tubule Detection

A+B+C+

Additional Features

Grade

Can help discover additional morphological features.

slide-41
SLIDE 41

Automated Breast Cancer Grading

  • A. Nuclear Atypia

1 2 3

Nuclei Segmentation Can we build a single AI model that can segment nuclei from any H&E image regardless or organ?

slide-42
SLIDE 42

Labeled Nuclei Segmentation Data

32 1000x1000 Slide Patches from 8 Different Organs

Small subjectively labeled datasets are not enough for capturing the diversity needed for a singular multi-organ nuclei segmentation network.

slide-43
SLIDE 43

Sparse Stain Normalization

Breast Kidney Lung Stomach

slide-44
SLIDE 44

Synthetic Data Generation for Nuclei Segmentation

Normalized sides from six organs Segmentation Masks Pathology Patches (B) Nuclei Segmentation Patches (A)

The variability in data can be captured using synthetic data. Unpaired Synthetic Data Generation

slide-45
SLIDE 45

Unpaired Mapping between random polygons and synthetic H&E patches.

slide-46
SLIDE 46

Unpaired Mapping between random polygons and synthetic H&E patches.

Synthetic Pathology Images with Nuclei Segmentation Masks

slide-47
SLIDE 47

Deep Learning for Medical Imaging – Major Challenges

  • 1. Limited Annotated Data
  • Under representation of rare conditions.
  • Limited experts available for annotation.
  • Privacy Issues
  • Cost
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.
  • 3. Structured Prediction
  • Global vs Local features.
  • 4. Incorporating Multimodal Information
  • Subjective diagnosis is multimodal.

Faisal Mahmood, Nicholas J. Durr et al."Unsupervised Reverse Domain Adaptation for Synthetic Medical Images via Adversarial Training." IEEE Transactions on Medical Imaging (2018).

slide-48
SLIDE 48

Faisal Mahmood, Daniel Borders, et al. "Deep Adversarial Training for Multi-Organ Nuclei Segmentation in Histopathology Images." arXiv preprint arXiv:1810.00236 (2018).

Context-Aware Nuclei Segmentation with No CRF Post-processing Step

Overlapping Nuclei

slide-49
SLIDE 49

Aggregated Jaccard Index

(F. Mahmood et al., 2018)

slide-50
SLIDE 50

Extension to Mitotic Event Detection

(F. Mahmood et al., 2018)

Synthetic Data Generation Context-Aware Mitosis Detection

94.6% Detection Accuracy

slide-51
SLIDE 51

Extension to Epithelium Segmentation

(F. Mahmood et al., 2018)

93.8% Segmentation Accuracy

slide-52
SLIDE 52

Extension to Tubule Segmentation

96.9% Segmentation Accuracy

(F. Mahmood et al., 2018)

slide-53
SLIDE 53

Tissue Level Semantic Segmentation

TMA Prediction Ground Truth

91.4% Segmentation Accuracy

slide-54
SLIDE 54

Predicted Feature Fusion

Grade Prediction

0.2 0.4 0.6 0.8 1

False Positive Rate

0.2 0.4 0.6 0.8 1

True Positive Rate H&E Only - ROC

H&E Only (AUC =0.8425)

Data: 2,210 H&E Slide ROIs (TCGA). Labeling: Labeled by 2 pathologists.

AUC=0.8425

slide-55
SLIDE 55

Predicted Feature Fusion

Grade Prediction

AUC Increases by 7.98% - Same Data.

0.2 0.4 0.6 0.8 1

False Positive Rate

0.2 0.4 0.6 0.8 1

True Positive Rate H&E + NS ROC

H&E + NS (AUC =0.9097)

AUC=0.9097

slide-56
SLIDE 56

Predicted Feature Fusion

Grade Prediction

AUC Increases by 13.74% - Same Data.

0.2 0.4 0.6 0.8 1

False Positive Rate

0.2 0.4 0.6 0.8 1

True Positive Rate H&E + Nuclei Seg +TF + ME ROC

H&E+NS+TF+ME (AUC =0.9583)

AUC=0.9583

slide-57
SLIDE 57

Error

Multimodal Densenet: General Framework for Multimodal Data Fusion

(F. Mahmood et al., 2018)

slide-58
SLIDE 58

Multimodal Densenet: General Framework for Multimodal Data Fusion

(F. Mahmood et al., 2018)

Multimodal Fusion Library

slide-59
SLIDE 59

Deep Learning for Medical Imaging – Major Challenges

  • 1. Limited Annotated Data
  • Under representation of rare conditions.
  • Limited experts available for annotation.
  • Privacy Issues
  • Cost
  • 2. Domain Adaptation
  • Diversity in data, different sensors, cites and patients.
  • Patient specific texture and color information.
  • 3. Structured Prediction
  • Global vs Local features.
  • 4. Incorporating Multimodal Information
  • Subjective diagnosis is multimodal.
slide-60
SLIDE 60

Future Work - Clinical

Aim 1: Segmentation, Classification and Grading

Nuclei Segmentation

Synthetic Data Generation Context-aware Adversarial Methods

Tissue Level Segmentation Whole Slide Classification Methods

Fusing spatial information Classes Malignant Benign DCIS UDH Dense Classifier

Aim 2: Biomarker Identification

Identify Features Used to Make Classification Decisions Correlate Identified Features with Patient Diagnosis and Prognosis

Feature-wise Kaplan– Meier Prognostic Model

Features used for classification Aim 4: Multimodal Fusion Aim 3: New Grading System / Response to Specific Therapeutic Agents

Fusing Morphological Features for Classification

Develop New Grading Scheme

Subjective + Cloud Based Analysis of Nottingham vs Proposed Grading Objective Analysis of Nottingham vs Proposed Grading DL Models on Cloud HPC Comparative Analysis

slide-61
SLIDE 61

Computational Biophotonics Lab Nicholas J. Durr Richard Chen Greg N. McKay Jordan Sweer Taylor Bobrow Mason Chen Bailey Surtees Eric Chiang JHU School of Medicine Alexander S. Baras Susan Harvey Saowanee Ngamruengphong Department of Computer Science Alan Yuille Greg Haiger Siemens Sandra Sudarsky Daphne Yu

Acknowledgements

Olympus Lee Zhen Google Jesus Trujillo Gomez

slide-62
SLIDE 62

faisalm@jhu.edu Code / Data Available at: http://faisal.ai

Thank You.