Label-free classification of ciliated cells using Deep Learning - - PowerPoint PPT Presentation

label free classification of ciliated cells using
SMART_READER_LITE
LIVE PREVIEW

Label-free classification of ciliated cells using Deep Learning - - PowerPoint PPT Presentation

Label-free classification of ciliated cells using Deep Learning Ketil Tvermosegaard PSI Webinar Acknowledgements: Steven Barrett, Gareth Wayne, James Porter, Luke Markham, Sam Bates, Paul Cooper One-slide summary non-ciliated (not hairy)


slide-1
SLIDE 1

Label-free classification of ciliated cells using Deep Learning

Ketil Tvermosegaard PSI Webinar Acknowledgements: Steven Barrett, Gareth Wayne, James Porter, Luke Markham, Sam Bates, Paul Cooper

slide-2
SLIDE 2

One-slide summary

2

ciliated (hairy) non-ciliated (not hairy)

slide-3
SLIDE 3

One-slide summary

3

ciliated (hairy) non-ciliated (not hairy) Gareth Wayne (Novel Human Genetics) New “image cytometer” (ca. £500K) Produces pictures of cells (thousands). Can we use ML to label cells?

slide-4
SLIDE 4

One-slide summary

4

ciliated (hairy) non-ciliated (not hairy) Gareth Wayne (Novel Human Genetics) New “image cytometer” (ca. £500K) Produces pictures of cells (thousands). Can we use ML to label cells? Ketil Tvermosegaard and Steven Barrett (Research Statistics) Sounds like a Deep Learning problem. But can we access images? And will DL work?

slide-5
SLIDE 5

One-slide summary

5

ciliated (hairy) non-ciliated (not hairy) Gareth Wayne (Novel Human Genetics) New “image cytometer” (ca. £500K) Produces pictures of cells (thousands). Can we use ML to label cells? Ketil Tvermosegaard and Steven Barrett (Research Statistics) Sounds like a Deep Learning problem. But can we access images? And will DL work? Paul Cooper, Sam Bates, Luke Markham (Tessella) Improve Ketil and Steven’s POC Build prototype code for production

slide-6
SLIDE 6

One-slide summary

6

ciliated (hairy) non-ciliated (not hairy) Gareth Wayne (Novel Human Genetics) New “image cytometer” (ca. £500K) Produces pictures of cells (thousands). Can we use ML to label cells? Ketil Tvermosegaard and Steven Barrett (Research Statistics) Sounds like a Deep Learning problem. But can we access images? And will DL work? Paul Cooper, Sam Bates, Luke Markham (Tessella) Improve Ketil and Steven’s POC Build prototype code for production Actual confusion matrix from trained network on data from NEW experiment!

slide-7
SLIDE 7

Key Learnings

7

slide-8
SLIDE 8

Key Learnings

8

DL is not hard to use POC network took about 100 lines

  • f R-code using keras package

And this code was mostly lifted from a tutorial example used to recognise images of fruit

slide-9
SLIDE 9

Key Learnings

9

DL is not hard to use POC network took about 100 lines

  • f R-code using keras package

And this code was mostly lifted from a tutorial example used to recognise images of fruit DL is hard to use well Essentially an infinite-dimensional

  • ptimisation problem
  • architecture
  • hyper-parameters
  • pre-processing

Solved by Tessella’s “test bench” approach

slide-10
SLIDE 10

Key Learnings

10

DL is not hard to use POC network took about 100 lines

  • f R-code using keras package

And this code was mostly lifted from a tutorial example used to recognise images of fruit DL is hard to use well Essentially an infinite-dimensional

  • ptimisation problem
  • architecture
  • hyper-parameters
  • pre-processing

Solved by Tessella’s “test bench” approach Is DL right for you? Is the data the right kind?

  • After a lot of work, we got access to images

Do you have enough?

  • 10s of thousands of manually labelled images

Is your problem actually a DL problem?

  • We had a clear visual phenotype

(recognisable to non-expert)

slide-11
SLIDE 11

Timeline of pre-TAP work

11

Problem Definition

  • Derived features
  • Proprietary file format

ML on features?

  • Failed

“Crack” file format?

  • Success

Deep Learning?

  • Fruits

POC

  • DL sufficient to proceed

Contact Tessella

  • Started Tessella Analytics

Partnership (TAP) project

……

slide-12
SLIDE 12

Timeline of pre-TAP work

12

Problem Definition

  • Derived features
  • Proprietary file format

ML on features?

  • Failed

“Crack” file format?

  • Success

Deep Learning?

  • Fruits

POC

  • DL sufficient to proceed

Contact Tessella

  • Started Tessella Analytics

Partnership (TAP) project

“My computer science definition of progress: generating new error messages”

……

slide-13
SLIDE 13

Timeline of TAP work

13

Build Refine Apply

slide-14
SLIDE 14

Timeline of TAP work

14

Build Refine Apply

▪ Start naïve (fruits) ▪ Network “test bench” ▪ Test architecture and hyper-params

slide-15
SLIDE 15

Timeline of TAP work

15

Build Refine Apply

▪ Data augmentation ▪ Architecture ▪ Optimisation

slide-16
SLIDE 16

Timeline of TAP work

16

Build Refine Apply

▪ New data ▪ New cell types ▪ Mis-labels

slide-17
SLIDE 17

Source: Gareth Wayne

Epithelial Differentiation screening

Insert your date / confidentiality text here 17

slide-18
SLIDE 18

Source: Gareth Wayne

Epithelial Differentiation screening

Insert your date / confidentiality text here 18

Does editing target change differentiation? (Medium throughput screen of candidate targets)

slide-19
SLIDE 19

Ciliated cells

19

… on cells. What is it good for? Important in respiratory indications like… transportation protection secretion biophysics.org

slide-20
SLIDE 20

In principle, flow cytometry is easy…

Flow cytometry (FACS)

20

slide-21
SLIDE 21

In principle, flow cytometry is easy…

Flow cytometry (FACS)

21

Label cells with fluorophore-antibody pairs

slide-22
SLIDE 22

In principle, flow cytometry is easy…

Flow cytometry (FACS)

22

Use laser to read wavelength of light emitted by each cell

slide-23
SLIDE 23

In principle, flow cytometry is easy…

Flow cytometry (FACS)

23

Classify cells based on label expression

slide-24
SLIDE 24

But in practice…

Flow cytometry (FACS)

24

Do we really get a single cell at a time? Fluorophores overlap! What is the appropriate “gating” to identify cell types? + Gating is sequential!

slide-25
SLIDE 25

– Image Flow Cytometry = Flow cytometry + Cell imaging camera … to the rescue?

Image from Merck Millipore / Amnis

Image Flow Cytometry

25

slide-26
SLIDE 26

– Scientist using FACS to determine if epithelial cells were ciliated (“hairy”) or not – Using single cell images (Image Flow Cytometry) to validate findings – Validation not always consistent with FACS

representative images

The “scientific problem”

26

ciliated non-ciliated ciliation marker

slide-27
SLIDE 27

– Many thousands of images (5,000 – 10,000 cells in a well, approx. 30 wells to a plate) – Derived numerical features available (sphericity, diameter, etc) – Image files in proprietary file format – Obvious visual phenotype

The “statistical problem”

27

slide-28
SLIDE 28

– Special case of (Artificial) Neural Network, characterized by having multiple layers

Deep Learning

28

slide-29
SLIDE 29

– Special case of (Artificial) Neural Network, characterized by having multiple layers

Deep Learning

29

slide-30
SLIDE 30

– Special case of (Artificial) Neural Network, characterized by having multiple layers – Many kinds of layers. We use activation, convolutional, pooling, dropout

Deep Learning

30

slide-31
SLIDE 31

– Special case of (Artificial) Neural Network, characterized by having multiple layers – Many kinds of layers. We use activation, convolutional, pooling, dropout

image: https://medium.com/@akankshamalhotra24/tutorial-on-feedforward-neural-network-part-1-659eeff574c3

Deep Learning

31

𝑕 𝑥′𝑦 + 𝑐 = 𝑕(Σ𝑗=1

𝑜 𝑥𝑗𝑦𝑗 + 𝑐) , with e.g. 𝑕 = tanh

Activation: each node has (scalar-valued) output # Parameters: (one weight vector w of same length as x plus one single bias scalar) X (# nodes) Intuition: combination of linear transformation and (softly) step-like functions = flexible function approximation NB: “The Universal Approximation Theorem”

I.e., each node =

slide-32
SLIDE 32

– Special case of (Artificial) Neural Network, characterized by having multiple layers – Many kinds of layers. We use activation, convolutional, pooling, dropout

Deep Learning

32

Convolution: each node has (array-valued) output. A “filter” array is multiplied element-wise onto the input array and the sum is taken. This filter is run across the entire input array yielding a new (smaller) array. Example output for 2-dimensional input: 𝑎𝑗,𝑘 = Σ𝑏=1

𝑙

Σ𝑐=1

𝑙

𝐺

𝑏,𝑐𝑦𝑗+𝑏,𝑘+𝑐

# Parameters: (#cells in filter) X (#nodes in layer) Intuition: each node can learn a “feature”. E.g. circles, horizontal lines, etc.

slide-33
SLIDE 33

– Special case of (Artificial) Neural Network, characterized by having multiple layers – Many kinds of layers. We use activation, convolutional, pooling, dropout

Deep Learning

33

Pooling: each node has (array-valued) output. The input array is divided into a grid and a simple “pooling function” is applied to all the cells in each “grid chunk”. # Parameters: none Hyper-parameters: size of filter, step size Intuition: data compression + trying to extract “salient features” (data might be “grainy”)

slide-34
SLIDE 34

– Special case of (Artificial) Neural Network, characterized by having multiple layers – Many kinds of layers. We use activation, convolutional, pooling, dropout

Deep Learning

34

Dropout: Every iteration of training, randomly drop a fixed proportion of nodes in the layer # Parameters: none Hyper-parameters: dropout rate / number Intuition: ”Robustification” against dominating/correlated features. Similar in spirit to randomly dropped features in random forests.

slide-35
SLIDE 35

– Special case of (Artificial) Neural Network, characterized by having multiple layers

Deep Learning

35

Gets a special name because It Works! ImageNet Challenge 2012

  • Vast improvement on earlier technologies
  • Many examples followed

(Google translation, speech recognition, etc.)

slide-36
SLIDE 36

Architecture & Hyperparameters

▪ Systematic experimentation to optimise deep-learning architecture and hyper-parameters ▪ Ensembling of several networks

Training-set Improvement ▪ Tessella deep-learning image analysis expertise critical to:

▪ Image pre-processing (standardisation of raw cell images) ▪ Image augmentation (perturbation of input to increase volume/diversity of training set)

▪ ‘Ground truth’ improvement (re-presentation of false+/-’s to human experts) ▪ Further human image labelling (more experiments, other primary cell donors) Process & Progress

Tessella Analytics Partnership Project

36

source: Steven Barrett

slide-37
SLIDE 37

– Facilitates experimentation with different network architectures – Iterate over hyper-parameters – Record results in a database-style format

Test bench, conceptually

37

K-fold validation Hyper-parameters Experiments database Training metrics Post-processing Network architecture Trained model

source: Tessella

slide-38
SLIDE 38

– “Wrapper” function which takes hyperparameters + architecture as input and returns uniquely ID’ed output (input + performance metrics) Simple, but highly effective

source: Tessella

Test bench, concretely

Insert your date / confidentiality text here 38

slide-39
SLIDE 39

Neural Network Ensembles

Augment 0.59 Predict 0.36 Average

𝑔

2 𝑦

𝑔

1 𝑦

𝑔

3 𝑦

0.52 0.51

𝑔

2 𝑦

𝑔

1 𝑦

0.18 0.19 0.15

𝑔

3 𝑦

𝑔

2 𝑦

𝑔

1 𝑦

0.39 0.40 0.30

𝑔

3 𝑦

– Combine the results of multiple trained classifiers in a weighted voting system – Each classifier has learnt different features and found slightly different optima

– Average of predictions more robust to unimportant differences between optima

source: Tessella

representative images

slide-40
SLIDE 40

Augmentation exposes areas which were not imaged How to fill in the gaps? – Initially tried nearest pixel method – Deforms cells in some cases Solution – Estimate background – Standardise image by piecewise linear transform, setting background to fixed value – Fill exposed pixels with the same value

Augmentation exposed areas:

40

source: Tessella

representative images

slide-41
SLIDE 41

Situation Gene KO experiment with two main readouts ▪ marker-based flow cytometry (FACS) ▪ image flow cytometry (ImageStream), processed with Deep Learning Purposes (i) to investigate whether KO of the hits modulates ciliation (ii) whether FACS and ImageStream concur Action Experimental design was provided: A complete block design with three blocks (each block a separate replication of the full KO experiment). Linear mixed effects models were fitted separately for each endpoint, to control for block effects. Impact Some strong disagreements between FACS and ImageStream. This was expected, based on previously observed disagreements as well as specific biological hypotheses regarding e.g. Target 2. The experiment and analysis provided actionable results (ImageStream is considered the “real” readout”) for further work and crucial confirmation

  • f conjectured problems with FACS

Hit Validation in Modulation of Ciliation

41

Ketil Tvermosegaard, Research Statistics

95% confidence intervals for estimated difference in proportion of ciliated cells between cells with given gene KO and cells with non-targeted sgRNA. Transformed to ratio (from log10 scale).

slide-42
SLIDE 42

Insert your date / confidentiality text here 16x9 core template 42

Thank you!