but were forced to find out
play

BUT WERE FORCED TO FIND OUT Ivan tajduhar istajduh@riteh.hr SSIP - PowerPoint PPT Presentation

EVERYTHING YOU NEVER WANTED TO KNOW ABOUT MACHINE LEARNING, BUT WERE FORCED TO FIND OUT Ivan tajduhar istajduh@riteh.hr SSIP 2019 27 TH SUMMER SCHOOL ON IMAGE PROCESSING, TIMISOARA, ROMANIA July 10 th 2019 EVERYTHING YOU NEVER WANTED TO KNOW


  1. EVERYTHING YOU NEVER WANTED TO KNOW ABOUT MACHINE LEARNING, BUT WERE FORCED TO FIND OUT Ivan Štajduhar istajduh@riteh.hr SSIP 2019 27 TH SUMMER SCHOOL ON IMAGE PROCESSING, TIMISOARA, ROMANIA July 10 th 2019

  2. EVERYTHING YOU NEVER WANTED TO KNOW ABOUT MACHINE LEARNING, BUT WERE FORCED TO FIND OUT IN INTRODUCTION AND MOTIVATION

  3. Challenges

  4. Solution • Model-based techniques – Manually tailored – Variation and complexity in clinical data – Limited by current insights into clinical conditions, diagnostic modelling and therapy – Hard to establish analytical solutions

  5. Hržić , Franko, et al. "Local-Entropy Based Approach for X-Ray Image Segmentation and Fracture Detection." Entropy 21.4 (2019): 338. (un uniri-tehn hnic-18 18-15) 15)

  6. Machine learning • Model-based techniques – Manually tailored – Variation and complexity in clinical data – Limited by current insights into clinical conditions, diagnostic modelling and therapy – Hard to establish analytical solutions • An alternative: learning from data – Minimising an objective function

  7. CT MRI PACS US PET

  8. Summary • Introduction and motivation • Representation, optimisation & stuff • Evaluation metrics & experimental setup • Improving model performance

  9. EVERYTHING YOU NEVER WANTED TO KNOW ABOUT MACHINE LEARNING, BUT WERE FORCED TO FIND OUT REPRESENTATION, , OPTIMISATION & STUFF

  10. Machine learning • Machine learning techniques mainly deal with representation, performance assessment and optimisation: – The learning process is always preceded by the choice of a formal representation of the model. A set of possible models is called the space of the hypothesis. – Learning algorithm uses a cost function to determine (evalu luate) how successful a model is – Op Optimisation is the process of choosing the most successful models

  11. Hypothesis • Learning type: supervised vs unsupervised unlabelled labelled data data • Hypothesis type: regression vs classification categorical continuous outcome outcome

  12. Hypothesis Data Learning algorithm Observation Outcome h (known variables, (prediction) easily obtainable)

  13. 16

  14. feature predictor extraction

  15. Hypothesis and parameter estimation

  16. Hypothesis and parameter estimation

  17. Regularisation • A way of reducing overfitting by ignoring non-informative features • What is overfitting? • Many types of regularisation – Quadratic regulariser

  18. Multilayer perceptron (MLP) • An extension of the logistic-regression idea

  19. Multilayer perceptron (MLP) • Parameters estimated through backpropagation algorithm evidence error CS231n: Convolutional Neural Networks for Visual Recognition, Stanford University http://cs231n.stanford.edu/2016/

  20. Multilayer perceptron (MLP) Convolutional layer Normalisation layer Activation function Fully-connected layer Normalisation layer Activation function CS231n: Convolutional Neural Networks for Visual Recognition, Stanford University http://cs231n.stanford.edu/2016/

  21. Support vector machine (SVM) • Maximum margin classifier

  22. Support vector machine (SVM) • Often used with kernels for dealing with linearly non- separable problems • Quadratic programming solver optimisation

  23. Support vector machine (SVM)

  24. Kernels • A possible way of dealing with linearly non-separable problems • A measure of similarity between data points • Kernel trick can implicitly escalate dimensionality

  25. Tree models • An alternative form of mapping • Partition the input space into cuboid regions – Can be easily interpretable Temperature hot cold moderate yes yes Sky rainy cloudy sunny yes no no

  26. Tree models • An alternative form of mapping • Partition the input space into cuboid regions – Can be easily interpretable

  27. CART • CART hypothesis • Local optimisation (greedy) – recursive partitioning • Cost function

  28. | CART |

  29. Short recap • Representation, optimisation & stuff |

  30. EVERYTHING YOU NEVER WANTED TO KNOW ABOUT MACHINE LEARNING, BUT WERE FORCED TO FIND OUT EVALUATION METRICS & EXPERIMENTAL SETUP

  31. Evaluation metrics • The choice of an adequate model-evaluation metric depends on the modelling goal • Common metrics for common problems: – Mean squared error (for regression) – Classification accuracy

  32. outcome = class = label Evaluation metrics Confusion matrix (normalised) • Confusion matrix – binary case – multiple classes (K>2) Observed outcome Predicted Predicted outcome outcome Observed True False outcome positive negative (TP) (FN) Observed False True outcome positive negative (FP) (TN) 36 Predicted outcome

  33. Evaluation metrics • Common metrics for class-imbalanced problems, or when misclassifications are not equally bad: – Sensitivity (recall, true positive rate) Predicted Predicted – Specificity outcome outcome Observed True False – F1 score outcome positive negative (TP) (FN) Observed False True outcome positive negative (FP) (TN) 37

  34. Evaluation metrics • For class-imbalanced problems, or when misclassifications are not equally bad (probabilistic classification): – Receiver operating characteristic (ROC) curve Predicted Predicted outcome outcome Observed True False outcome positive negative (TP) (FN) 1-SPEC Observed False True outcome positive negative (FP) (TN) Fawcett, Tom. "An introduction to ROC analysis." Pattern recognition letters 27.8 (2006): 861-874.

  35. Evaluation metrics • For highly-skewed class distributions (probabilistic classification): – Precision-recall (PR) curve – Area under the curve (AUC) AUROC Davis, Jesse, and Mark Goadrich. "The relationship between Precision-Recall and ROC curves." Proceedings of the 23rd international conference on Machine learning. ACM, 2006.

  36. Evaluation metrics • Uncertain labellings, e.g., censored survival data • Kaplan-Meier estimate of a survival function Kaplan-Meier estimate • Alternative evaluation metrics: – Log-rank test – Concordance index – Explained residual variation of integrated Brier score Risk group: LOW HIGH

  37. Experimental setup • Setting up an unbiased experiment – How well will the model perform on new, yet unseen, data? train test • Dataset split: training • n-fold cross-validation or leave-one-out test • Fold stratification of class-wise distribution • Multiple iterations of fold splits

  38. Experimental setup • Estimating a model wisely – validation data used for tuning hyperparameters Data Learning algorithm training MoA

  39. Experimental setup training • Fair estimate of a classifier / regression model – data preprocessing can bias your conclusions – watch out for naturally correlated observations, e.g. • diagnostic scan of a patient now and before • different imaging modalities of the same subject – artificially generated data can cause a mess – use testing data only after you are done with the training • Fair estimate of a segmenter – splitting pixels of the same image into separate sets is usually not the best idea – also, use a separate testing set of images How well will the model perform on new, yet unseen, data?

  40. Method performance comparison • You devised a new method – What makes your method better? – Is it significantly better? • Is the sample representative of a population? • Hypothesis: It is better (maybe) POPULATION

  41. Method performance comparison • Hypothesis: A equals B! – Non-parametric statistical tests – A level of significance α is used to determine at which level the hypothesis may be rejected – The smaller the p -value, the stronger the evidence against the null hypothesis Demšar, Janez. "Statistical comparisons of classifiers over multiple data sets." Journal of Machine learning research 7.Jan ( 2006): 1-30. Derrac, Joaquín, et al. "A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms." Swarm and Evolutionary Computation 1.1 (2011): 3-18.

  42. Two classifiers • Comparing performance against baseline (over multiple datasets) • Wilcoxon signed ranks test Demšar, Janez. "Statistical comparisons of classifiers over multiple data sets." Journal of Machine learning research 7.Jan ( 2006): 1-30.

  43. Multiple classifiers • Comparing performance of multiple classifiers (over multiple datasets) • Friedman test • Post-hoc tests Demšar, Janez. "Statistical comparisons of classifiers over multiple data sets." Journal of Machine learning research 7.Jan ( 2006): 1-30.

  44. Multiple classifiers • Friedman test Demšar, Janez. "Statistical comparisons of classifiers over multiple data sets." Journal of Machine learning research 7.Jan ( 2006): 1-30.

  45. Multiple classifiers • Post-hoc tests • All-vs-all – Nemenyi test • One-vs-all – Bonferroni-Dunn test worse Demšar, Janez. "Statistical comparisons of classifiers over multiple data sets." Journal of Machine learning research 7.Jan ( 2006): 1-30.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend