Machine Learning Basics
- Prof. Kuan-Ting Lai
2020/4/4
Machine Learning Basics Prof. Kuan-Ting Lai 2020/4/4 Machine - - PowerPoint PPT Presentation
Machine Learning Basics Prof. Kuan-Ting Lai 2020/4/4 Machine Learning Francois Chollet , Deep Learning with Python, Manning, 2017 Machine Learning Flow ( ) ( ) ( ) Data Evaluation Training
2020/4/4
Francois Chollet, “Deep Learning with Python,” Manning, 2017
(收集資料) Data (評估準確度) Evaluation (Loss Function) (訓練模型) Training (Optimization)
4
Supervised Learning
Classification Regression
Unsupervised Learning
Clustering Dimensionality Reduction
Reinforcement Learning
Deep Reinforcement Learning
Supervised Learning
Classification Regression
Unsupervised Learning
Clustering Dimensionality Reduction
Reinforcement Learning
Deep Reinforcement Learning
5
Has a teacher to label data!
6
Supervised Learning
Classification (分門別類) Regression (回歸分析)
Unsupervised Learning
Clustering (物以類聚) Dimensionality Reduction (化繁為簡)
Reinforcement Learning
Deep Reinforcement Learning
(連續資料) (離散資料)
7
9
(Discrete) (Continuous)
https://towardsdatascience.com/data-types-in-statistics-347e152e8bee
machine-learning-statistics-and-data-mining/
Supervised Learning
Regression Classification
Unsupervised Learning
Clustering Dimension Reduction
Jebaseelan Ravi @ Medium
are not linearly separable
https://www.tensorflow.org/tutorials/customization/custom_training_walkthrough
𝑦1 𝑦2
https://datascience.stackexchange.com/questions/17536/kernel-trick-explanation
https://www.tensorflow.org/tutorials/customization/custom_training_walkthrough
Supervised Learning
Regression Classification
Unsupervised Learning
Clustering Dimension Reduction
errors
Supervised Learning
Regression Classification
Unsupervised Learning
Clustering Dimension Reduction
𝑇 𝑦 = 𝑓𝑦 𝑓𝑦 + 1 = 1 1 + 𝑓−𝑦
https://en.wikipedia.org/wiki/Sigmoid_function
𝑇 𝑦 = 𝑇 𝑦 (1- 𝑇 𝑦 )
S-shaped curve
𝑧′ = 𝑄 𝑦, 𝑥 = 𝑄𝜄 𝑦 = 1 1 + 𝑓− 𝒙𝑈𝒚+𝑐 𝑧′ = ቊ0, 𝑦 < 𝑢 1, 𝑦 ≥ 𝑢
loss= ൝− log 1 − 𝑄𝜄 𝑦 , 𝑗𝑔 𝑧 = 0 − log 𝑄𝜄 𝑦 , 𝑗𝑔 𝑧 = 1
https://towardsdatascience.com/a-guide-to-neural-network-loss-functions-with-applications-in-keras-3a3baa9f71c5
loss= ൝− log 1 − 𝑄𝜄 𝑦 , 𝑗𝑔 𝑧 = 0 − log 𝑄𝜄 𝑦 , 𝑗𝑔 𝑧 = 1 ⇒ 𝑀𝜄(x) = −𝑧 log 𝑄𝜄 𝑦 + − (1 − y)log 1 − 𝑄𝜄 𝑦 ∇𝑀𝑋(x) = − 𝑧 − 𝑄𝜄 𝑦 𝑦
https://towardsdatascience.com/a-guide-to-neural-network-loss-functions-with-applications-in-keras-3a3baa9f71c5
https://towardsdatascience.com/workflow-of-a-machine-learning-project-ec1dba419b94
Overfitting Underfitting
https://en.wikipedia.org/wiki/Overfitting
to training data and does not generalize on unseen test data
http://scott.fortmann-roe.com/docs/BiasVariance.html
course/regularization-for-sparsity/l1-regularization
46
https://en.wikipedia.org/wiki/Confusion_matrix
https://en.wikipedia.org/wiki/Confusion_matrix
Coronavirus Example
https://www.facebook.com/numeracylab/posts/2997362376951435
−P: positive samples, N: negative samples, P’: predicted positive samples, TP: true positives, TN: true negatives
TP P
TP P′
TP+TN 𝑄+N
2
1 𝑠𝑓𝑑𝑏𝑚𝑚 + 1 𝑞𝑠𝑓𝑑𝑗𝑡𝑗𝑝𝑜
Characteristic) Curve
Recall Precision False Positive Rate (FPR) True Positive Rate (TPR)
53
Pedro Domingos, “A Few Useful Things to Know about Machine Learning,” Commun. ACM, 2012
examples in the training set
−Every learner must embody some knowledge or assumptions beyond the data
programs
training data but only 50% on test data, when in fact it could have 75% on both, it has overfit.
variance
− Cross validation − Add regularization term
high-dimensional
dimensionality of the examples grows
understand and drive force for algorithm design
classification-in-python/