introduction to ml dl
play

Introduction to ML & DL Shan-Hung Wu shwu@cs.nthu.edu.tw - PowerPoint PPT Presentation

Introduction to ML & DL Shan-Hung Wu shwu@cs.nthu.edu.tw Department of Computer Science, National Tsing Hua University, Taiwan Machine Learning Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 1 / 22 Outline Whats Machine


  1. Introduction to ML & DL Shan-Hung Wu shwu@cs.nthu.edu.tw Department of Computer Science, National Tsing Hua University, Taiwan Machine Learning Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 1 / 22

  2. Outline What’s Machine Learning? 1 What’s Deep Learning? 2 About this Course... 3 FAQ 4 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 2 / 22

  3. Outline What’s Machine Learning? 1 What’s Deep Learning? 2 About this Course... 3 FAQ 4 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 3 / 22

  4. Prior vs. Posteriori Knowledge To solve a problem, we need an algorithm E.g., sorting Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 4 / 22

  5. Prior vs. Posteriori Knowledge To solve a problem, we need an algorithm E.g., sorting A priori knowledge is enough Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 4 / 22

  6. Prior vs. Posteriori Knowledge To solve a problem, we need an algorithm E.g., sorting A priori knowledge is enough For some problem, however, we do not have the a priori knowledge E.g., to tell if an email is spam or not The correct answer varies in time and from person to person Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 4 / 22

  7. Prior vs. Posteriori Knowledge To solve a problem, we need an algorithm E.g., sorting A priori knowledge is enough For some problem, however, we do not have the a priori knowledge E.g., to tell if an email is spam or not The correct answer varies in time and from person to person Machine learning algorithms use the a posteriori knowledge to solve problems Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 4 / 22

  8. Prior vs. Posteriori Knowledge To solve a problem, we need an algorithm E.g., sorting A priori knowledge is enough For some problem, however, we do not have the a priori knowledge E.g., to tell if an email is spam or not The correct answer varies in time and from person to person Machine learning algorithms use the a posteriori knowledge to solve problems Learnt from examples (as extra input) Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 4 / 22

  9. Example Data X as Extra Input Unsupervised: i = 1 , where x ( i ) ∈ R D X = { x ( i ) } N E.g., x ( i ) an email Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 5 / 22

  10. Example Data X as Extra Input Unsupervised: i = 1 , where x ( i ) ∈ R D X = { x ( i ) } N E.g., x ( i ) an email Supervised: i = 1 , where x ( i ) ∈ R D and y ( i ) ∈ R K , X = { ( x ( i ) , y ( i ) ) } N E.g., y ( i ) ∈ { 0 , 1 } a spam label Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 5 / 22

  11. General Types of Learning (1/2) Supervised learning : learn to predict the labels of future data points X ∈ R N × D : x ′ ∈ R D : y ′ ∈ R K : y ∈ R N × K : [ e ( 6 ) , e ( 1 ) , e ( 9 ) , e ( 4 ) , e ( 2 ) ] ? Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 6 / 22

  12. General Types of Learning (1/2) Supervised learning : learn to predict the labels of future data points X ∈ R N × D : x ′ ∈ R D : y ′ ∈ R K : y ∈ R N × K : [ e ( 6 ) , e ( 1 ) , e ( 9 ) , e ( 4 ) , e ( 2 ) ] ? Unsupervised learning : learn patterns or latent factors in X Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 6 / 22

  13. General Types of Learning (2/2) Reinforcement learning : learn from “good”/“bad” feedback of actions (instead of correct labels) to maximize the goal Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 7 / 22

  14. General Types of Learning (2/2) Reinforcement learning : learn from “good”/“bad” feedback of actions (instead of correct labels) to maximize the goal AlphaGo [1] is a hybrid of reinforcement learning and supervised learning Supervised learning from the game records Then, reinforcement learning from self-play Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 7 / 22

  15. General Machine Learning Steps Data collection, preprocessing (e.g., integration, cleaning, etc.), and 1 exploration Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 8 / 22

  16. General Machine Learning Steps Data collection, preprocessing (e.g., integration, cleaning, etc.), and 1 exploration Split a dataset into the training and testing datasets 1 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 8 / 22

  17. General Machine Learning Steps Data collection, preprocessing (e.g., integration, cleaning, etc.), and 1 exploration Split a dataset into the training and testing datasets 1 Model development 2 Assume a model { f ( · ; w ) } that is a collection of candidate functions 1 f ’s (representing posteriori knowledge) we want to discover f is assumed to be parametrized by w 1 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 8 / 22

  18. General Machine Learning Steps Data collection, preprocessing (e.g., integration, cleaning, etc.), and 1 exploration Split a dataset into the training and testing datasets 1 Model development 2 Assume a model { f ( · ; w ) } that is a collection of candidate functions 1 f ’s (representing posteriori knowledge) we want to discover f is assumed to be parametrized by w 1 Define a cost function C ( w ; X ) (or functional C [ f ; X ] ) that measures 2 “how good a particular f can explain the training data” Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 8 / 22

  19. General Machine Learning Steps Data collection, preprocessing (e.g., integration, cleaning, etc.), and 1 exploration Split a dataset into the training and testing datasets 1 Model development 2 Assume a model { f ( · ; w ) } that is a collection of candidate functions 1 f ’s (representing posteriori knowledge) we want to discover f is assumed to be parametrized by w 1 Define a cost function C ( w ; X ) (or functional C [ f ; X ] ) that measures 2 “how good a particular f can explain the training data” Training : employ an algorithm that finds the best (or good enough) 3 function f ∗ ( · ; w ∗ ) in the model that minimizes the cost function w ∗ = argmin w C ( w ; X ) Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 8 / 22

  20. General Machine Learning Steps Data collection, preprocessing (e.g., integration, cleaning, etc.), and 1 exploration Split a dataset into the training and testing datasets 1 Model development 2 Assume a model { f ( · ; w ) } that is a collection of candidate functions 1 f ’s (representing posteriori knowledge) we want to discover f is assumed to be parametrized by w 1 Define a cost function C ( w ; X ) (or functional C [ f ; X ] ) that measures 2 “how good a particular f can explain the training data” Training : employ an algorithm that finds the best (or good enough) 3 function f ∗ ( · ; w ∗ ) in the model that minimizes the cost function w ∗ = argmin w C ( w ; X ) Testing : evaluate the performance of the learned f ∗ using the testing 4 dataset Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 8 / 22

  21. General Machine Learning Steps Data collection, preprocessing (e.g., integration, cleaning, etc.), and 1 exploration Split a dataset into the training and testing datasets 1 Model development 2 Assume a model { f ( · ; w ) } that is a collection of candidate functions 1 f ’s (representing posteriori knowledge) we want to discover f is assumed to be parametrized by w 1 Define a cost function C ( w ; X ) (or functional C [ f ; X ] ) that measures 2 “how good a particular f can explain the training data” Training : employ an algorithm that finds the best (or good enough) 3 function f ∗ ( · ; w ∗ ) in the model that minimizes the cost function w ∗ = argmin w C ( w ; X ) Testing : evaluate the performance of the learned f ∗ using the testing 4 dataset Apply the model in the real world 5 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 8 / 22

  22. Example for Spam Detection Random split of your past emails and labels 1 Training dataset: X = { ( x ( i ) , y ( i ) ) } i 1 Testing dataset: X ′ = { ( x ′ ( i ) , y ′ ( i ) ) } i 2 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 9 / 22

  23. Example for Spam Detection Random split of your past emails and labels 1 Training dataset: X = { ( x ( i ) , y ( i ) ) } i 1 Testing dataset: X ′ = { ( x ′ ( i ) , y ′ ( i ) ) } i 2 Model development 2 Model : { f : f ( x ; w ) = w ⊤ x } 1 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 9 / 22

  24. Example for Spam Detection Random split of your past emails and labels 1 Training dataset: X = { ( x ( i ) , y ( i ) ) } i 1 Testing dataset: X ′ = { ( x ′ ( i ) , y ′ ( i ) ) } i 2 Model development 2 Model : { f : f ( x ; w ) = w ⊤ x } 1 Cost function : C ( w ; X ) = Σ i 1 ( w ; f ( x ( i ) ; w ) � = y ( i ) ) 2 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 9 / 22

  25. Example for Spam Detection Random split of your past emails and labels 1 Training dataset: X = { ( x ( i ) , y ( i ) ) } i 1 Testing dataset: X ′ = { ( x ′ ( i ) , y ′ ( i ) ) } i 2 Model development 2 Model : { f : f ( x ; w ) = w ⊤ x } 1 Cost function : C ( w ; X ) = Σ i 1 ( w ; f ( x ( i ) ; w ) � = y ( i ) ) 2 Training : to solve w ∗ = argmin w Σ i 1 ( w ; f ( x ( i ) ; w ) � = y ( i ) ) 3 Shan-Hung Wu (CS, NTHU) Introduction Machine Learning 9 / 22

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend