overview of decision trees ensemble
play

Overview of Decision Trees, Ensemble Methods and Reinforcement - PowerPoint PPT Presentation

Overview of Decision Trees, Ensemble Methods and Reinforcement Learning CMSC 678 UMBC Outline Decision Trees Ensemble Methods Bagging Random Forests Reinforcement Learning Decision Trees 20 Questions: http://20q.net/ Goals: 1.


  1. Overview of Decision Trees, Ensemble Methods and Reinforcement Learning CMSC 678 UMBC

  2. Outline Decision Trees Ensemble Methods Bagging Random Forests Reinforcement Learning

  3. Decision Trees “20 Questions”: http://20q.net/ Goals: 1. Figure out what questions to ask 2. In what order 3. Determine how many questions are enough 4. What to predict at the end Adapted from Hamed Pirsiavash

  4. Example: Learning a decision tree Course ratings dataset Adapted from Hamed Pirsiavash

  5. Example: Learning a decision tree Course ratings dataset Rating is the label Adapted from Hamed Pirsiavash

  6. Example: Learning a decision tree Course ratings dataset Questions are features Rating is the label Adapted from Hamed Pirsiavash

  7. Example: Learning a decision tree Course ratings dataset Questions are features Responses are feature values Rating is the label Idea: Predict the label by forming a tree where each node branches on values of particular features Adapted from Hamed Pirsiavash

  8. Example: Learning a decision tree Course ratings dataset Questions are features Responses are feature values Rating is the label Easy? Adapted from Hamed Pirsiavash

  9. Example: Learning a decision tree Course ratings dataset Questions are features Responses are feature values Rating is the label Easy? Easy: yes Easy: no AI? Adapted from Hamed Pirsiavash

  10. Example: Learning a decision tree Course ratings dataset Questions are features Responses are feature values Rating is the label Easy? Easy: yes Easy: no AI? AI: yes AI: no …. Adapted from Hamed Pirsiavash

  11. Example: Learning a decision tree Course ratings dataset Questions are features Responses are feature values Rating is the label Easy? Easy: yes Easy: no AI? Sys? AI: yes AI: no …. Adapted from Hamed Pirsiavash

  12. Example: Learning a decision tree Course ratings dataset Questions are features Responses are feature values Rating is the label Easy? Easy: yes Easy: no AI? Sys? AI: yes Sys: yes Sys: no AI: no …. …. Adapted from Hamed Pirsiavash

  13. Predicting with a Decision Tree is Done Easily and Recursively CIML, Ch 1

  14. There Are Many Ways to Learn a Decision Tree 1. Greedy/Count: What is the most accurate feature at each decision point? 1. See CIML Ch. 1 (and next slides) 2. Maximize information gain at each step 1. Most popular approaches: ID3, C4.5 3. Account for statistical significance 1. Example: Chi-square automatic interaction detection (CHAID) 4. Other task-specific ones (including clustering based)

  15. CIML, Ch 1

  16. counting CIML, Ch 1

  17. counting recursive CIML, Ch 1

  18. simple base cases counting recursive CIML, Ch 1

  19. Outline Decision Trees Ensemble Methods Bagging Random Forests Reinforcement Learning

  20. Ensembles Key Idea: “Wisdom of the crowd“ groups of people can often make better decisions than individuals Apply this to ML Learn multiple classifiers and combine their predictions

  21. Combining Multiple Classifiers by Voting Train several classifiers and take majority of predictions Courtesy Hamed Pirsiavash

  22. Combining Multiple Classifiers by Voting Train several classifiers and take majority of predictions For regression use mean or median of the predictions For ranking and collective classification use some form of averaging

  23. Combining Multiple Classifiers by Voting Train several classifiers and take majority of predictions For regression use mean or median of the predictions For ranking and collective classification use some form of averaging A common family of approaches is called bagging

  24. Bagging: Split the Data Q: What can go wrong with option 1? Option 1: Split the data into K pieces and train a classifier on each

  25. Bagging: Split the Data Q: What can go wrong with option 1? Option 1: Split the data into K pieces and train a classifier on each A: Small sample → poor performance

  26. Bagging: Split the Data Q: What can go wrong with option 1? Option 1: Split the data into K pieces and train a classifier on each A: Small sample → poor performance Option 2: Bootstrap aggregation (bagging) resampling

  27. Bagging: Split the Data Q: What can go wrong with option 1? Option 1: Split the data into K pieces and train a classifier on each A: Small sample → poor performance Option 2: Bootstrap aggregation (bagging) Given a dataset D… resampling Obtain datasets D 1 , D 2 , … , D N sampling with using bootstrap resampling replacement from D get new datasets D̂ by random sampling with replacement from D Courtesy Hamed Pirsiavash

  28. Bagging: Split the Data Q: What can go wrong with option 1? Option 1: Split the data into K pieces and train a classifier on each A: Small sample → poor performance Option 2: Bootstrap aggregation (bagging) Given a resampling dataset D… Obtain datasets D 1 , D 2 , … , D N using bootstrap resampling sampling with replacement from D Train classifiers on each dataset and average their get new datasets D̂ by predictions random sampling with replacement from D Courtesy Hamed Pirsiavash

  29. Why does averaging work? Averaging reduces the variance of estimators Courtesy Hamed Pirsiavash

  30. Why does averaging work? Averaging reduces the variance of estimators f: Generating line y: observed data 𝑕 𝑗 : Learned polynomial regression Courtesy Hamed Pirsiavash

  31. Why does averaging work? Averaging reduces the variance of estimators 50 samples Averaging is a form of regularization: each model can individually overfit but the average is able to overcome the overfitting Courtesy Hamed Pirsiavash

  32. Bagging Decision Trees How would it work?

  33. Bagging Decision Trees How would it work? Bootstrap sample S samples {(X 1 , Y 1 ), …, (X S , Y S )} Train a tree t s on (X s , Y s ) At test time: ො 𝑧 = avg(𝑢 1 𝑦 , … 𝑢 𝑇 𝑦 )

  34. Random Forests Bagging trees with one modification At each split point, choose a random subset of features of size k and pick the best among these Train decision trees of depth d Average results from multiple randomly trained trees Q : What’s the difference between bagging decision trees and random forests? Courtesy Hamed Pirsiavash

  35. Random Forests Bagging trees with one modification At each split point, choose a random subset of features of size k and pick the best among these Train decision trees of depth d Average results from multiple randomly trained trees A : Bagging → highly Q : What’s the difference between bagging decision correlated trees (reuse good trees and random forests? features) Courtesy Hamed Pirsiavash

  36. Random Forests: Human Pose Estimation (Shotton et al., CVPR 2011) Training: 3 trees, 20 deep, 300k training images per tree, 2000 training example pixels per image, 2000 candidate features θ, and 50 candidate thresholds τ per feature (Takes about 1 day on a 1000 core cluster)

  37. (Shotton et al., CVPR 2011)

  38. Outline Decision Trees Ensemble Methods Bagging Random Forests Reinforcement Learning

  39. There’s an entire book! http://incompleteideas. net/book/the-book- 2nd.html

  40. Reinforcement Learning environment agent https://static.vecteezy.com/system/resources/previews/000/0 Robot image: openclipart.org 90/451/original/four-seasons-landscape-illustrations-vector.jpg

  41. Reinforcement Learning take action environment agent https://static.vecteezy.com/system/resources/previews/000/0 Robot image: openclipart.org 90/451/original/four-seasons-landscape-illustrations-vector.jpg

  42. Reinforcement Learning take action get new state environment and/or reward agent https://static.vecteezy.com/system/resources/previews/000/0 Robot image: openclipart.org 90/451/original/four-seasons-landscape-illustrations-vector.jpg

  43. Reinforcement Learning take action get new state environment and/or reward agent https://static.vecteezy.com/system/resources/previews/000/0 Robot image: openclipart.org 90/451/original/four-seasons-landscape-illustrations-vector.jpg

  44. Markov Decision Process: Formalizing Reinforcement Learning take action get new state environment and/or reward agent Markov Decision (𝒯, 𝒝, ℛ, 𝑞, 𝛿) Process: https://static.vecteezy.com/system/resources/previews/000/0 Robot image: openclipart.org 90/451/original/four-seasons-landscape-illustrations-vector.jpg

  45. Markov Decision Process: Formalizing Reinforcement Learning take action get new state environment and/or reward agent set of possible actions Markov Decision (𝒯, 𝒝, ℛ, 𝑞, 𝛿) Process: set of possible states https://static.vecteezy.com/system/resources/previews/000/0 Robot image: openclipart.org 90/451/original/four-seasons-landscape-illustrations-vector.jpg

  46. Markov Decision Process: Formalizing Reinforcement Learning take action get new state environment and/or reward agent set of possible actions Markov Decision (𝒯, 𝒝, ℛ, 𝑞, 𝛿) Process: set of reward of possible (state, states action) pairs https://static.vecteezy.com/system/resources/previews/000/0 Robot image: openclipart.org 90/451/original/four-seasons-landscape-illustrations-vector.jpg

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend