what we did not cover
play

What We Did Not Cover COMPSCI 371D Machine Learning COMPSCI 371D - PowerPoint PPT Presentation

What We Did Not Cover COMPSCI 371D Machine Learning COMPSCI 371D Machine Learning What We Did Not Cover 1 / 9 What We Did Not Cover 1 Much More Detail 2 Statistical Machine Learning 3 Other Supervised Techniques 4 Reducing the Burden of


  1. What We Did Not Cover COMPSCI 371D — Machine Learning COMPSCI 371D — Machine Learning What We Did Not Cover 1 / 9

  2. What We Did Not Cover 1 Much More Detail 2 Statistical Machine Learning 3 Other Supervised Techniques 4 Reducing the Burden of Labeling 5 Unsupervised Methods 6 Addressing Multiple Learning Tasks Together 7 Prediction over Time COMPSCI 371D — Machine Learning What We Did Not Cover 2 / 9

  3. Much More Detail Much More Detail • Computationally efficient training algorithms: Optimization techniques • Deep learning architectures for special problems: Image motion analysis, video analysis, ... COMPSCI 371D — Machine Learning What We Did Not Cover 3 / 9

  4. Statistical Machine Learning Statistical Machine Learning • How to measure the size of H : Vapnik-Chervonenkis dimension, Rademacher complexity • How large must T be to get an h that is within ǫ of a performance target with probability greater than 1 − δ : Probably Approximately Correct (PAC) learning (We saw a glimpse of PAC when we looked at the sample complexity of linear predictors and SVMs) • H is learnable if there exists a size of T that is large enough for this goal to be achieved • Which H s are learnable? • How large must S be to get a performance measure accurate within ǫ : Concentration bounds, statistical estimation theory, PAC-like techniques COMPSCI 371D — Machine Learning What We Did Not Cover 4 / 9

  5. Other Supervised Techniques Other Supervised Techniques • Boosting: How to use many bad predictors to make one good one Similar in principle to ensemble predictors, different assumptions and techniques • Learning to rank Example: Learning a better Google COMPSCI 371D — Machine Learning What We Did Not Cover 5 / 9

  6. Reducing the Burden of Labeling Reducing the Burden of Labeling • Semi-supervised methods: Build models of the data x to leverage sparse labels y • Domain adaptation: Train a classifier on source-domain labeled data ( x , y ) and target-domain unlableled data x so that is works well in the target domain TRAIN Vegas Building Detector Vegas Image Vegas Building Detector Shanghai Image ADAPT Shanghai Building Detector Shanghai Image https://en.wikipedia.org COMPSCI 371D — Machine Learning What We Did Not Cover 6 / 9

  7. Unsupervised Methods Unsupervised Methods • Dimensionality reduction: Compressing X ⊆ R d to X ′ ⊆ R d ′ with d ′ ≪ d • Principal or Independent Component Analysis (PCA, ICA) • Manifold learning • Clustering: • K -means • Expectation-Maximization • Agglomerative methods • Splitting methods COMPSCI 371D — Machine Learning What We Did Not Cover 7 / 9

  8. Addressing Multiple Learning Tasks Together Addressing Multiple Learning Tasks Together • Multi-task learning: How to learn representations that are common to different but related prediction tasks COMPSCI 371D — Machine Learning What We Did Not Cover 8 / 9

  9. Prediction over Time Prediction over Time • State-space methods • Time series analysis • Stochastic state estimation • System identification • Recurrent neural networks • Reinforcement learning: Actions over time Learning policies underlying observed sequences COMPSCI 371D — Machine Learning What We Did Not Cover 9 / 9

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend