feature selection gavin brown cs man ac uk gbrown the
play

Feature Selection Gavin Brown www.cs.man.ac.uk/~gbrown The Usual - PowerPoint PPT Presentation

COMP61011 : Machine Learning Feature Selection Gavin Brown www.cs.man.ac.uk/~gbrown The Usual Supervised Learning Approach data + labels Learning Algorithm Model Predicted label Testing data Predicting Recurrence of Lung Cancer ofi Only a


  1. COMP61011 : Machine Learning Feature Selection Gavin Brown www.cs.man.ac.uk/~gbrown

  2. The Usual Supervised Learning Approach data + labels Learning Algorithm Model Predicted label Testing data

  3. Predicting Recurrence of Lung Cancer ofi Only a few genes actually matter! Need small, interpretable subset to help doctors! fl

  4. Text classi fi cation.... is this news story “ i nterestin g” ? “ B ag-of-Words ” representation: x = { 0 , 3 , 0 , 0 , 1 , ..., 2 , 3 , 0 , 0 , 0 , 1 } one entry per word! Easily 50,000 words! Very sparse - easy to over fi t! Need accuracy, otherwise we lose visitors to our news website!

  5. The Usual Supervised Learning Approach ????? data + labels Learning Algorithm – OVERWHELMED! Model Predicted label Testing data

  6. With big data….  Time complexity  Computational cost  Cost in data collection  Over-fitting  Lack of interpretability Feature selection

  7. Some things matter, Some do not. Relevant features - those that we need to perform well Irrelevant features - those that are simply unnecessary Redundant features - those that become irrelevant in the presence of others

  8. 3 main categories of Feature Selection techniques: Wrappers, Filters, Embedded methods

  9. Wrappers: Evaluation method Feature set Pros :  Model-oriented  Usually gets good performance for the model you choose. Trains a model Cons : Trains a model  Hugely computationally expensive. Outputs accuracy

  10. Wrappers: Search strategy  With an exhaustive search 101110000001000100001000000000100101010 20 features … 1 million feature sets to check 25 features … 33.5 million sets 30 features … 1.1 billion sets  Need for a search strategy  Sequential forward selection  Recursive backward elimination  Genetic algorithms  Simulated annealing  …

  11. Wrappers: Sequential Forward Selection

  12. Search Complexity for Sequential Forward Selection

  13. Feature Selection (2): Filters

  14. Search Complexity for Filter Methods Pros :  A lot less expensive! Cons :  Not model-oriented

  15. Feature Selection (3): Embedded methods Principle : the classifier performs feature selection as part of the learning procedure Example : the logistic LASSO (Tibshirani, 1996) With Error Function: Cross-entropy error Regularizing term Pros :  Performs feature selection as part of learning the procedure Cons :  Computationally demanding

  16. Conclusions on Feature Selection Potential benefits Wrappers generally infeasible on the modern “big data” problem. Filters mostly heuristics, but can be formalized in some cases. - Manchester MLO group works on this challenge.

  17. This is the End of the Course Unit…. That’s it. We’re done. Exam in January – past papers on website. MSc students: Projects due Friday, 4pm CDT/MRes students: 1 week later. You need to submit a hardcopy to SSO: - your 6 page (maximum) report You need to send by email to Gavin : - the report as PDF, and a ZIP file of your code.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend