feature selection
play

Feature Selection CE-725: Statistical Pattern Recognition Sharif - PowerPoint PPT Presentation

Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Soleymani Fall 2018 Some slides are based on Isabelle Guyons slides : Introduction to feature selection, 2007. Outline } Dimensionality


  1. Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Soleymani Fall 2018 Some slides are based on Isabelle Guyon’s slides : “Introduction to feature selection”, 2007.

  2. Outline } Dimensionality reduction } Filter univariate methods } Multi-variate filter & wrapper methods } Search strategies } Embedded approach 2

  3. Dimensionality reduction: Feature selection vs. feature extraction } Feature selection } Select a subset of a given feature set } Feature extraction (e.g., PCA, LDA) } A linear or non-linear transform on the original feature space 𝑦 & ' 𝑦 " 𝑦 " 𝑧 " 𝑦 " ⋮ ⋮ → ⋮ ⋮ ⋮ → = 𝑔 𝑦 $ 𝑦 & () 𝑦 $ 𝑧 $ ) 𝑦 $ Feature Feature Selection Extraction ( 𝑒 + < 𝑒 ) 3

  4. Feature selection } Data may contain many irrelevant and redundant variables and often comparably few training examples } Consider supervised learning problems where the number of features 𝑒 is very large (perhaps 𝑒 ≫ 𝑜 ) } E.g., datasets with tens or hundreds of thousands of features and (much) smaller number of data samples ( text or document processing, gene expression array analysis) 𝑒 2 3 𝑒 − 1 1 4 𝒚 (") ⋮ ⋮ … ⋮ 𝒚 (5) 𝑗 " 𝑗 < … 𝑗 $ ) 4

  5. Why feature selection? } FS is a way to find more accurate, faster, and easier to understand classifiers. } Performance: enhancing generalization ability } alleviating the effect of the curse of dimensionality } the higher the ratio of the no. of training patterns 𝑂 to the number of free classifier parameters, the better the generalization of the learned classifier } Efficiency: speeding up the learning process } Interpretability: resulting a model that is easier to understand by human (") (") 𝑦 " ⋯ 𝑦 $ 𝑧 (") 𝑗 " , 𝑗 < , … , 𝑗 $ ) 𝒀 = ⋮ ⋱ ⋮ , 𝑍 = ⋮ Feature Selection (5) (5) 𝑧 (5) 𝑦 " ⋯ 𝑦 $ The selected features Supervised feature selection: Given a labeled set of data points, select a subset of features for data representation 5

  6. Noise (or irrelevant) features } Eliminating irrelevant features can decrease the classification error on test data 𝑦 < Noise feature 𝑦 " 𝑦 " SVM Decision Boundary SVM Decision Boundary 6

  7. Drug Screening 𝑂 = 1909 compounds 𝑒 = 139,351 binary features three-dimensional properties of the molecule. Number of features 7 [Weston et al, Bioinformatics, 2002]

  8. Text Filtering Reuters : 21578 news wire, 114 semantic categories. 20 newsgroups : 19997 articles, 20 categories. WebKB : 8282 web pages, 7 categories. Bag-of-words : >100000 features. Top 3 words of some categories: } Alt.atheism : atheism, atheists, morality } Comp.graphics : image, jpeg, graphics } Sci.space : space, nasa, orbit } Soc.religion.christian : god, church, sin } Talk.politics.mideast : israel, armenian, turkish 8 [Bekkerman et al, JMLR, 2003]

  9. Face Male/Female Classification 𝑂 = 1000 training images 𝑒 = 60×85 = 5100 features 100 500 1000 Relief: Simba: [Navot, Bachrach, and Tishby, ICML, 2004] 9

  10. Some definitions } One categorization of feature selection methods: } Univariate method (variable ranking) : considers one variable (feature) at a time. } Multivariate method : considers subsets of features together. } Another categorization: } Filter method : ranks features or feature subsets independent of the classifier as a preprocessing step. } Wrapper method : uses a classifier to evaluate the score of features or feature subsets. } Embedded method : Feature selection is done during the training of a classifier } E.g., Adding a regularization term 𝒙 " in the cost function of linear classifiers 10

  11. Filter: univariate } Univariate filter method } Score each feature 𝑙 based on the 𝑙 -th column of the data matrix and the label vector } Relevance of the feature to predict labels: Can the feature discriminate the patterns of different classes? } Rank features according to their score values and select the ones with the highest scores. } How do you decide how many features k to choose? e.g., using cross- validation to select among the possible values of k } Advantage: computational and statistical scalability 11

  12. � � � Pearson Correlation Criteria (&) − 𝑦 R 𝑧 (&) − 𝑧 5 ∑ 𝑦 R X 𝑑𝑝𝑤(𝑌 R , 𝑍) &Y" 𝑆 𝑙 = ≈ 𝑤𝑏𝑠(𝑌 R ) 𝑤𝑏𝑠(𝑍) < & − 𝑦 R 𝑧 & − 𝑧 5 5 ∑ ∑ X < 𝑦 R &Y" &Y" 𝑦 < 𝑆(1) ≫ 𝑆(2) 𝑦 " } can only detect linear dependencies between feature and target. 12

  13. µ + µ - Single Variable Classifier ROC Curve x 2 density Y=1 Y=-1 x 1 𝑦 < x Sensitivity=TPR s + 2 s - µ + µ - Y=1 density Area Under Curve Y=-1 𝑦 " (AUC) 1- Specificity=FPR s + x 1 s - 13

  14. Univariate Mutual Information } Independence: 𝑄(𝑌, 𝑍) = 𝑄(𝑌)𝑄(𝑍) } Mutual information as a measure of dependence: 𝑁𝐽 𝑌, 𝑍 = 𝐹 ^,_ log 𝑄(𝑌, 𝑍) 𝑄 𝑌 𝑄(𝑍) } Score of 𝑌 R based on MI with 𝑍 : } 𝐽 𝑙 = 𝑁𝐽 𝑌 R , 𝑍 14

  15. Filter – univariate: Disadvantage } Univariate methods may fail: } a feature may be important in combination with other features. } Redundant features: } They can select a group of dependent variables that carry similar information about the output, i.e. it is sufficient to use only one (or a few) of these variables. 15

  16. Univariate methods: Failure } Samples on which univariate feature analysis and scoring fails: 16 [Guyon-Elisseeff, JMLR 2004; Springer 2006]

  17. Redundant features What is the relation between redundancy and correlation: Are highly correlated features necessarily redundant? What about completely correlated ones? 17

  18. Multi-variate feature selection } Search in the space of all possible combinations of features. } all feature subsets: For 𝑒 features, 2 $ possible subsets. } high computational and statistical complexity. 18

  19. Search space for feature selection ( 𝑒 = 4 ) 0,0,0,0 1,0,0,0 0,0,1,0 0,1,0,0 0,0,0,1 1,1,0,0 0,1,1,0 0,1,0,1 1,0,1,0 1,0,0,1 0,0,1,1 1,1,1,0 1,0,1,1 1,1,0,1 0,1,1,1 1,1,1,1 [Kohavi-John,1997] 19

  20. Multivariate methods: General procedure Original Subset feature set Subset Subset generation evaluation No Yes Stopping Validation criterion Subset Generation: select a candidate feature subset for evaluation Subset Evaluation: compute the score (relevancy value) of the subset Stopping criterion: when stopping the search in the space of feature subsets Validation: verify that the selected subset is valid 20

  21. Stopping criteria } Predefined number of features is selected } Predefined number of iterations is reached } Addition (or deletion) of any feature does not result in a better subset } An optimal subset (according to the evaluation criterion) is obtained. 21

  22. Filter and wrapper methods } Wrappers use the classifier performance to evaluate the feature subset utilized in the classifier. } Training 2 $ classifiers is infeasible for large 𝑒 . } Most wrapper algorithms use a heuristic search. } Filters use an evaluation function that is cheaper to compute than the performance of the classifier } e.g. correlation coefficient 22

  23. Filters vs. wrappers rank subsets of useful features Feature Original feature set Classifier Filter subset Multiple Original feature set Classifier feature subsets Wrapper take classifier into account to rank feature subsets (e.g., using cross validation to evaluate features) 23

  24. Wrapper methods: Performance assessment } For each feature subset, train classifier on training data and assess its performance using evaluation techniques like cross-validation 24

  25. Filter methods: Evaluation criteria } Distance (Eulidean distance) } Class separability: Features supporting instances of the same class to be closer in terms of distance than those from different classes } Information (Information Gain) } Select S1 if IG(S1,Y)>IG(S2,Y) } Dependency (correlation coefficient) } good feature subsets contain features highly correlated with the class, yet uncorrelated with each other } Consistency (min-features bias) } Selects features that guarantee no inconsistency in data } inconsistent instances have the same feature vector but different class labels } Prefers smaller subset with consistency (min-feature) f 1 f 2 class instance 1 a b c1 25 inconsistent instance 2 a b c2

  26. How to search the space of feature subsets? } NP-hard problem. } Complete search is possible only for small number of features. } Greedy search is often used ( forward selection or backward elimination ). 26

  27. Subset selection or generation } Search direction } Forward } Backward } Random } Search strategies } Exhaustive - Complete } Branch & Bound } Best first search } Heuristic or greedy } Sequential forward selection } Sequential backward elimination } Plus-l Minus-r Selection } Bidirectional Search } Sequential floating Selection } Non-deterministic } Simulated annealing } Genetic algorithm 27

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend