random subwindows for robust image classification
play

Random Subwindows for Robust Image Classification Rapha el Mar - PowerPoint PPT Presentation

Introduction Our Approach Experiments Conclusions Random Subwindows for Robust Image Classification Rapha el Mar ee, Pierre Geurts, Justus Piater, Louis Wehenkel Institut Montefiore, University of Li` ege, Belgium CVPR05, 22th June


  1. Introduction Our Approach Experiments Conclusions Random Subwindows for Robust Image Classification Rapha¨ el Mar´ ee, Pierre Geurts, Justus Piater, Louis Wehenkel Institut Montefiore, University of Li` ege, Belgium CVPR05, 22th June 2005 Mar´ ee et al. Random Subwindows + Extra-Trees (1 / 25)

  2. Introduction Our Approach Image classification Experiments Approaches Conclusions Image classification Given a training set of N labelled images (i.e. each image is associated with a class), build a model to predict the class of new images Challenges To avoid manual adaptation to specific task To be able to discriminate between a lot of classes To be robust to uncontrolled conditions Illumination/scale/viewpoint/orientation changes Partial occlusions, cluttered backgrounds . . . Mar´ ee et al. Random Subwindows + Extra-Trees (2 / 25)

  3. Introduction Our Approach Image classification Experiments Approaches Conclusions Approaches General scheme [MO04] Detection of “interesting” regions in images [MTS + 05] Harris, Hessian, MSER, edge-based, local variance, . . . Description by feature vectors [MS05] SIFT, PCA, DCT, moment invariants, . . . Matching of feature vectors Nearest neighbor with Euclidian, Mahalanobis distance, . . . Mar´ ee et al. Random Subwindows + Extra-Trees (3 / 25)

  4. Introduction Our Approach Image classification Experiments Approaches Conclusions Approaches General scheme [MO04] Detection of “interesting” regions in images [MTS + 05] Harris, Hessian, MSER, edge-based, local variance, . . . Random extraction of square patches Description by feature vectors [MS05] SIFT, PCA, DCT, moment invariants, . . . Pixel-based normalized representation Matching of feature vectors Nearest neighbor with Euclidian, Mahalanobis distance, . . . Recent machine learning algorithms able to handle high-dimensional data, e.g.: Ensemble of Decision Trees, SVMs Mar´ ee et al. Random Subwindows + Extra-Trees (3 / 25)

  5. ✁ � ✞✟ ✂✄ ✁ � � Introduction Our Approach Random Subwindows Experiments Extra-Trees Conclusions Detector: Random Subwindows ✠✡✠✡✠✡✠✡✠ ☛✡☛✡☛✡☛✡☛ ☎✆☎ ✝✆✝ ✠✡✠✡✠✡✠✡✠ ☛✡☛✡☛✡☛✡☛ ✝✆✝ ☎✆☎ ✠✡✠✡✠✡✠✡✠ ☛✡☛✡☛✡☛✡☛ ✠✡✠✡✠✡✠✡✠ ☛✡☛✡☛✡☛✡☛ Extract Subwindows of random sizes, at random locations Mar´ ee et al. Random Subwindows + Extra-Trees (4 / 25)

  6. ✍✎✏ ✄☎ ✜✢ ✘✙ ✗ ✖✗ ✖ ✖ ✑ ✑ ✏ ✡☛ Introduction Our Approach Random Subwindows Experiments Extra-Trees Conclusions Descriptor: 16x16 Hue-Saturation-Value ✣✝✣✝✣✝✣✝✣ ✤✝✤✝✤✝✤✝✤ ✚✁✚ ✛✁✛ ✣✝✣✝✣✝✣✝✣ ✤✝✤✝✤✝✤✝✤ ✛✁✛ ✚✁✚ ✣✝✣✝✣✝✣✝✣ ✤✝✤✝✤✝✤✝✤ ✣✝✣✝✣✝✣✝✣ ✤✝✤✝✤✝✤✝✤ ✆✝✆✝✆ �✁� ✂✁✂ ✠✁✠ ✟✁✟ ✆✝✆✝✆ ✞✝✞✝✞ �✁� ✂✁✂ �✁� ✂✁✂ ✠✁✠ ✟✁✟ ✞✝✞✝✞ 16x16 ☞✁☞ ✌✁✌ ☞✁☞ ☞✁☞ ✌✁✌ ☞✁☞ ✌✁✌ : 1 768 : 1 768 ✒✁✒ ✓✁✓ : ✒✁✒ ✓✁✓ 1 768 ✔✝✔✝✔ ✕✝✕✝✕ : ✔✝✔✝✔ ✕✝✕✝✕ 1 768 : 1 768 Resize each subwindow to 16 × 16 Describe each subwindow by its 768 pixel values (in HSV) Mar´ ee et al. Random Subwindows + Extra-Trees (5 / 25)

  7. ✴✵ ✕ ✗ ✗ ✗ ✖ ✖ ✖ ✖ ✖ ✖ ✕ ✕ ✕ ✕ ✚ ✕ ✕ ✕ ✕ ✕ ✲✳ ✕ ✕ ✕ ✕ ✔ ✔ ✗ ✚✛ ✔ ✹ ✸ ✸ ✸ ✸ ✸ ✸ ✸ ✸ ✹ ✹ ✹ ✹ ✹ ✛ ❀ ❀ ❀ ❀ ❀ ❀ ❀ ❀ ❁ ❁ ❁ ❁ ❁ ✔ ✕ ✔ ✔ ❁ ☎✆ ✄ ✂✄ ✂ ✁ �✁ � ✔ ✔ ☛☞ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ Introduction Our Approach Random Subwindows Experiments Extra-Trees Conclusions Learning: subwindow classification model Extract N w ( >> N ) subwindows from training images Random detector, 16x16 HSV descriptor Label each subwindow with the class of its parent image ✏✞✏ ✡✞✡✞✡ ✠✞✠✞✠ ✘✞✘✞✘✞✘ ✙✞✙✞✙ ✡✞✡✞✡ ✠✞✠✞✠ ✏✞✏ ✎✞✎ ✘✞✘✞✘✞✘ ✙✞✙✞✙ ✡✞✡✞✡ ✠✞✠✞✠ ✘✞✘✞✘✞✘ ✙✞✙✞✙ ✥✤✥ ✣✤✣ ✠✞✠✞✠ ✡✞✡✞✡ ✎✞✎ ✏✞✏ ✘✞✘✞✘✞✘ ✙✞✙✞✙ ✝✞✝✞✝✞✝ ✠✞✠✞✠ ✡✞✡✞✡ ✓✒✓ ✑✒✑ ✘✞✘✞✘✞✘ ✙✞✙✞✙ ✣✤✣ ✥✤✥ ✍✞✍✞✍ ✌✞✌✞✌ ✎✞✎ ✝✞✝✞✝✞✝ ✟✞✟✞✟ ✜✞✜✞✜ ✑✒✑ ✓✒✓ ✥✤✥ ✣✤✣ ✍✞✍✞✍ ✌✞✌✞✌ ✝✞✝✞✝✞✝ ✟✞✟✞✟ ✜✞✜✞✜ ✢✞✢ ✢✞✢ ✟✞✟✞✟ C1 C2 C3 ★✞★ ✪✞✪ ✰✞✰ ✶✞✶✞✶ ★✞★ ✪✞✪ ✰✞✰ ✦✞✦ ✧✞✧ ✩✞✩ ★✞★ ✫✞✫ ✪✞✪ ✭✞✭ ✬✞✬ ✮✞✮ ✯✞✯ ✰✞✰ ✱✞✱ ✷✞✷✞✷ ✶✞✶✞✶ ✺✞✺✞✺ ✻✞✻✞✻ ✼✞✼✞✼ ✽✞✽✞✽ ✾✞✾✞✾ ❂✤❂ ❃✤❃ ✿✞✿ ✦✞✦ ✧✞✧ ✩✞✩ ✫✞✫ ✱✞✱ ✼✞✼✞✼ ✽✞✽✞✽ ✾✞✾✞✾ ✿✞✿ ✧✞✧ ✦✞✦ ✩✞✩ ✫✞✫ ✬✞✬ ✭✞✭ ✯✞✯ ✮✞✮ ✱✞✱ ✷✞✷✞✷ ✺✞✺✞✺ ✻✞✻✞✻ ✽✞✽✞✽ ✼✞✼✞✼ ✾✞✾✞✾ ❂✤❂ ✿✞✿ ❃✤❃ C1 C1 C1 C1 C1 C2 C2 C2 C2 C2 C3 C3 C3 C3 C3 Build a subwindow classification model by supervised learning T1 T2 T3 T4 T5 Mar´ ee et al. Random Subwindows + Extra-Trees (6 / 25)

  8. Introduction Our Approach Random Subwindows Experiments Extra-Trees Conclusions Learning: Extra-Trees [Geu02, GEW05] a a a a a a a 768 Class ...... a < 31 1 2 3 4 766 767 3 ✟✁✟ ✠✁✠ 60 0.1 99 17 1 0.23 164 C1 ...... ✟✁✟ ✠✁✠ ✟✁✟ ✠✁✠ ✡✁✡ ☛✁☛ 60 0.37 113 23 ...... 29 0.07 230 C1 Y N ✡✁✡ ☛✁☛ ✡✁✡ T1 T2 T3 T4 T5 ☛✁☛ ☞✁☞ ✌✁✌ 75 0.03 210 1 77 0.05 255 C1 ...... ✂✁✂ �✁� a < 0.5 C1 ☞✁☞ ✌✁✌ 8 �✁� ✂✁✂ 2 0.1 97 2 ...... 0 0.23 88 C2 ✄☎✄☎✄ �✁� ✆☎✆☎✆ ✂✁✂ ✄☎✄☎✄ ✆☎✆☎✆ 3 0.2 180 18 0 0.12 145 C2 ...... Y N ✄☎✄☎✄ ✆☎✆☎✆ ...... ...... ...... ...... ...... ...... ...... ...... ...... ...... C2 C3 2 0.06 55 10 10 0.54 100 C3 ...... ✞☎✞☎✞ ✝☎✝☎✝ ✞☎✞☎✞ ✝☎✝☎✝ Ensemble of T decision trees, generated independently Top-down growing by recursive partitioning Internal test nodes compare a pixel-location-channel to a threshold ( a i < v i ), terminal nodes output class probability estimates Choice of internal tests at random Fully developed (perfect fit on LS ) Mar´ ee et al. Random Subwindows + Extra-Trees (7 / 25)

  9. ✆ ✗ ✗✘ ✘ ✆ Introduction Our Approach Random Subwindows Experiments Extra-Trees Conclusions Recognition: aggregation of subwindows and tree votes ✄☎✄ ✂✁✂✁✂ �✁�✁� ✄☎✄ �✁�✁� ✂✁✂✁✂ ✟✁✟✁✟ ✠✁✠✁✠ ✡☛✡ ✂✁✂✁✂ �✁�✁� ☞☛☞ ✠✁✠✁✠ ✟✁✟✁✟ ✞✁✞ ✝✁✝ ✌✁✌ ✍✁✍ ✝✁✝ ✞✁✞ ✌✁✌ ✍✁✍ ✌✁✌ ✍✁✍ ✝✁✝ ✌✁✌ ✞✁✞ ✍✁✍ ? ✙✁✙✁✙ ✛☎✛☎✛ ✎✁✎✁✎ ✚✁✚ ✜☎✜☎✜ ✏✁✏✁✏ ✑✁✑ ✢✁✢ ✒✁✒ ✑✁✑ ✙✁✙✁✙ ✛☎✛☎✛ ✢✁✢ ✎✁✎✁✎ ✚✁✚ ✜☎✜☎✜ ✣✁✣ ✏✁✏✁✏ ✒✁✒ ✑✁✑ ✢✁✢ ✣✁✣ ✒✁✒ ✑✁✑ ✢✁✢ ✙✁✙✁✙ ✚✁✚ ✛☎✛☎✛ ✜☎✜☎✜ ✣✁✣ ✏✁✏✁✏ ✎✁✎✁✎ ✒✁✒ ? ? ? ? ? ? T1 T2 T3 T4 T5 ✓✁✓ ✔✁✔ ✓✁✓ ✔✁✔ ✓✁✓ ✔✁✔ 0 4 0 0 0 0 1 0 0 0 C1 C2 CM ✖✁✖ ✕✁✕ ✖✁✖ ✕✁✕ ✖✁✖ ✕✁✕ + 0 4 0 0 0 0 0 0 0 1 C1 C2 CM = 6 49 4 3 2 10 1 1 5 5 C1 C2 CM C2 Mar´ ee et al. Random Subwindows + Extra-Trees (8 / 25)

  10. Introduction Methodology Our Approach Datasets Experiments Results Conclusions Experiments Standard classification datasets (4 in the paper + 4) Multi-class (up to 201 classes) Illumination/scale/viewpoint changes, partial occlusions, cluttered backgrounds Standard protocols Independent test set or leave-one-out validation Directly comparable to other results in the literature Parameters Number of learning subwindows: N w = 120000 (total) Number of trees built: T = 10 Number of test subwindows: N w , test = 100 (per image) Mar´ ee et al. Random Subwindows + Extra-Trees (9 / 25)

  11. Introduction Methodology Our Approach Datasets Experiments Results Conclusions Datasets: COIL-100 [MN95] (100 classes) Mar´ ee et al. Random Subwindows + Extra-Trees (10 / 25)

  12. Introduction Methodology Our Approach Datasets Experiments Results Conclusions Datasets: ETH-80 [LS03] (8 classes) Mar´ ee et al. Random Subwindows + Extra-Trees (11 / 25)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend