Random Subwindows for Robust Image Classification Rapha el Mar - - PowerPoint PPT Presentation

random subwindows for robust image classification
SMART_READER_LITE
LIVE PREVIEW

Random Subwindows for Robust Image Classification Rapha el Mar - - PowerPoint PPT Presentation

Introduction Our Approach Experiments Conclusions Random Subwindows for Robust Image Classification Rapha el Mar ee, Pierre Geurts, Justus Piater, Louis Wehenkel Institut Montefiore, University of Li` ege, Belgium CVPR05, 22th June


slide-1
SLIDE 1

Introduction Our Approach Experiments Conclusions

Random Subwindows for Robust Image Classification

Rapha¨ el Mar´ ee, Pierre Geurts, Justus Piater, Louis Wehenkel

Institut Montefiore, University of Li` ege, Belgium

CVPR05, 22th June 2005

Mar´ ee et al. Random Subwindows + Extra-Trees (1 / 25)

slide-2
SLIDE 2

Introduction Our Approach Experiments Conclusions Image classification Approaches

Image classification

Given a training set of N labelled images (i.e. each image is associated with a class), build a model to predict the class of new images Challenges

To avoid manual adaptation to specific task To be able to discriminate between a lot of classes To be robust to uncontrolled conditions

Illumination/scale/viewpoint/orientation changes Partial occlusions, cluttered backgrounds . . .

Mar´ ee et al. Random Subwindows + Extra-Trees (2 / 25)

slide-3
SLIDE 3

Introduction Our Approach Experiments Conclusions Image classification Approaches

Approaches

General scheme [MO04]

Detection of “interesting” regions in images [MTS+05]

Harris, Hessian, MSER, edge-based, local variance, . . .

Description by feature vectors [MS05]

SIFT, PCA, DCT, moment invariants, . . .

Matching of feature vectors

Nearest neighbor with Euclidian, Mahalanobis distance, . . .

Mar´ ee et al. Random Subwindows + Extra-Trees (3 / 25)

slide-4
SLIDE 4

Introduction Our Approach Experiments Conclusions Image classification Approaches

Approaches

General scheme [MO04]

Detection of “interesting” regions in images [MTS+05]

Harris, Hessian, MSER, edge-based, local variance, . . . Random extraction of square patches

Description by feature vectors [MS05]

SIFT, PCA, DCT, moment invariants, . . . Pixel-based normalized representation

Matching of feature vectors

Nearest neighbor with Euclidian, Mahalanobis distance, . . . Recent machine learning algorithms able to handle high-dimensional data, e.g.: Ensemble of Decision Trees, SVMs

Mar´ ee et al. Random Subwindows + Extra-Trees (3 / 25)

slide-5
SLIDE 5

Introduction Our Approach Experiments Conclusions Random Subwindows Extra-Trees

Detector: Random Subwindows

✁ ✂✄ ☎✆☎ ☎✆☎ ✝✆✝ ✝✆✝ ✞✟ ✠✡✠✡✠✡✠✡✠ ✠✡✠✡✠✡✠✡✠ ✠✡✠✡✠✡✠✡✠ ✠✡✠✡✠✡✠✡✠ ☛✡☛✡☛✡☛✡☛ ☛✡☛✡☛✡☛✡☛ ☛✡☛✡☛✡☛✡☛ ☛✡☛✡☛✡☛✡☛

Extract Subwindows of random sizes, at random locations

Mar´ ee et al. Random Subwindows + Extra-Trees (4 / 25)

slide-6
SLIDE 6

Introduction Our Approach Experiments Conclusions Random Subwindows Extra-Trees

Descriptor: 16x16 Hue-Saturation-Value

✁ ✁ ✁ ✂✁✂ ✂✁✂ ✂✁✂ ✄☎ ✆✝✆✝✆ ✆✝✆✝✆ ✞✝✞✝✞ ✞✝✞✝✞ ✟✁✟ ✟✁✟ ✠✁✠ ✠✁✠ ✡☛

16x16

768 1 768 1 768 1 768 1 768 1

☞✁☞ ☞✁☞ ☞✁☞ ☞✁☞ ✌✁✌ ✌✁✌ ✌✁✌ ✍✎✏ ✏ ✑ ✑ ✒✁✒ ✒✁✒ ✓✁✓ ✓✁✓ ✔✝✔✝✔ ✔✝✔✝✔ ✕✝✕✝✕ ✕✝✕✝✕

: : : : :

✖ ✖ ✖✗ ✗ ✘✙ ✚✁✚ ✚✁✚ ✛✁✛ ✛✁✛ ✜✢ ✣✝✣✝✣✝✣✝✣ ✣✝✣✝✣✝✣✝✣ ✣✝✣✝✣✝✣✝✣ ✣✝✣✝✣✝✣✝✣ ✤✝✤✝✤✝✤✝✤ ✤✝✤✝✤✝✤✝✤ ✤✝✤✝✤✝✤✝✤ ✤✝✤✝✤✝✤✝✤

Resize each subwindow to 16 × 16 Describe each subwindow by its 768 pixel values (in HSV)

Mar´ ee et al. Random Subwindows + Extra-Trees (5 / 25)

slide-7
SLIDE 7

Introduction Our Approach Experiments Conclusions Random Subwindows Extra-Trees

Learning: subwindow classification model

Extract Nw (>>N) subwindows from training images

Random detector, 16x16 HSV descriptor Label each subwindow with the class of its parent image

✁ ✂ ✂✄ ✄ ☎✆ ✝✞✝✞✝✞✝ ✝✞✝✞✝✞✝ ✝✞✝✞✝✞✝ ✟✞✟✞✟ ✟✞✟✞✟ ✟✞✟✞✟ ✠✞✠✞✠ ✠✞✠✞✠ ✠✞✠✞✠ ✠✞✠✞✠ ✠✞✠✞✠ ✡✞✡✞✡ ✡✞✡✞✡ ✡✞✡✞✡ ✡✞✡✞✡ ✡✞✡✞✡ ☛☞ ✌✞✌✞✌ ✌✞✌✞✌ ✍✞✍✞✍ ✍✞✍✞✍ ✎✞✎ ✎✞✎ ✎✞✎ ✏✞✏ ✏✞✏ ✏✞✏ ✑✒✑ ✑✒✑ ✓✒✓ ✓✒✓ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✔ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✕ ✖ ✖ ✖ ✖ ✖ ✖ ✗ ✗ ✗ ✗ ✘✞✘✞✘✞✘ ✘✞✘✞✘✞✘ ✘✞✘✞✘✞✘ ✘✞✘✞✘✞✘ ✘✞✘✞✘✞✘ ✙✞✙✞✙ ✙✞✙✞✙ ✙✞✙✞✙ ✙✞✙✞✙ ✙✞✙✞✙ ✚ ✚✛ ✛ ✜✞✜✞✜ ✜✞✜✞✜ ✢✞✢ ✢✞✢ ✣✤✣ ✣✤✣ ✣✤✣ ✥✤✥ ✥✤✥ ✥✤✥ ✦✞✦ ✦✞✦ ✦✞✦ ✧✞✧ ✧✞✧ ✧✞✧ ★✞★ ★✞★ ★✞★ ✩✞✩ ✩✞✩ ✩✞✩ ✪✞✪ ✪✞✪ ✪✞✪ ✫✞✫ ✫✞✫ ✫✞✫ ✬✞✬ ✬✞✬ ✭✞✭ ✭✞✭ ✮✞✮ ✮✞✮ ✯✞✯ ✯✞✯ ✰✞✰ ✰✞✰ ✰✞✰ ✱✞✱ ✱✞✱ ✱✞✱ ✲✳ ✴✵ ✶✞✶✞✶ ✶✞✶✞✶ ✷✞✷✞✷ ✷✞✷✞✷ ✸ ✸ ✸ ✸ ✸ ✸ ✸ ✸ ✹ ✹ ✹ ✹ ✹ ✹ ✺✞✺✞✺ ✺✞✺✞✺ ✻✞✻✞✻ ✻✞✻✞✻ ✼✞✼✞✼ ✼✞✼✞✼ ✼✞✼✞✼ ✽✞✽✞✽ ✽✞✽✞✽ ✽✞✽✞✽ ✾✞✾✞✾ ✾✞✾✞✾ ✾✞✾✞✾ ✿✞✿ ✿✞✿ ✿✞✿ ❀ ❀ ❀ ❀ ❀ ❀ ❀ ❀ ❁ ❁ ❁ ❁ ❁ ❁ ❂✤❂ ❂✤❂ ❃✤❃ ❃✤❃

C1 C2 C3 C1 C1 C1 C1 C1 C2 C2 C2 C2 C2 C3 C3 C3 C3 C3

Build a subwindow classification model by supervised learning

T2 T1 T3 T4 T5

Mar´ ee et al. Random Subwindows + Extra-Trees (6 / 25)

slide-8
SLIDE 8

Introduction Our Approach Experiments Conclusions Random Subwindows Extra-Trees

Learning: Extra-Trees [Geu02, GEW05]

T2 T1 T3 T4 T5

✁ ✁ ✁ ✂✁✂ ✂✁✂ ✂✁✂ ✄☎✄☎✄ ✄☎✄☎✄ ✄☎✄☎✄ ✆☎✆☎✆ ✆☎✆☎✆ ✆☎✆☎✆ ✝☎✝☎✝ ✝☎✝☎✝ ✞☎✞☎✞ ✞☎✞☎✞ ✟✁✟ ✟✁✟ ✟✁✟ ✠✁✠ ✠✁✠ ✠✁✠ ✡✁✡ ✡✁✡ ✡✁✡ ☛✁☛ ☛✁☛ ☛✁☛ ☞✁☞ ☞✁☞ ✌✁✌ ✌✁✌

18 3 60 60 2 75 1 1 29 17 23 77 2

a a a a a

2 3 4

a

1

a

768 Class

...... ...... ...... ...... ...... ......

767 766

C1 C1 C1 C2 C2 0.1 0.03 0.2 0.1 0.37 180 97 210 113 99 0.23 145 88 255 230 164 0.05 0.07 0.12 0.23 0.06 55 10 10 0.54 100 C3 2

...... ...... ...... ...... ...... ...... ...... ...... ...... ...... ...... N N Y Y

< 31 a C2 C3 a < 0.5

8 3

C1

Ensemble of T decision trees, generated independently Top-down growing by recursive partitioning

Internal test nodes compare a pixel-location-channel to a threshold (ai < vi), terminal nodes output class probability estimates Choice of internal tests at random Fully developed (perfect fit on LS)

Mar´ ee et al. Random Subwindows + Extra-Trees (7 / 25)

slide-9
SLIDE 9

Introduction Our Approach Experiments Conclusions Random Subwindows Extra-Trees

Recognition: aggregation of subwindows and tree votes

T2 T1 T3 T4 T5

C1 C2 CM 0 0 0 1 4 C1 C2 CM 0 0 0 0 0 0 0 1 4 C1 C2 CM

✁✁ ✁✁ ✁✁ ✂✁✂✁✂ ✂✁✂✁✂ ✂✁✂✁✂ ✄☎✄ ✄☎✄ ✆ ✆ ✝✁✝ ✝✁✝ ✝✁✝ ✞✁✞ ✞✁✞ ✞✁✞ ✟✁✟✁✟ ✟✁✟✁✟ ✠✁✠✁✠ ✠✁✠✁✠ ✡☛✡ ☞☛☞ ✌✁✌ ✌✁✌ ✌✁✌ ✌✁✌ ✍✁✍ ✍✁✍ ✍✁✍ ✍✁✍ ✎✁✎✁✎ ✎✁✎✁✎ ✎✁✎✁✎ ✏✁✏✁✏ ✏✁✏✁✏ ✏✁✏✁✏ ✑✁✑ ✑✁✑ ✑✁✑ ✑✁✑ ✒✁✒ ✒✁✒ ✒✁✒ ✒✁✒ ✓✁✓ ✓✁✓ ✓✁✓ ✔✁✔ ✔✁✔ ✔✁✔ ✕✁✕ ✕✁✕ ✕✁✕ ✖✁✖ ✖✁✖ ✖✁✖ ✗ ✗✘ ✘ ✙✁✙✁✙ ✙✁✙✁✙ ✙✁✙✁✙ ✚✁✚ ✚✁✚ ✚✁✚ ✛☎✛☎✛ ✛☎✛☎✛ ✛☎✛☎✛ ✜☎✜☎✜ ✜☎✜☎✜ ✜☎✜☎✜ ✢✁✢ ✢✁✢ ✢✁✢ ✢✁✢ ✣✁✣ ✣✁✣ ✣✁✣

? ? ? ? ? ? ?

+ =

3 5 10 4 6 1 5 2 49 1

C2 Mar´ ee et al. Random Subwindows + Extra-Trees (8 / 25)

slide-10
SLIDE 10

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Experiments

Standard classification datasets (4 in the paper + 4)

Multi-class (up to 201 classes) Illumination/scale/viewpoint changes, partial occlusions, cluttered backgrounds

Standard protocols

Independent test set or leave-one-out validation Directly comparable to other results in the literature

Parameters

Number of learning subwindows: Nw = 120000 (total) Number of trees built: T = 10 Number of test subwindows: Nw,test = 100 (per image)

Mar´ ee et al. Random Subwindows + Extra-Trees (9 / 25)

slide-11
SLIDE 11

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: COIL-100 [MN95] (100 classes)

Mar´ ee et al. Random Subwindows + Extra-Trees (10 / 25)

slide-12
SLIDE 12

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: ETH-80 [LS03] (8 classes)

Mar´ ee et al. Random Subwindows + Extra-Trees (11 / 25)

slide-13
SLIDE 13

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: ZuBuD [SSV03] (201 classes)

Mar´ ee et al. Random Subwindows + Extra-Trees (12 / 25)

slide-14
SLIDE 14

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: WANG [CW04] (10 classes)

Mar´ ee et al. Random Subwindows + Extra-Trees (13 / 25)

slide-15
SLIDE 15

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: MNIST [LBBH98] (10 classes)

Mar´ ee et al. Random Subwindows + Extra-Trees (14 / 25)

slide-16
SLIDE 16

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: AR Expression Variant Faces [MB98] (100 classes)

Learning: Session 1: Session 2:

Mar´ ee et al. Random Subwindows + Extra-Trees (15 / 25)

slide-17
SLIDE 17

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: TSG-20 [FSPB05] (20 classes)

Mar´ ee et al. Random Subwindows + Extra-Trees (16 / 25)

slide-18
SLIDE 18

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Datasets: IRMA [LGD+05] [iCS05] (57 classes)

(ImageCLEF 2005 [iCS05]) (courtesy of TM Lehmann, Dept. of Medical Informatics, RWTH Aachen, Germany) Mar´ ee et al. Random Subwindows + Extra-Trees (17 / 25)

slide-19
SLIDE 19

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Results: Misclassification error rates

DB ls/ts class us worst best COIL-100 1800/5400 100 0.50% 12.50% 0.10% [MO04] COIL-100 100/7100 100 13.58% 50% 24% [MO04] ZuBuD 1005/115 201 4.35% 59% 0% [MO04] ETH-80 3280/3280 8 25.49% 35.15% 13.60% [LS03] WANG 1000/1000 10 15.90% 62.5% 15.90% [DKN04a] MNIST 60000/10000 10 2.13% 12% 0.50% [DKN04b] AR EVF 100/600 100 15.83% 29.83% 12% [TCZ+05] TSG-20 40/40 20 5.0% 2.5% 0% [FSPB05] IRMA 9000/1000 57 14.7% 73.3% 12.6% [iCS05]

Mar´ ee et al. Random Subwindows + Extra-Trees (18 / 25)

slide-20
SLIDE 20

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

COIL-100: robustness to viewpoint changes

0% 10% 20% 30% 40% −180° −135° −90° −45° 0° 45° 90° 135° 180° error rate test image azimuthal angle Extra−Trees + Random Subwindows

COIL-100: error rate depending on azimuthal test angle, learning

  • nly from the frontal view (0◦).

Mar´ ee et al. Random Subwindows + Extra-Trees (19 / 25)

slide-21
SLIDE 21

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Some observations: subwindow classification

correct: misclassified:

Mar´ ee et al. Random Subwindows + Extra-Trees (20 / 25)

slide-22
SLIDE 22

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Robustness to orientation changes

C1 C2 C3 C1 C1 C1 C1 C1 C2 C2 C2 C2 C2 C3 C3 C3 C3 C3

Mar´ ee et al. Random Subwindows + Extra-Trees (21 / 25)

slide-23
SLIDE 23

Introduction Our Approach Experiments Conclusions Methodology Datasets Results

Why does it work?

Random Subwindows

Aggregation of a large amount of information

Use both local, global, (un)homogeneous regions, . . .

Pixel-based normalized representation

Normalization to a fixed size HSV limits the effect of illumination changes

Tolerance to partial occlusions and cluttered backgrounds

Extra-trees

Accurate even with high-dimensional data (variance reduction)

Mar´ ee et al. Random Subwindows + Extra-Trees (22 / 25)

slide-24
SLIDE 24

Introduction Our Approach Experiments Conclusions Summary Future work

Summary

Novel image classification method that...

combines Random Subwindows and Extra-Trees yields quite good results on a variety of tasks could be quickly evaluated on new classification problems

few parameters (the more trees/subwindows, the better) fast learning (± 6m30s on ZuBuD) fast classification (tree depth ± 18.26 on ZuBuD)

is now implemented in Java: http://www.montefiore.ulg.ac.be/~maree/

Mar´ ee et al. Random Subwindows + Extra-Trees (23 / 25)

slide-25
SLIDE 25

Introduction Our Approach Experiments Conclusions Summary Future work

Extensions and Future Work

Method

Comparison with other detectors and other descriptors Comparison with other machine learning algorithms

CART, Bagging, Boosting, Random Forests: [MGPW05] KNN, SVM

Filtering Subwindows for heavily cluttered backgrounds?

Evaluation

ALOI, Butterflies, Birds, Caltech 101, NORB, . . . , ? Ongoing real-world applications: metal powders, marbles, flowers, license plates, . . .

Mar´ ee et al. Random Subwindows + Extra-Trees (24 / 25)

slide-26
SLIDE 26

Introduction Our Approach Experiments Conclusions Summary Future work

Acknowledgments

Rapha¨ el Mar´ ee is supported by GIGA-Interdisciplinary Cluster for Applied Genoproteomics, hosted by the University of Li` ege Pierre Geurts is a Postdoctoral Researcher at the National Fund for Scientific Research (FNRS, Belgium) IRMA database courtesy of TM Lehmann, Dept. of Medical Informatics, RWTH Aachen, Germany PEPITe for the release of PiXiT, a Java implementation of the method, available for evaluation purpose at: http://www.montefiore.ulg.ac.be/~maree/

Mar´ ee et al. Random Subwindows + Extra-Trees (25 / 25)

slide-27
SLIDE 27

Introduction Our Approach Experiments Conclusions Summary Future work

  • Y. Chen and J. Z. Wang.

Image categorization by learning and reasoning with regions. Journal of Machine Learning Research, 5:913–939, August 2004.

  • T. Deselaers, D. Keysers, and H. Ney.

Features for image retrieval: A quantitative comparison. In Proc. 26th DAGM Symposium on Pattern Recognition (DAGM 2004), volume LNCS 3175, pages 228–236, August/September 2004.

  • T. Deselaers, D. Keysers, and H. Ney.

Classification error rate for quantitative evaluation of content-based image retrieval systems. In Proc. 17th International Conference on Pattern Recognition (ICPR), 2004.

  • G. Fritz, C. Seifert, L. Paletta, and H. Bischof.

Learning informative sift descriptors for attentive object recognition. In Proc. 1st Austrian Cognitive Vision Workshop (ACVW 2005), 2005.

  • P. Geurts.

Contributions to decision tree induction: bias/variance tradeoff and time series classification. PhD thesis, Department of Electrical Engineering and Computer Science, University of Li` ege, May 2002.

  • P. Geurts, D. Ernst, and L. Wehenkel.

Extremely randomized trees. To appear in Machine Learning Journal, 2005.

  • S. L. N. in Computer Science, editor.
  • Proc. of Cross Language Evaluation Forum (CLEF), to appear, 2005.
  • Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner.

Gradient-based learning applied to document recognition.

  • Proc. of the IEEE, 86(11):2278–2324, 1998.

Mar´ ee et al. Random Subwindows + Extra-Trees (25 / 25)

slide-28
SLIDE 28

Introduction Our Approach Experiments Conclusions Summary Future work

  • T. Lehmann, M. G¨

uld, T. Deselaers, D. Keysers, H. Schubert, K. Spitzer, H. Ney, and B. Wein. Automatic categorization of medical images for content-based retrieval and data mining. Computerized Medical Imaging and Graphics, 29(2):143–155, 2005.

  • B. Leibe and B. Schiele.

Analyzing appearance and contour based methods for object categorization. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR’03), Madison, WI, June 2003.

  • A. Martinez and R. Benavente.

The ar face database. Technical report, School of Electrical & Computer Engineering Purdue University, West Lafayette, Indiana, 1998.

  • R. Mar´

ee, P. Geurts, J. Piater, and L. Wehenkel. Decision trees and random subwindows for object recognition. In ICML workshop on Machine Learning Techniques for Processing Multimedia Content (MLMM2005), 2005.

  • H. Murase and S. K. Nayar.

Visual learning and recognition of 3d objects from appearance. International Journal of Computer Vision, 14(1):5–24, 1995.

  • J. Matas and S. Obdr˘

z´ alek. Object recognition methods based on transformation covariant features. In Proc. 12th European Signal Processing Conference (EUSIPCO 2004), Vienna, Austria, September 2004.

  • K. Mikolajczyk and C. Schmid.

A performance evaluation of local descriptors. PAMI, to appear, 2005. Mar´ ee et al. Random Subwindows + Extra-Trees (25 / 25)

slide-29
SLIDE 29

Introduction Our Approach Experiments Conclusions Summary Future work

  • K. Mikolajczyk, T. Tuytelaars, C. Schmid, A. Zisserman, J. Matas, F. Schaffalitzky, T. Kadir, and L. V.

Gool. A comparison of affine region detectors. International Journal of Computer Vision, to appear, 2005.

  • H. Shao, T. Svoboda, and L. Van Gool.

Zubud - Zurich building database for image based recognition. Technical Report TR-260, Computer Vision Lab, Swiss Federal Institute of Technology, Switzerland, 2003.

  • X. Tan, S. Chen, Z.-H. Zhou, , and F. Zhang.

Recognizing partially occluded, expression variant faces from single training image per person with som and soft knn ensemble. IEEE Transactions on Neural Networks, 2005. Mar´ ee et al. Random Subwindows + Extra-Trees (25 / 25)