announcements
play

Announcements Recognition wrap-up Assignment 1 due Sept 22 11:59 pm - PDF document

9/19/2017 Announcements Recognition wrap-up Assignment 1 due Sept 22 11:59 pm on Canvas & Hw2 is out and due Wed Oct 11 Self-supervised representation learning Next week: CNN hands-on tutorial with Ruohan Gao and Tushar


  1. 9/19/2017 Announcements Recognition wrap-up • Assignment 1 due Sept 22 11:59 pm on Canvas & • Hw2 is out and due Wed Oct 11 Self-supervised representation learning • Next week: CNN hands-on tutorial with Ruohan Gao and Tushar Nagarajan • Bring laptop Kristen Grauman • Set up your TACC portal account in advance UT-Austin Wed Sept 20, 2017 Last time: Three landmark case Outline studies for image classification • Last time • Spatial verification for instance recognition • Recognizing categories • Today • Wrap up on categories/classifiers • Self-supervised learning • External papers & assigned paper discussion Boosting + face SVM + person NN + scene Gist • Shuffle and Learn (Yu-Chuan) detection detection classification • Colorization (Keivaun) • Curious Robot (Ginevra) Viola & Jones e.g., Hays & Efros e.g., Dalal & Triggs • Experiment • Network dissection (Thomas and Wonjoon) Slide credit: Kristen Grauman Linear classifiers Last time • Intro to categorization problem • Object categorization as discriminative classification • Boosting + fast face detection example • Nearest neighbors + scene recognition example • Support vector machines + pedestrian detection example • Pyramid match kernels, spatial pyramid match • Convolutional neural networks + ImageNet example 1

  2. 9/19/2017 Linear classifiers Support Vector Machines (SVMs) • Find linear function to separate positive and negative examples • Discriminative    x positive : x w b 0 classifier based on i i    x negative : x w b 0 optimal separating i i hyperplane • Maximize the margin between the positive and negative training Which line examples is best? Support vector machines Support vector machines • Want line that maximizes the margin. • Want line that maximizes the margin.         x positive ( y 1) : x w b 1 x positive ( y 1) : x w b 1 i i i i i i             x negative ( y 1) : x w b 1 x negative ( y 1) : x w b 1 i i i i i i x i w   b   1 x i w   b   1 For support, vectors, For support, vectors,   | x w b | Distance between point i and line: || w || For support vectors: w Τ x  b  1  1 1 2     M Support vectors Support vectors w w Margin M w w w Margin C. Burges, A Tutorial on Support Vector Machines for Pattern Recognition, Data Mining and Knowledge Discovery, 1998 Support vector machines Finding the maximum margin line • Want line that maximizes the margin. 1. Maximize margin 2/|| w || 2. Correctly classify all training data points:         x positive ( y 1) : x w b 1 x positive ( y 1) : x w b 1 i i i i i i       x negative ( y 1) : x w b 1       x negative ( y 1) : x w b 1 i i i i i i     x i w b 1 For support, vectors, Quadratic optimization problem :   | x w b | Distance between point i 1 and line: || w || w T Minimize w 2 Therefore, the margin is 2 / || w || Subject to y i ( w · x i + b ) ≥ 1 Support vectors Margin M 2

  3. 9/19/2017 Finding the maximum margin line Finding the maximum margin line       w i y x w i y x • Solution: • Solution: i i i i i i b = y i – w · x i (for any support vector)        w x b y x x b learned Support i i i i weight vector • Classification function:    f ( x ) sign ( w x b)        sign y x x b i i i i C. Burges, A Tutorial on Support Vector Machines for Pattern Recognition, Data Mining and Knowledge Discovery, 1998 Person detection Non-linear SVMs with HoG’s & linear SVM’s  Datasets that are linearly separable with some noise work out great: • Map each grid cell in the x 0 input window to a histogram  But what are we going to do if the dataset is just too hard? counting the gradients per x 0 orientation.  How about … mapping data to a higher-dimensional space: • Train a linear SVM using x 2 training set of pedestrian vs. non-pedestrian windows. Dalal & Triggs, CVPR 2005 0 Code available: x http://pascal.inrialpes.fr/soft/olt/ Nonlinear SVMs Example • The kernel trick : instead of explicitly computing 2-dimensional vectors x=[ x 1 x 2 ]; the lifting transformation φ ( x ), define a kernel let K (x i ,x j )=(1 + x i T x j ) 2 function K such that K ( x i , x j j ) = φ ( x i ) · φ ( x j ) Need to show that K (x i ,x j )= φ(x i ) T φ(x j ): K (x i ,x j )=(1 + x i T x j ) 2 , • This gives a nonlinear decision boundary in the = 1+ x i1 2 x j1 2 + 2 x i1 x j1 x i2 x j2 + x i2 2 x j2 2 + 2 x i1 x j1 + 2 x i2 x j2 original feature space: = [1 x i1 2 √ 2 x i1 x i2 x i2 2 √ 2 x i1 √ 2 x i2 ] T    y K ( x , x ) b i i i [1 x j1 2 √ 2 x j1 x j2 x j2 2 √ 2 x j1 √ 2 x j2 ] i = φ(x i ) T φ(x j ), where φ(x) = [1 x 1 2 √ 2 x 1 x 2 x 2 2 √ 2 x 1 √ 2 x 2 ] 3

  4. 9/19/2017 Examples of kernel functions SVMs for recognition  Linear:  T K ( x , x ) x x 1. Define your representation for each i j i j example. 2. Select a kernel function. 2  x x i j 3. Compute pairwise kernel values    Gaussian RBF: K ( x ,x ) exp( ) i j  2 2 between labeled examples 4. Use this “kernel matrix” to solve for SVM support vectors & weights.  Histogram intersection: 5. To classify a new example: compute   K ( x , x ) min( x ( k ), x ( k )) kernel values between new input i j i j and support vectors, apply weights, k check sign of output. Kristen Grauman Partially matching sets of features What about a matching kernel? Optimal match: O(m 3 ) Greedy match: O(m 2 log m) Pyramid match: O(m) ( m =num pts) We introduce an approximate matching kernel that makes it practical to compare large sets of features based on their partial correspondences. Local feature correspondence useful similarity [Previous work: Indyk & Thaper, Bartal, Charikar, Agarwal & measure for generic object categories Varadarajan, …] Kristen Grauman Kristen Grauman Pyramid match: main idea Pyramid match: main idea Feature space partitions serve to “match” the local descriptors within successively wider regions. descriptor space Histogram intersection counts number of possible matches at a given partitioning. Kristen Grauman Kristen Grauman 4

  5. 9/19/2017 Pyramid match kernel Pyramid match kernel Optimal match: O(m 3 ) Pyramid match: O(mL) measures number of newly matched difficulty of a pairs at level match at level optimal partial • For similarity, weights inversely proportional to bin size matching (or may be learned) • Normalize these kernel values to avoid favoring large sets [Grauman & Darrell, ICCV 2005] Kristen Grauman Unordered sets of local features: Spatial pyramid match No spatial layout preserved! • Make a pyramid of bag-of-words histograms. • Provides some loose (global) spatial layout information Too much? Too little? [Lazebnik, Schmid & Ponce, CVPR 2006] Spatial pyramid match Spatial pyramid match • Can capture scene categories well---texture-like patterns • Make a pyramid of bag-of-words histograms. but with some variability in the positions of all the local • Provides some loose (global) spatial layout pieces. information Sum over PMKs computed in image coordinate space, one per word. [Lazebnik, Schmid & Ponce, CVPR 2006] 5

  6. 9/19/2017 SVMs: Pros and cons Spatial pyramid match • Pros • Can capture scene categories well---texture-like patterns • Kernel-based framework is very powerful, flexible but with some variability in the positions of all the local • Often a sparse set of support vectors – compact at test time pieces. • Work very well in practice, even with very small training • Sensitive to global shifts of the view sample sizes • Cons • No “direct” multi-class SVM, must combine two-class SVMs • Can be tricky to select best kernel function for a problem • Computation, memory – During training time, must compute matrix of kernel values for every pair of examples – Learning can take a very long time for large-scale problems Confusion table Adapted from Lana Lazebnik Traditional Image Categorization: Recall: Evolution of methods Training phase Training Training Training Images Labels • Hand-crafted models • Hand-crafted features • “End-to-end” learning of Image Classifier Trained • 3D geometry • Learned models Features Training Classifier features and • Hypothesize and align • Data-driven models*,** Slide credit: Jia-Bin Huang Traditional Image Categorization: Learning a Hierarchy of Feature Extractors Testing phase • Each layer of hierarchy extracts features from output Training Training Training Images Labels of previous layer • All the way from pixels  classifier Image Classifier Trained • Layers have the (nearly) same structure Features Training Classifier Labels Image/video Image/Video Simple Testing Pixels Layer 1 Layer 1 Layer 2 Layer 2 Layer 3 Layer 3 Classifier Prediction Image Trained • Train all layers jointly Classifier Features Outdoor Test Image Slide credit: Jia-Bin Huang Slide: Rob Fergus 6

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend