1
play

1 Classification by image fragments Cross-generalization = Let - PDF document

Motivations Object classification methods generally require large number of training examples. Cross-generalization: Constructing training set is very time-consuming task, and it learning novel classes from a costs a lot. (time == money)


  1. Motivations � Object classification methods generally require large number of training examples. Cross-generalization: � Constructing training set is very time-consuming task, and it learning novel classes from a costs a lot. (time == money) single example by feature replacement � Do we need many training examples for ALL classes? There exist a lot of classes in real world, and it is not possible to collect training images for all of them. Presented by: Goo Jun � Humans can distinguish >>10k classes CS395T Object Recognition Spring 2007 Overview Introduction Hypothesis : A feature is likely to be useful for a novel class (e.g. dogs) � Develop a feature-based object classification model that can � if a similar feature proved effective for a similar class (e.g. cows). learn a novel class from a single training example. Assume a sufficient number of training examples are available for a set � of object classes (say, cows, horses, and flowers) to extract suitable � Experience from already learned classes facilitate the learning discriminating features. and constrain overfitting. These classes are referred to as “known” or “familiar” classes. � � Features for a novel class are obtained by adapting the features The objective is to learn a new class, say dog, from a single example. from similar familiar (already learned) classes. � • Challenge : Obtain suitable features (restrict overfitting). • Proposed solution : Adapt the features from similar familiar classes. � Cross-generalized model outperforms the stand-alone algorithm on a large dataset with 107 classes of objects. Related Works Using Image Fragments as Features Feature Extraction � Data Manufacturing � � Extract sub-images of multiple sizes from multiple locations � Pros : Can significantly improve classification when generative model is available. � With each fragment, its location in the original image is stored and used to � Cons : Constructing generative models which reflect natural variations of determine relative locations of different fragments visual objects is very difficult. � Features are selected in a greedy manner that maximize the mutual information between the feature and the class it represents. Perona et al. proposed a parametric class model and obtains a prior for � parameters for a novel class based on the examples of familiar class. For Classification � Pros : Avoid inaccurate parameter estimate and increases performance � compared to no prior. � Set of fragments is searched for in the image, using the absolute value of � Cons : A single prior is used for all novel classes, hence biases novel class normalized cross-correlation. parameters towards frequently appearing familiar class. � For each fragment F , the relevant locations in the image are determined by the location of F relative to the common reference frame. Freeman et al . proposed feature sharing between classes. � � Image patches at the relevant locations are compared with F , and the location � Pros : Reduces the total number of representative features. with the highest correlation is selected. If this highest correlation exceeds a � Cons : Produces generic features (like edges) and requires simultaneous training of all classes. pre-determined threshold , the fragment is considered present, or active in θ F the image Proposed algorithm obtains class-specific features and novel class features are adapted only from simila similar familiar amiliar class features. [ 2 ] E.Bart, E.Byvatov, and S. Ullman. “View-invariant recognition using corresponding object fragments.” In ECCV, pages 113-127, 2002. [17] S.Ullman, M.Vidal-Naquet, and E.Sali. “Visual features of intermediate complexity and their use in classification.” Nature Neuroscience, 5(7):682-687, 202. 1

  2. Classification by image fragments Cross-generalization = Let indicate whether f 1 or 0 exist or not in the test image. F A feature F is likely to be useful for class C if a similar feature F’ proved � � i i effective for a similar class C’. The classifier labels the image as belonging to the class C if � For an example image E , each familiar fragment is searched within E using k � F i Normalized cross-correlation. = ( | ) ∑ P f d C > = ∈ ( ) where ( ) log i , { 0 , 1 } w f T w d d i i i = i The location is selected with maximum cross-correlation and a fragment P ( f d | C ) k new � S F i i is extracted from the same location of image E . i = represents the probability that fragment is present in � P ( f d | C ) F i images belonging to class k. The process continues to choose J new fragments from E that corresponds to J � highest cross-correlation among all familiar fragments . k F ( J = 25) i It is needed to invoke a new class if an image under testing does not get � The original fragments extracted from E and their relative locations are new F � classified into one of the familiar classes. j used for the design of the new classifier. Should be able to estimate reliable features from only one example of the � Note that features are chosen independently without incorporating any spatial � new class. correlation between them, although familiar classes offer this information. Classification of Novel Class Results (1) Classification results are compared � Each novel fragment was nominated by some fragment to which it was new k � F F j i against a stand-alone algorithm (SA). similar. SA uses two examples for each class new Since and belong to different classes one needs to prevent being F new k � F F j j i and two examples from a non-class for detected in images that contain , so a threshold higher that k k is required. F S i i training. The authors set the threshold to . 1 . 1 k � S i A new classifier that uses the fragments has the form new � F j The tests are performed on a set of � 107 object classes, with each class ∑ new new > new new ∈ w ( f ) T f { 0 , 1 } j j j j having 40 to 100 examples, along with k = = a set of 400 non-class images. ( | ) p f d c k ∑ > = ∈ k k k k i w ( f ) T where w ( d ) log , d { 0 , 1 } i i i i = = k p ( f d | c k ) i The weights cannot be empirically estimates as shown before (one data point to Leave-one-out scheme was used, � � count). The authors suggest where performance on each class as a novel class was tested with training new = k new = k ( 0 ) ( 0 ) ( 1 ) ( 1 ) w w and w w j i j i done on the rest 106 classes. Test images consist of the novel class and a Alternative : new = new = offers poor performance! ( 0 ) 0 ( 1 ) 1 � w and w j j set of non-class images. Results (2) Conclusions � A classification algorithm is proposed that is able to learn a novel class from a single example. � New features are adapted from the similar features of familiar classes. � The algorithm tries to mimic human cognition system. Will not work for a completely new class (having nothing common with Crayfish class Mandolin Class familiar classes). Performance margin = � No negative example is required to learn the new classifier 2(Area-0.5) (although negative examples are used to learn the familiar class discrimination). Hit rate difference = Hit rate CG – Hit rate SA � No spatial correlation is used to obtain features for the novel class. Scope for future research. Averaged over all examples 2

  3. That’s it! Questions? 3

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend