Introduction to Pattern Recognition
Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr
CS 551, Spring 2007 c 2007, Selim Aksoy
Introduction to Pattern Recognition Selim Aksoy Department of - - PowerPoint PPT Presentation
Introduction to Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr c CS 551, Spring 2007 2007, Selim Aksoy Human Perception Humans have developed highly sophisticated skills
CS 551, Spring 2007 c 2007, Selim Aksoy
◮ recognizing a face, ◮ understanding spoken words, ◮ reading handwriting, ◮ distinguishing fresh food from its smell.
CS 551, Spring 2007 c 2007, Selim Aksoy 1/35
◮ fingerprint image, ◮ handwritten word, ◮ human face, ◮ speech signal, ◮ DNA sequence, ◮ . . .
◮ observe the environment, ◮ learn to distinguish patterns of interest, ◮ make sound and reasonable decisions about the
CS 551, Spring 2007 c 2007, Selim Aksoy 2/35
CS 551, Spring 2007 c 2007, Selim Aksoy 3/35
Table 1: Example pattern recognition applications.
Problem Domain Application Input Pattern Pattern Classes Document image analysis Optical character recognition Document image Characters, words Document classification Internet search Text document Semantic categories Document classification Junk mail filtering Email Junk/non-junk Multimedia database retrieval Internet search Video clip Video genres Speech recognition Telephone directory assistance Speech waveform Spoken words Natural language processing Information extraction Sentences Parts of speech Biometric recognition Personal identification Face, iris, fingerprint Authorized users for access control Medical Diagnosis Microscopic image Cancerous/healthy cell Military Automatic target recognition Optical or infrared image Target type Industrial automation Printed circuit board inspection Intensity or range image Defective/non-defective product Industrial automation Fruit sorting Images taken on a conveyor belt Grade of quality Remote sensing Forecasting crop yield Multispectral image Land use categories Bioinformatics Sequence analysis DNA sequence Known types of genes Data mining Searching for meaningful patterns Points in multidimensional space Compact and well-separated clusters CS 551, Spring 2007 c 2007, Selim Aksoy 4/35
Figure 1: English handwriting recognition.
CS 551, Spring 2007 c 2007, Selim Aksoy 5/35
Figure 2: Chinese handwriting recognition.
CS 551, Spring 2007 c 2007, Selim Aksoy 6/35
Figure 3: Fingerprint recognition.
CS 551, Spring 2007 c 2007, Selim Aksoy 7/35
Figure 4: Biometric recognition.
CS 551, Spring 2007 c 2007, Selim Aksoy 8/35
Figure 5: Cancer detection and grading using microscopic tissue data.
CS 551, Spring 2007 c 2007, Selim Aksoy 9/35
Figure 6: Cancer detection and grading using microscopic tissue data.
CS 551, Spring 2007 c 2007, Selim Aksoy 10/35
Figure 7: Land cover classification using satellite data.
CS 551, Spring 2007 c 2007, Selim Aksoy 11/35
Figure 8: Building and building group recognition using satellite data.
CS 551, Spring 2007 c 2007, Selim Aksoy 12/35
Figure 9: License plate recognition: US license plates.
CS 551, Spring 2007 c 2007, Selim Aksoy 13/35
Figure 10: Clustering of microarray data.
CS 551, Spring 2007 c 2007, Selim Aksoy 14/35
◮ sea bass, ◮ salmon.
Figure 11: Picture taken from a camera.
CS 551, Spring 2007 c 2007, Selim Aksoy 15/35
◮ length, width, weight, number and shape of fins, tail
◮ lighting conditions, position of fish on the conveyor
◮ capture image → isolate fish → take measurements
CS 551, Spring 2007 c 2007, Selim Aksoy 16/35
CS 551, Spring 2007 c 2007, Selim Aksoy 17/35
Figure 12: Histograms of the length feature for two types of fish in training samples. How can we choose the threshold l∗ to make a reliable decision?
CS 551, Spring 2007 c 2007, Selim Aksoy 18/35
CS 551, Spring 2007 c 2007, Selim Aksoy 19/35
Figure 13: Histograms of the lightness feature for two types of fish in training samples. It looks easier to choose the threshold x∗ but we still cannot make a perfect decision.
CS 551, Spring 2007 c 2007, Selim Aksoy 20/35
◮ Customers who buy salmon will object vigorously if
◮ Customers who buy sea bass will not be unhappy if
CS 551, Spring 2007 c 2007, Selim Aksoy 21/35
◮ lightness: x1 ◮ width: x2
CS 551, Spring 2007 c 2007, Selim Aksoy 22/35
Figure 14: Scatter plot of lightness and width features for training samples. We can draw a decision boundary to divide the feature space into two regions. Does it look better than using only lightness?
CS 551, Spring 2007 c 2007, Selim Aksoy 23/35
◮ Avoid unreliable features. ◮ Be careful about correlations with existing features. ◮ Be careful about measurement costs. ◮ Be careful about noise in the measurements.
CS 551, Spring 2007 c 2007, Selim Aksoy 24/35
Figure 15: We may distinguish training samples perfectly but how can we predict how well we can generalize to unknown samples?
CS 551, Spring 2007 c 2007, Selim Aksoy 25/35
Figure 16: Different criteria lead to different decision boundaries.
CS 551, Spring 2007 c 2007, Selim Aksoy 26/35
Physical environment Data acquisition/sensing Pre−processing Feature extraction Features Classification Post−processing Decision Model learning/estimation Features Feature extraction/selection Pre−processing Training data Model
Figure 17: Object/process diagram of a pattern recognition system.
CS 551, Spring 2007 c 2007, Selim Aksoy 27/35
◮ Measurements of physical variables. ◮ Important issues: bandwidth, resolution, sensitivity,
◮ Removal of noise in data. ◮ Isolation of patterns of interest from the background.
◮ Finding a new representation in terms of features.
CS 551, Spring 2007 c 2007, Selim Aksoy 28/35
◮ Learning a mapping between features and pattern
◮ Using features and learned models to assign a pattern
◮ Evaluation of confidence in decisions. ◮ Exploitation of context to improve performance. ◮ Combination of experts.
CS 551, Spring 2007 c 2007, Selim Aksoy 29/35
model Train classifier Evaluate classifier Collect data features Select Select Figure 18: The design cycle.
◮ Collecting training and testing data. ◮ How can we know when we have adequately large
CS 551, Spring 2007 c 2007, Selim Aksoy 30/35
◮ Domain dependence and prior information. ◮ Computational cost and feasibility. ◮ Discriminative features.
◮ Invariant
◮ Robust features with respect to occlusion, distortion,
CS 551, Spring 2007 c 2007, Selim Aksoy 31/35
◮ Domain dependence and prior information. ◮ Definition of design criteria. ◮ Parametric vs. non-parametric models. ◮ Handling of missing features. ◮ Computational complexity. ◮ Types of models:
◮ How can we know how close we are to the true
CS 551, Spring 2007 c 2007, Selim Aksoy 32/35
◮ How can we learn the rule from data? ◮ Supervised learning: a teacher provides a category
◮ Unsupervised learning: the system forms clusters or
◮ Reinforcement learning: no desired category is given
CS 551, Spring 2007 c 2007, Selim Aksoy 33/35
◮ How can we estimate the performance with training
◮ How can we predict the performance with future
◮ Problems of overfitting and generalization.
CS 551, Spring 2007 c 2007, Selim Aksoy 34/35
CS 551, Spring 2007 c 2007, Selim Aksoy 35/35