1 yaniv taigman 1 ming yang 1 marc aurelio ranzato 2 lior
play

1 Yaniv Taigman 1 Ming Yang 1 MarcAurelio Ranzato 2 Lior Wolf 1 - PowerPoint PPT Presentation

DeepFace for Unconstrained Face Recognition 1 Yaniv Taigman 1 Ming Yang 1 MarcAurelio Ranzato 2 Lior Wolf 1 Facebook AI Research 2 Tel Aviv University 11/26/2014 Era of big visual data 1.6M daily uploads 60M daily uploads 6B photos (12/2013)


  1. DeepFace for Unconstrained Face Recognition 1 Yaniv Taigman 1 Ming Yang 1 Marc’Aurelio Ranzato 2 Lior Wolf 1 Facebook AI Research 2 Tel Aviv University 11/26/2014

  2. Era of big visual data 1.6M daily uploads 60M daily uploads 6B photos (12/2013) 20B photos (3/2014) 215M daily uploads 350M daily uploads ?B photos 0B photos (11/2013) (11/2013) 100 hours video 400M daily uploads per min 350B photos (4/2014) (3/2014) • 1.75B smartphone users in 2014 • 880B digital photos will be taken in 2014 Sources: www.expandedramblings.com , www.emarketer.com

  3. Tag suggestions No automatic face recognition service in EU countries

  4. Facerec main objective Find a representation & similarity measure such that: • Intra-subject similarity is high • Inter-subject similarity is low

  5. Milestones in face recognition 1997 1999 1973 1991 1964 1999 2001 2006 Kanade’s Belhumeur Blanz & Vetter Turk & Bledsoe Wiskott Viola & Ahonen LBP Fisherfaces Morphable Thesis Pentland Face EBGM Jones faces Eigenfaces Recognition Boosting Slightly modified version of Anil Jain’s timeline

  6. Problem solved? NIST FRVT’s best - performer’s on: 1. Verification: FRR=0.3% at FAR=0.1% 2. Identification: with 1.6 million identities: 95.9% 3. Identification: on LFW with 4,249 identities: 56.7%  Answer: No. • L. Best-Rowden, H. Han, C. Otto, B. Klare, and A. K. Jain. Unconstrained face recognition: Identifying a person of interest from a media collection. IEEE Trans. Information Forensics and Security, 2014.

  7. Constrained vs. unconstrained UNCONSTRAINED CONSTRAINED Labeled Faces in the Wild FRVT property constrained unconstrained resolution about 2000x2000 50x50 viewpoint fully frontal rotated, loose illumination controlled arbitrary occlusion disallowed allowed

  8. Challenges in unconstrained face recognition Gallery 1.Pose 2.Illumination Probes for example 3.Expression 4.Aging 5.Occlusion

  9. A case study • Gallery images: 1 million mug-shot + 6 web images • Probe images: 5 faces • Ranking results – w/o or with demographic filtering Probe faces: A case study of automated face recognition: the Boston Marathon bombing suspects, J. C. Klontz and A.K. Jain, IEEE Computer, 2013

  10. Unconstrained Face Recognition Era: The Labeled Faces in the Wild (LFW) 13,233 photos of 5,749 celebrities celebrities Labeled faces in the wild: A database for studying face recognition in unconstrained environments, Huang, Jain, Learned- Miller, ECCVW, 2008

  11. Face verification (1:1) = !=

  12. Human-level performance • User study on Mechanical Turk – 10 different workers per face pair – Average human performance – Original images, tight crops, inverse crops 99.20% “These results suggest that automatic face verification algorithms should not use regions outside of the face, as they 97.53% could artificially boost accuracy in a manner not applicable on real data .” 94.27% Attribute and simile classifiers for face verification, Kumar, et al., ICCV 2009

  13. LFW: Progress over the recent 7 years • Labeled faces in the wild: A database for studying face recognition in unconstrained environments, ECCVW, 2008. • Attribute and simile classifiers for face verification, ICCV 2009. • Multiple one-shots for utilizing class label information, BMVC 2009. • Large scale strongly supervised ensemble metric learning, with applications to face verification and retrieval, NEC Labs TR, 2012.  Learning hierarchical representations for face verification with convolutional deep belief networks, CVPR, 2012. • Bayesian face revisited: A joint formulation, ECCV 2012. • Tom-vs-pete classifiers and identity preserving alignment for face verification, BMVC 2012. • Blessing of dimensionality: High-dimensional feature and its efficient compression for face verification, CVPR 2013. • Probabilistic elastic matching for pose variant face verification, CVPR 2013. • Fusing robust face region descriptors via multiple metric learning for face recognition in the wild, CVPR 2013. • Fisher vector faces in the wild, BMVC 2013. • A practical transfer learning algorithm for face verification, ICCV 2013.  Hybrid deep learning for computing face similarities, ICCV 2013.  Employed deep learning models for face verification on LFW. Please check http://vis-www.cs.umass.edu/lfw/ for the latest updates.

  14. LFW: Progress over the recent 7 years Accuracy / year Reduction of error wrt human / year 92.58% 95.17% 96.33% 97.53% 85.54% 88.00% 78.47% 73.93% 60.02% 52.32% 49.15% 48.06% 37.08% 37.09% 20.52% 19.24% Labeled Faces in the Wild: A Database for Studying Face Recognition in Unconstrained Environments (results page), Gary B. Huang, Manu Ramesh, Tamara Berg and Erik Learned-Miller.

  15. High-dim LBP • Accurate (27) dense facial landmarks • Concatenate multi-scale descriptors – ~100K-dim LBP, SIFT, Garbor, etc. • Transfer learning: Joint Bayesian Likelihood ratio test: EM update of the between/within class covariance • WDRef dataset – 99,773 images of 2,995 individuals – 95.17% => 96.33% on LFW (unrestricted protocol) Face alignment by explicit shape regression, Cao, et al., CVPR 2012 Bayesian face revisited: A joint formulation, Chen, et al., ECCV 2012 Blessing of dimensionality: High-dimensional feature and its efficient compression for face verification, Chen, et al., CVPR 2013 A practical transfer learning algorithm for face verification, Cao, et al., ICCV 2013

  16. Hybrid deep learning • 12X5 Siamese ConvNets X8 + RBM classification 12 face regions 8 pairs of inputs CelebFaces dataset 87,628 images of 5,436 individuals Hybrid deep learning for computing face similarities, Sun, Wang, Tang, ICCV 2013.

  17. Face recognition pipeline Align Detect Represent Classify Yaniv Lubomir Marc’Aurelio

  18. Faces are 3D objects

  19. Reconstruction accuracy and discriminability Bornstein et al. 2007

  20. Face alignment (‘ Frontalization ’ ) Detect 2D-Aligned 3D-Aligned

  21. 2D alignment f Localize 2D Align

  22. 3D alignment +67 x 2d 2D Align Pnts Piece-wise affine

  23. Rendering of new views

  24. Network architecture ION SFC labels ENTATION ESENTAT REPRES L4: L6: C1: M2: C3: L5: F7: F8: Calista_Flockhart_0002.jpg 16 x 16 x 32 filters 3x3 16 filters 16 x Frontalization 4096d 4030d Detection & Localization 9 x 9 x 16 5 x 5 x 16 11x11 9x9 7 x 7 x 16 Globally Localization Front-End ConvNet Local (Untied) Connected Convolutions

  25. SFC Training dataset 4.4 million photos blindly sampled, containing more than 4,000 identities (permission granted)

  26. Transferred Similarities (Test) (a) Cosine angle DeepFace Replica (b) Kernel Methods DeepFace Replica (c) Siamese Network

  27. Results on LFW

  28. Youtube face dataset (YTF) • Data collection – 3,425 Youtube videos 1,595 celebrities (a subset of LFW subjects) – 5,000 video pairs in 10 splits – Detected and roughly aligned face frames available. • Metric: mean recognition accuracy over 10 folds – Restricted protocol: only same/not-same labels – Unrestricted protocol: face identities, additional training pairs Face recognition in unconstrained videos with matched background similarity, Wolf, Hassner, Maoz, ICCV 2011

  29. Results on YouTube Faces (Video)

  30. Trade-offs (LFW Acc. %) 1. Alignment: 97.35 94.3 93.7 not “astonishing” 91.3 87.9 No Alignment 3D Pertrubation 2D Alignment 3D Alignment 3D Alignment + LBP 2. Dimensionality: 97.17 97 96.72 96.07 95.87 95.53 4096 4096 1024 1024 256 256 bits bits bits 1 3. Sparsity @ 4k dims: 0.8 0.6 0.4 0.2 0 0.1 0.2 0.3 0.4 0.5 0.6 7 0.8 0.9 1

  31. Trade-offs – Cont’d DB Size / DNN Test Error (%) 4. Training data size: 20.7 15.1 10.9 8.74 100% of the data 50% of the data 20% of the data 10% of the data 5. Network Architecture: 13.5 12.6 11.2 8.74 C1+M2+C3+L4+L5+L6+F7 -C3 -L4 -L5 -C3 -L4 -L5

  32. Failure cases • All false negatives on LFW (1%) age sunglasses occlusion/ hats profile errata

  33. Failure cases • All false positive on LFW (0.65%)

  34. Failure cases • Sample false negatives on YTF

  35. Failure cases • Sample false positives on YTF

  36. Face identification (1:N) Probe Gallery Unaccounted challenges in verification: = I.Reliability II.Large confusion (P x G) III.Different distributions != IV.Unknown class

  37. LFW identification (1:N) protocols 2 1. Close Set Gallery Probe #Gallery 1 : - 4,249 - #Probes: 3,143 Measured 3 by Rank-1 rate. Impostor Probe 2. Open Set #Gallery 1 : 596 - … - #Probes: 596 9,491 (‘unknown class’) - #Impostors: UNKNOWN Measured 3 by Rank-1 rate @ 1% False Alarm Rate. 1 Each identity with a single example 2 Unconstrained Face Recognition: Identifying a Person of Interest from a Media Collection Best-Rowden, Han, Otto, Klare and Jain ( IEEE Trans. Information Forensics and Security,) 3 Training is not permitted on LFW (‘unsupervised’)

  38. LFW identification (1:N) results Gallery Probe NIST’s Impostor Probe Cosine similarity measure (‘unsupervised’) : … P Confusion Matrix = G T *P UNKNOWN G is 4096x 4249 G P is 4096x 3143

  39. Bottleneck regularizes transfer learning 0 0 0 0 0 0 0 0 1 0 0 SOFTMAX FC8 FC7 DNN Labels Web-Scale Training for Face Identification; Taigman, Yang, Ranzato, Wolf

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend