Lei Wang VILA group School of Computing and Information Technology University of Wollongong, Australia
22-OCT-2018
Learning SPD-matrix-based Representation for Visual Recognition Lei - - PowerPoint PPT Presentation
Learning SPD-matrix-based Representation for Visual Recognition Lei Wang VILA group School of Computing and Information Technology University of Wollongong, Australia 22-OCT-2018 Introduction How to represent an image? Scale,
Lei Wang VILA group School of Computing and Information Technology University of Wollongong, Australia
22-OCT-2018
Interest point detection
Dense sampling An image becomes “A bag of features”
“Cat”
Depth Height Width
6
vs.
A set of local descriptors
x1 x2 . . . xn
How to pool?
– Discriminatively Learning Covariance Representation – Exploring Sparse Inverse Covariance Representation – Moving to Kernel-matrix-based Representation (KSPD) – Learning KSPD in deep neural networks
vs.
10
Image is from http://www.statsref.com/HTML/index.html?multivariate_distributions.html
11
12
13
14
Geodesic distance
Pennec X, Fillard P, Ayache N. A Riemannian framework for tensor
Fletcher P T, Principal geodesic analysis
diffusion tensors. Computer Vision and Mathematical Methods in Medical and Biomedical Image Analysis., 2004 Förstner W, Moonen B. A metric for covariance matrices, Geodesy-The Challenge of the 3rd Millennium, 2003 Lenglet C, Statistics on the manifold of multivariate normal distributions: Theory and application to diffusion tensor MRI processing. Journal of Mathematical Imaging and Vision, 2006
2003 2004 2006
15
Euclidean mapping
Arsigny V, Log-Euclidean metrics for fast and simple calculus on diffusion tensors. Magnetic resonance in medicine, 2006, Veeraraghavan A, Matching shape sequences in video with applications in human movement
Tuzel O, Pedestrian detection via classification on riemannian manifolds. PAMI, IEEE Transactions on, 2008
2005 2006 2008
16
Kernel methods
Vemulapalli R, Pillai J K, Chellappa R. Kernel learning for extrinsic classification
Harandi M et al. Sparse coding and dictionary learning for SPD matrices: a kernel approach, ECCV, 2012
manifold of symmetric positive definite matrices, CVPR 2013. Sra S. Positive definite matrices and the S-
2011. Wang R., et. al., Covariance discriminative learning: A natural and efficient approach to image set classification, CVPR, 2012
2011 2012 2013
Quang, Minh Ha, et. Al., Log-Hilbert- Schmidt metric between positive definite
2014
17
Integration with deep learning
Li et al., Is Second-order Information Helpful for Large-scale Visual Recognition? ICCV2017 Huang et al., A Riemannian Network for SPD Matrix Learning, AAAI2017 Improved Bilinear Pooling with CNN, Lin and Maji, BMVC2017 Lin et al, Bilinear CNN Models for Fine-grained Visual Recognition, ICCV2015 Ionescu et al, , Matrix Backpropagation for Deep Networks with Structured Layers, ICCV2015
2015 2017
Koniusz et al., A Deeper Look at Power Normalizations,, CVPR 2018
2018
– Discriminatively Learning Covariance Representation – Exploring Sparse Inverse Covariance Representation – Moving to Kernel-matrix-based Representation (KSPD) – Learning KSPD in deep neural networks
Covariance Matrix
20
21
22
Class 1 Class 2
23
Class 1 Class 2 Class 1 Class 2
24
adjustment
Power-based adjustment Coefficient-based adjustment
25
26
27
29
The most difficult 15 pairs of Brodatz texture data set
30
The most difficult 15 pairs of Brodatz texture data set
[1] X. Mestre, “Improved estimation of eigenvalues and eigenvectors of covariance matrices using their sample estimates,” IEEE Trans. Inf. Theory, vol. 54, pp. 5113–5129, Nov. 2008. [2] B. Efron and C. Morris, “Multivariate empirical Bayes and estimation of covariance matrices,” Ann. Stat., vol. 4, pp. 22–32, 1976. [3] A. Ben-David and C. E. Davidson, “Eigenvalue estimation of hyper-spectral Wishart covariance matrices from limited number of samples,” IEEE Trans. Geosci. Remote Sens., vol. 50, pp. 4384–4396, May 2012.
31
– Discriminatively Learning Covariance Representation – Exploring Sparse Inverse Covariance Representation – Moving to Kernel-matrix-based Representation (KSPD) – Learning KSPD in deep neural networks
33
Small sample 10 ~ 300 High dimensions 50 ~ 400
34
35
(SICE)
36
37
38
40
– Discriminatively Learning Covariance Representation – Exploring Sparse Inverse Covariance Representation – Moving to Kernel-matrix-based Representation (KSPD) – Learning KSPD in deep neural networks
42
43
i-th feature j-th feature Just a linear kernel function!
44
45
Advantages:
what the feature dimensions and sample size are.
computational load.
46
47
48
49
58 60 62 64 66 68 70 72 74 76 78 80 Alex Net (F7) VGG-19 Net (Conv5) Fisher Vector (CVPR15) Cov-RP Ker-RP (RBF)
Comparison on MIT Indoor Scenes Data Set
(Classification accuracy in percentage)
50
51
– Discriminatively Learning Covariance Representation – Exploring Sparse Inverse Covariance Representation – Moving to Kernel-matrix-based Representation (KSPD) – Learning KSPD in deep neural networks
Bilinear CNN Models for Fine-grained Visual Recognition, Lin et al, ICCV2015
Matrix Backpropagation for Deep Networks with Structured Layers, Ionescu et al, ICCV2015
Improved Bilinear Pooling with CNN, Lin and Maji, BMVC2017
Is Second-order Information Helpful for Large-scale Visual Recognition?, Li et al., ICCV2017
57
The kernel-matrix-based SPD representation has not been developed upon deep local descriptors has not been jointly learned via deep learning Existing matrix backpropagation for learning covariance-
encounters numerical stability issue
58
59
60
H = f(K) on the kernel matrix K
Matrix Backpropagation for Deep Networks with Structured Layers, Ionescu et al, ICCV2015
Kernel Pooling for Convolutional Neural Networks, Cui et al, CVPR2017
(d) Low-rank Bilinear Pooling for Fine-Grained Classification, Kong et al, CVPR2017 (c) Compact Bilinear Pooling, Gao et al, CVPR2016 Statistically-motivated Second-order Pooling, Yu and Salzmann, ECCV2018
covariance representation
covariance representation
1.
for SPD Matrices and Its Applications, IEEE Transactions on Neural Networks and Learning Systems (TNNLS), Vol. 27, Issue 5, pp. 1020-1033, May 2016. 2.
Covariance-based Visual Representation, arXiv:1610.08619 [cs.CV]. 3.
Representation with Nonlinear Kernel Matrices, IEEE International Conference
4.
based SPD Representation for Fine-grained Image Recognition, The 15th European Conference on Computer Vision (ECCV), September 2018.
– What is it modelling, relationship to other pooling schemes?
– Optimisation on manifold, kernel learning, prior knowledge?
– Deal with high-dimensional features and large data set?
– Rectangular matrix – Higher order information – Spatial or temporal order
74
Brain Networks with Different Sparsity, Pattern Recognition, 63 642-652, 2017.
Learning for SPD Matrix based Visual Representation, IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
Bayesian Networks from High-dimensional Continuous Neuroimaging Data, IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), Volume: 38 , Issue: 11 , Nov. 1 2016 .
With Compact Representation of SICE Matrices, IEEE Transactions on Biomedical Engineering, 62 (6), 1623-1634, 2015.
Matrix: Application in Brain Functional Network Classification, IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), June 2014
Learning for Discriminative Bayesian Network from Neuroimaging Data, In the 17th International Conference on MICCAI, September 2014.
Images Courtesy of Google Image