Deep Semantic Ranking Based Hashing for Multi-Label Image Retrieval
Presenter: MinKu Kang
1
Fang Zhao, et al. CVPR 2015
Hashing for Multi-Label Image Retrieval Fang Zhao, et al. CVPR - - PowerPoint PPT Presentation
Deep Semantic Ranking Based Hashing for Multi-Label Image Retrieval Fang Zhao, et al. CVPR 2015 Presenter: MinKu Kang 1 Previous Presentation Presented by Youngki Kwon 2 Introduction Ranking Based Image Retrieval Similarity based on
Presenter: MinKu Kang
1
Fang Zhao, et al. CVPR 2015
2
3
4
Pairwise Similarity Similar Dissimilar
5
6
CNN
. . .
CNN
. . .
CNN
. . .
𝐽 𝐽+ 𝐽− 𝐺(𝐽) 𝐺(𝐽+) 𝐺(𝐽−)
Weights are shared. Weights are shared. Triplet Ranking Loss
7
more-similar less-similar
8
𝑠
1 = 3
𝑠
2 = 2
𝐵𝑜𝑑ℎ𝑝𝑠
𝑢𝑠𝑓𝑓, 𝑡𝑣𝑜, 𝑡𝑙𝑧 𝑢𝑠𝑓𝑓, 𝑡𝑣𝑜, 𝑡𝑙𝑧 𝑢𝑠𝑓𝑓, 𝑡𝑙𝑧
Similar dissimilar
9
𝑠
1 = 3
𝑠
2 = 2
𝐵𝑜𝑑ℎ𝑝𝑠
𝑢𝑠𝑓𝑓, 𝑡𝑣𝑜, 𝑡𝑙𝑧 𝑢𝑠𝑓𝑓, 𝑡𝑣𝑜, 𝑡𝑙𝑧 𝑢𝑠𝑓𝑓, 𝑡𝑙𝑧
Similar dissimilar
10
𝑠
1 = 3
𝑠
2 = 2
𝑠
3 = 0 𝑢𝑠𝑓𝑓, 𝑡𝑣𝑜, 𝑡𝑙𝑧 𝑢𝑠𝑓𝑓, 𝑡𝑣𝑜, 𝑡𝑙𝑧 𝑢𝑠𝑓𝑓, 𝑡𝑙𝑧 𝑐𝑣𝑗𝑚𝑒𝑗𝑜, 𝑑𝑏𝑠 𝑠
1 = 3
𝒙𝒊𝒇𝒐 𝒋 = 𝟐 𝑠
2 = 2
𝑠
3 = 0 𝑢𝑠𝑓𝑓, 𝑡𝑣𝑜, 𝑡𝑙𝑧 𝑢𝑠𝑓𝑓, 𝑡𝑙𝑧
𝑐𝑣𝑗𝑚𝑒𝑗𝑜, 𝑑𝑏𝑠 𝑠
1 = 3
𝒙𝒊𝒇𝒐 𝒋 = 𝟑
Possible Triplets
𝒓 𝒚𝒋 𝒚𝒌 𝒚𝒌 Loss for an anchor =
11
Loss for an anchor Loss for all anchors
12
: Encourages each bit averaged over the training data to be mean-zero : Penalized large weights
…
13
Bypassing Connection Skipping Layer
Utilize diverse feature information biased toward visual appearance
14
discontinuous smooth, differentiable Relaxation
15
discontinuous smooth, differentiable Relaxation
But, the sigmoid function had a bad influence on the convergence of the network Many Deep Learning Libraries support automatic, symbolic differentiations
16
𝑠
1 = 3
𝑠2 = 2 𝑠
1 = 3
𝑠3 = 1 Higher weight Ranking Discrepancy
17
MIRFLICKR-25K : Multi-label(24) images from social photography website NUS-WIDE: Multi-label(81) images
18
Average Ranking
19
MIRFLICKR-25K NUS-WIDE
20
MIRFLICKR-25K NUS-WIDE
21
1. What is the most appropriate role of the bypassing connection? a) to increase the capacity of the network b) to utilize the diverse feature information c) to make the training procedure efficient. d) to prevent the overfitting training dataset 2. Choose the term which is not included in the final loss function. a) mean-zero relaxation b) sign function => sigmoid function term c) large-weight penalizing term d) L1 regularization term for the hash codes
22
23