Dagstuhl Seminar Inpainting-Based Image Compression
Optimising Data for PDE-Based Inpainting and Compression
Laurent Hoeltgen
hoeltgen@b-tu.de
Optimising Data for PDE-Based Inpainting and Compression Laurent - - PowerPoint PPT Presentation
Dagstuhl Seminar Inpainting-Based Image Compression Optimising Data for PDE-Based Inpainting and Compression Laurent Hoeltgen hoeltgen@b-tu.de Chair for Mathematics of Engineering & Numerical Optimisation Brandenburg University of
Dagstuhl Seminar Inpainting-Based Image Compression
hoeltgen@b-tu.de
Motivation (1)
alternative to transform based approaches has three important pillars each requires advanced optimisation
Motivation (2)
How to reconstruct the image?
PDE has significant influence on our design principles
What are our optimisation criteria?
Which data should we optimise? Should we maximise the quality or minimise the file size?
Outlook
Data Reconstruction (1)
simple to analyse
linear, parameter free, ...
applicable to any domain and co-domain
no restriction on image type, size, quantisation, ...
fast to carry out
Data Reconstruction (2)
consider Laplace equation with mixed boundary conditions:
ΩK: represents known data Ω \ ΩK: region to be inpainted (i.e. missing data) image reconstructions given as solutions u
Data Selection (1)
data locations have no apparent structure data values are real valued large amounts of data are preferred
data positions stored in memory efficient structures data values are quantised sparse sets of data are preferred
Data Selection (2)
We can’t optimise accuracy and file size at the same time.
Using all image data yields perfect accuracy. Using no image data at all yields perfect file size.
We target a fixed amount of image data and optimise accuracy. In case of identical accuracy, we chose the one with smaller size.
Data Selection (3)
brute force search of optimal pixel locations is impossible 105600 ways to choose 5% of data from 0.07 megapixel image most smart-phones yield 12 megapixel photographs for comparison:
Data Selection in the Domain (1)
Data Selection in the Domain (2)
PDE makes also sense if c : Ω → R
can be seen as regularisation
Regions with c > 1 violate Max-Min principle
PDE resembles backward diffusion process may also be interpreted as a Helmholtz equation contrast enhancement
unconditional existence of a solution difficult to show
Data Selection in the Domain (3)
(H., Setzer, Weickert 2013)
u, c
Ω
model has strong similarities to Belhachmi et al. 2009
2 (u − g)2 favours accurate reconstructions λ|c| favours sparse sets of data
2|c|2 required for technical reasons PDE enforces feasible solutions
Data Selection in the Domain (4)
u, c
Ω
energy reflects trade-off between accuracy and sparsity
λ steers sparsity of the interpolation data
small, positive λ: many data points, but good reconstruction large, positive λ: few data points, but bad reconstruction
non-convex, non-smooth, and large scale optimisation
Data Selection in the Domain (5)
linearise constraint to handle non-convexity:
add proximal term and solve iteratively
u, c
Ω
Data Selection in the Domain (6)
2 − uk − g2 2
2 − ck2 2
Data Selection in the Domain (7)
Data Selection in the Domain (8)
Data Selection in the Domain (8)
Data Selection in the Domain (8)
Data Selection in the Domain (8)
Data Selection in the Domain (9)
Data Selection in the Co-Domain (1)
quantisation to n ≪ 256 colours essential for compression
best number of colours hard to determine
hard to predict
Data Selection in the Co-Domain (2)
given mask c, solution u = u(c) of
x
Ω
Data Selection in the Co-Domain (3)
least squares model suggested by Mainberger et al. 2011 yields arbitrary values in R equivalence to mask optimisation
maximises reconstruction quality storage of colour values in R is very expensive
Data Selection in the Co-Domain (4)
memory efficient representation of arbitrary binary masks good colour quantisation strategies
Data Selection in the Co-Domain (5)
Optimal colour values gather around few dominant colours. replacing all colours by their closest dominant colour yields:
reconstruction error becomes slightly larger data becomes much easier to compress
achievable by using clustering strategies
Which clustering method is the best? How to chose our feature values?
Data Selection in the Co-Domain (6)
Data Selection in the Co-Domain (7)
k-means++ hierarchical clustering Gaussian mixture models
Data Selection in the Co-Domain (8)
(H., Breuß 2016)
apply k-means++ on colour values from mask locations
10 15 20 25 30 35 40 45 50 55 60 65 70 75 45 50 55 number of clusters mean squared error
k-means++ 30 colours outperforms original data (170 colours)
Data Selection in the Co-Domain (9)
Silhouette plots allow to evaluate clustering results. Each bar represents a feature.
Values close to 1 indicate a correct labelling. Values close to -1 are probably labelled false.
we obtained:
−1 −0.5 0.5 1 Hierarchical Mean: 0.59, Median: 0.8 −1 −0.5 0.5 1 k-means++ Mean: 0.63, Median: 0.73
Data Selection in the Co-Domain (10)
proposed clustering optimises reconstruction error distribution might be unsuitable for data compression final compressed file size is very hard to predict efficient clustering based on file size is difficult (but necessary)
Summary and Conclusion
trade-off between quality and file size necessary file compression add further constraints some tasks remain difficult to solve
Thank You
http://www-user.tu-cottbus.de/~hoeltgen/