graph cut convex relaxation and continuous max flow
play

Graph cut, convex relaxation and continuous max-flow problem Egil - PowerPoint PPT Presentation

Graph cut, convex relaxation and continuous max-flow problem Egil Bae (UCLA) and Xue-Cheng Tai (U. of Bergen), May 5, 2014 Context of this presentation Overview over recent combinatorial graph cut methods and convex relaxation methods in


  1. Three problems PCLMS or Binary LM (Lie-Lysaker-T.,2005): � min f 1 (1 − u ) + f 2 u + g ( x ) |∇ u | dx . u ( x ) ∈{ 0 , 1 } Ω Convex problem (CEN, (Chan-Esdoglu-Nikolova,2006)) � min f 1 (1 − u ) + f 2 u + g ( x ) |∇ u | dx . u ( x ) ∈ [0 , 1] Ω Graph-cut (Boykov-Kolmogorov,2001) � max p s dx subject to: p s , p t , q Ω p s ( x ) ≤ f 1 ( x ) , p t ( x ) ≤ f 2 ( x ) , | p ( x ) | ≤ g ( x ) , div p ( x ) − p s ( x ) + p t ( x ) = 0 .

  2. Remarks The following approaches are solving the same problem, but did not know each other: ◮ max-flow and min-cut. ◮ Chan-Esedougla-Nikolova 2006 (convex relaxation approach) ◮ Binary Level set methods and PCLSM (piecewise constant level set method) ◮ A cut is nothing else, but the Lagrangian multiplier for the flow conservation constraint!!!

  3. Continuous Max-Flow: Remarks ◮ Min-cut problem is minimizing an energy functional. (Many existing algorithms) Are not using the decent (gradient) info of the energy. ◮ Continuous max-flow/min-cut is a convex minimization problem. A lot of choices, can use decent (gradient) info.

  4. Continuous Max-Flow: How to solve it (Only 2-phase case)? ◮ Popular (discrete) Min-cut algorithms: Augmented Path. Push-relabel, etc, ◮ Available continuouse max-flow/Min-cut approaches: Split-Bregman, Augmented Lagrangian, Primal-Dual approaches. We can use these approach to solve the convex min-cut problem.

  5. Continuous Max-Flow and Min-Cut Multiplier-Based Maximal-Flow Algorithm Augmented lagrangian functional (Glowinski & Le Tallec, 1989) � − c 2 | div q − p s + p t | 2 dx . � � L c ( p s , p t , q , λ ) := p s dx + λ div q − p s + p t Ω minmax subject to: p s ( x ) ≤ f 1 ( x ) , p t ( x ) ≤ f 2 ( x ) , | q ( x ) | ≤ g ( x ) ADMM algorithm: For k=1,... until convergence, solve q k +1 := arg � q � ∞ ≤ α L c ( p k s , p k t , q , λ k ) max p k +1 p s ( x ) ≤ f 1 ( x ) L c ( p s , p k t , q k +1 , λ k ) := arg max s p k +1 p t ( x ) ≤ f 2 ( x ) L c ( p k +1 , p t , q k +1 , λ k ) := arg max t s λ k +1 = λ k − c (div q k +1 − p k +1 + p k +1 ) s t

  6. Continuous Max-Flow and Min-Cut Other algorithms for solving the relaxed problem: add p = ∇ u ◮ Bresson et. al. ◮ fix µ k and solve ROF problem |∇ λ ( x ) | dx + 1 � λ k +1 := arg min � 2 θ � λ ( x ) − µ k ( x ) � 2 � α λ Ω ◮ fix λ k +1 and solve � 1 � µ k +1 := arg min 2 θ � µ ( x ) − λ k +1 � 2 + � � � µ ( x ) f 1 ( x ) − f 2 ( x ) dx µ ∈ [0 , 1] Ω ◮ Goldstein-Osher: Split Bregman / augmented lagrangian

  7. Convergence Figure : Red line: max-flow algorithm. Blue line: Splitting algorithm (Bresson et. al. 2007)

  8. Metrication error, Parallel, GPU, ... Experiment of mean-curvature driven 3D surface evolution (volume size: 150X150X150 voxels). (a) The radius plot of the 3D ball evolution driven by its mean-curvature flow, which is computed by the proposed continuous max-flow algorithm; its function is theoretically r ( t ) = √ C − 2 t . (b) The computed 3D ball at one discrete time frame, which fits a perfect 3D ball shape. This is in contrast to (c), the computation result by graph cut [15] with a 3D 26-connected graph. The computation time of the continuous max-flow algorithm for each discrete time evolution is around 1 sec., which is faster than the graph cut method (120 sec.) Ref: Y. Yuan, E. Ukwatta, X. Tai, A. Fenster, and C. Schnorr. A fast global optimization-based approach to evolving contours with generic shape prior. Technical report, also UCLA Tech. Report CAM 12-38, 2012.

  9. Metrication error, Parallel, GPU, ... ◮ Fully parallel, easy GPU implementation. ◮ linear grow of computational cost (per iteration): 2D, 3D, ...

  10. Gamma convergence When h �→ 0, the energy at the minimizer on the discrete graph converges to energy at the minimizer of the continuous problem both for isotropic and anisotropic TV: � � � | u x | 2 + | u y | 2 dx . TV ( u ) = ( | u x | + | u y | ) dx , TV ( u ) = Ω Ω Γ = ⇒

  11. Gamma convergence When h �→ 0, the energy at the minimizer on the discrete graph converges to energy at the minimizer of the continuous problem both for isotropic and anisotropic TV: � � � | u x | 2 + | u y | 2 dx . TV ( u ) = ( | u x | + | u y | ) dx , TV ( u ) = Ω Ω Γ = ⇒ ◮ Anisotropic TV:

  12. Gamma convergence When h �→ 0, the energy at the minimizer on the discrete graph converges to energy at the minimizer of the continuous problem both for isotropic and anisotropic TV: � � � | u x | 2 + | u y | 2 dx . TV ( u ) = ( | u x | + | u y | ) dx , TV ( u ) = Ω Ω Γ = ⇒ ◮ Anisotropic TV: sub-modular,

  13. Gamma convergence When h �→ 0, the energy at the minimizer on the discrete graph converges to energy at the minimizer of the continuous problem both for isotropic and anisotropic TV: � � � | u x | 2 + | u y | 2 dx . TV ( u ) = ( | u x | + | u y | ) dx , TV ( u ) = Ω Ω Γ = ⇒ ◮ Anisotropic TV: sub-modular, Can use standard graph cut methods. ◮ Isotropic TV:

  14. Gamma convergence When h �→ 0, the energy at the minimizer on the discrete graph converges to energy at the minimizer of the continuous problem both for isotropic and anisotropic TV: � � � | u x | 2 + | u y | 2 dx . TV ( u ) = ( | u x | + | u y | ) dx , TV ( u ) = Ω Ω Γ = ⇒ ◮ Anisotropic TV: sub-modular, Can use standard graph cut methods. ◮ Isotropic TV: not sub-modular,

  15. Gamma convergence When h �→ 0, the energy at the minimizer on the discrete graph converges to energy at the minimizer of the continuous problem both for isotropic and anisotropic TV: � � � | u x | 2 + | u y | 2 dx . TV ( u ) = ( | u x | + | u y | ) dx , TV ( u ) = Ω Ω Γ = ⇒ ◮ Anisotropic TV: sub-modular, Can use standard graph cut methods. ◮ Isotropic TV: not sub-modular, Cannot use stander graph cut methods,

  16. Gamma convergence When h �→ 0, the energy at the minimizer on the discrete graph converges to energy at the minimizer of the continuous problem both for isotropic and anisotropic TV: � � � | u x | 2 + | u y | 2 dx . TV ( u ) = ( | u x | + | u y | ) dx , TV ( u ) = Ω Ω Γ = ⇒ ◮ Anisotropic TV: sub-modular, Can use standard graph cut methods. ◮ Isotropic TV: not sub-modular, Cannot use stander graph cut methods, primal-dual approach is faster even !

  17. Literature: Gamma convergence for discrete TV ◮ Isotropic TV: h �→ 0 h = mesh size. Braides-2002, Chambolle-2004, Chambolle-Caselles-Novaga-Cremers-Pock-2010, ◮ Anisotropic TV: h �→ 0 h = mesh size. Chambolle-Caselles-Novaga-Cremers-Pock-2010, Gennip-Bertozzi-2012, Trillo-Slepcev-2013,

  18. Multiphase Approaches Multiphase Approaches

  19. From Two-phase to multi-phases ◮ Related to garph cut, α -expansion and α − β swap are mostly popular approaches for multiphase ”labelling”. ◮ Approximations are made and upper bounded has been given. Ref: Y. Boykov and O. Veksler and R. Zabih: Fast approximate energy minimization via graph cuts, IEEE Transactions on Pattern Analysis and Machine Intelligence, 23, 1222-1239, (2001).

  20. Multiphase problems – Approach I 1 0.8 0.6 0.4 0.2 0 0 20 40 60 100 90 80 80 60 70 50 40 30 We need to identify n 100 10 20 0 characteristic functions 1 0.8 0.6 ψ i ( x ) , i = 1 , 2 · · · n : 0.4 0.2 0 0 n 20 40 � 60 ψ i ( x ) ∈ { 0 , 1 } , ψ i ( x ) = 1 . 100 80 90 80 70 60 50 40 20 30 100 10 0 i =1 1 0.8 0.6 0.4 0.2 0 0 20 40 60 90 100 80 80 70 60 50 30 40 20 100 10 0

  21. Multiphase problems – Approach II Each point x ∈ Ω is labelled by a vector function: u ( x ) = ( u 1 (2) , u 2 ( x ) , · · · u d ( x )) .

  22. Multiphase problems – Approach II Each point x ∈ Ω is labelled by a vector function: u ( x ) = ( u 1 (2) , u 2 ( x ) , · · · u d ( x )) . ◮ Multiphase: Total number of phases n = 2 d . (Chan-Vese) u i ( x ) ∈ { 0 , 1 } .

  23. Multiphase problems – Approach II Each point x ∈ Ω is labelled by a vector function: u ( x ) = ( u 1 (2) , u 2 ( x ) , · · · u d ( x )) . ◮ Multiphase: Total number of phases n = 2 d . (Chan-Vese) u i ( x ) ∈ { 0 , 1 } . ◮ More than binary labels: Total number of phases n = B d . u i ( x ) ∈ { 0 , 1 , 2 , · · · B } .

  24. Multiphase problems – Approach III Each point x ∈ Ω is labelled by i = 1 , 2 , · · · n . 3 u ( x ) = i , 2.5 2 1.5 ◮ One label function is enough 1 0 20 for any n phases. 40 ◮ More generall 60 100 90 80 80 70 60 50 u ( x ) = ℓ i , i = 1 , 2 , · · · n . 40 30 20 100 10 0

  25. Literature: Multiphase Approach I Zach-et-al-2008 (VMV), Lellmann-Kappes-Yuan-Becker-Schn¨ orr-2009 (SSVM), Lellman-Schnorr-2011 (SIIMS), Li-Ng-Zeng-Chen-2010 (SIIMS), Lellman-Lellman-Widman-Schnorr-2013 (IJCV), Qiao-Wang-Ng-2013, Bae-Yuan-Tai-2011 (IJCV)

  26. Literature: Multiphase Approach II Vese-Chan-2002 (IJCV),Lie-Lysaker-Tai-2006 (IEEE TIP), Brown-Chan-Bresson-2010 (cam-report 10-43), Bae-Tai 2009/2014 (JMIV)

  27. Literature: Multiphase Approach III Chung-Vese-2005 (EMMCVPR), Lie-Lysaker-Tai-2006 (Math. Comp), Ishikawa-2004 (PAMI), Pock-Chambolle-Bischof-Cremers 2008/2010 (SIIMS), Kim-Kang-2012 (IEEE TIP), Jung-Kang-Shen-2007, Wei-Wang-2009, Luo-Tong-Luo-Wei-Wang-2009, Bae-Yuan-Tai-Boykov 2010/2014

  28. Multiphase Approach Multiphase Approach (I) Graph for characteristic functions Ref: Yuan-Bae-T.-Boykov (ECCV10): A continuous max-flow approach to Potts model, Computer Vision–ECCV (2010), pp. 379–392. Ref: Bae-Yuan-Tai: Global minimization for continuous multiphase partitioning problems using a dual approach, International journal of computer vision, 92, 112–129(2011).

  29. Multi-partitioning problem Multi-partitioning problem (Pott’s model) n n � � � � min f i dx + g ( x ) ds , { Ω i } Ω i ∂ Ω i i =1 i =1 ∪ n ∩ n such that i =1 Ω i = Ω , i =1 Ω i = ∅

  30. Multi-partitioning problem Multi-partitioning problem (Pott’s model) n n � � � � min f i dx + g ( x ) ds , { Ω i } Ω i ∂ Ω i i =1 i =1 ∪ n ∩ n such that i =1 Ω i = Ω , i =1 Ω i = ∅ Pott’s model in terms of characteristic functions n n n � � � � � min u i ( x ) f i ( x ) dx + g ( x ) |∇ u i | dx , s.t. u i ( x ) = 1 u i ( x ) ∈{ 0 , 1 } Ω Ω i =1 i =1 i =1

  31. Multi-partitioning problem Multi-partitioning problem (Pott’s model) n n � � � � min f i dx + g ( x ) ds , { Ω i } Ω i ∂ Ω i i =1 i =1 ∪ n ∩ n such that i =1 Ω i = Ω , i =1 Ω i = ∅ Pott’s model in terms of characteristic functions n n n � � � � � min u i ( x ) f i ( x ) dx + g ( x ) |∇ u i | dx , s.t. u i ( x ) = 1 u i ( x ) ∈{ 0 , 1 } Ω Ω i =1 i =1 i =1 � 1 , x ∈ Ω i u i ( x ) = χ Ω i ( x ) := , i = 1 , . . . , n 0 , x / ∈ Ω i

  32. A convex relaxation approach Relaxed Potts’ model in terms of characteristic functions (primal model) n n � � � � E P ( u ) = g ( x ) |∇ u i | dx , min u i ( x ) f i ( x ) dx + u Ω Ω i =1 i =1

  33. A convex relaxation approach Relaxed Potts’ model in terms of characteristic functions (primal model) n n � � � � E P ( u ) = g ( x ) |∇ u i | dx , min u i ( x ) f i ( x ) dx + u Ω Ω i =1 i =1 n � s.t. u ∈ △ + = { ( u 1 ( x ) , . . . , u n ( x )) | u i ( x ) = 1 ; u i ( x ) ≥ 0 } i =1 ◮ Convex optimization problem ◮ Optimization techniques: Zach et. al. alternating TV minimization. Lellmann et. al: Douglas Rachford splitting and special thresholding, Bae-Yuan-T. (2010), Chambolle-Crmer-Pock (2012).

  34. Dual formulation of relaxation: Bae-Yuan-T. (IJCV, 2011) Dual model: C λ := { p : Ω �→ R 2 | | p ( x ) | 2 ≤ g ( x ) , p n | ∂ Ω = 0 } , ◮ Hence the primal-dual model can be optimized pointwise for u n n � � � � min u i ( x ) f i ( x ) dx + g ( x ) |∇ u i | dx , u ∈△ + Ω Ω i =1 i =1

  35. Dual formulation of relaxation: Bae-Yuan-T. (IJCV, 2011) Dual model: C λ := { p : Ω �→ R 2 | | p ( x ) | 2 ≤ g ( x ) , p n | ∂ Ω = 0 } , ◮ Hence the primal-dual model can be optimized pointwise for u n n � � � � min u i ( x ) f i ( x ) dx + g ( x ) |∇ u i | dx , u ∈△ + Ω Ω i =1 i =1 n � � max u ∈△ + E ( u , p ) = min u i ( f i + div p i ) dx p i ∈ C λ Ω i =1

  36. Dual formulation of relaxation: Bae-Yuan-T. (IJCV, 2011) Dual model: C λ := { p : Ω �→ R 2 | | p ( x ) | 2 ≤ g ( x ) , p n | ∂ Ω = 0 } , ◮ Hence the primal-dual model can be optimized pointwise for u n n � � � � min u i ( x ) f i ( x ) dx + g ( x ) |∇ u i | dx , u ∈△ + Ω Ω i =1 i =1 n � � max u ∈△ + E ( u , p ) = min u i ( f i + div p i ) dx p i ∈ C λ Ω i =1 n � � = max min u i ( x )( f i ( x ) + div p i ( x )) dx p i ∈ C λ u ( x ) ∈△ + Ω i =1

  37. Dual formulation of relaxation: Bae-Yuan-T. (IJCV, 2011) Dual model: C λ := { p : Ω �→ R 2 | | p ( x ) | 2 ≤ g ( x ) , p n | ∂ Ω = 0 } , ◮ Hence the primal-dual model can be optimized pointwise for u n n � � � � min u i ( x ) f i ( x ) dx + g ( x ) |∇ u i | dx , u ∈△ + Ω Ω i =1 i =1 n � � max u ∈△ + E ( u , p ) = min u i ( f i + div p i ) dx p i ∈ C λ Ω i =1 n � � = max min u i ( x )( f i ( x ) + div p i ( x )) dx p i ∈ C λ u ( x ) ∈△ + Ω i =1 � � � = max min( f 1 + div p 1 , . . . , f n + div p n ) dx p i ∈ C λ Ω

  38. Dual formulation of relaxation: Bae-Yuan-T. (IJCV, 2011) Dual model: C λ := { p : Ω �→ R 2 | | p ( x ) | 2 ≤ g ( x ) , p n | ∂ Ω = 0 } , ◮ Hence the primal-dual model can be optimized pointwise for u n n � � � � min u i ( x ) f i ( x ) dx + g ( x ) |∇ u i | dx , u ∈△ + Ω Ω i =1 i =1 n � � max u ∈△ + E ( u , p ) = min u i ( f i + div p i ) dx p i ∈ C λ Ω i =1 n � � = max min u i ( x )( f i ( x ) + div p i ( x )) dx p i ∈ C λ u ( x ) ∈△ + Ω i =1 � � � = max min( f 1 + div p 1 , . . . , f n + div p n ) dx p i ∈ C λ Ω E D ( p ) = max p i ∈ C λ

  39. Convex relaxation over unit simplex � p ∗ ∈ arg sup E D ( p ) = min( f 1 + div p 1 , . . . , f n + div p n ) dx p i ∈ C α Ω n � u ∗ ∈ arg min E ( u , p ∗ ) = � u i ( x )( f i ( x ) + div p ∗ min i ( x )) dx u ( x ) ∈△ + u ∈△ + Ω i =1 Theorem Let p ∗ be optimal to the dual model. For each x ∈ Ω, a binary primal optimal variable u ∗ ( x ) can be recovered by � 1 if k = arg min i =1 ,..., n ( f i ( x ) + div p ∗ i ( x )) u ∗ k ( x ) = , 0 otherwise provided the n values ( f 1 ( x ) + div p ∗ 1 ( x ) , ..., f n ( x ) + div p ∗ n ( x )) have a unique minimizer. Then ( u ∗ , p ∗ ) is a saddle point, i.e. E P ( u ∗ ) = E ( u ∗ , p ∗ ) = E D ( p ∗ )

  40. Convex relaxation over unit simplex � p ∗ ∈ arg sup E D ( p ) = min( f 1 + div p 1 , . . . , f n + div p n ) dx p i ∈ C α Ω n � u ∗ ∈ arg min E ( u , p ∗ ) = � u i ( x )( f i ( x ) + div p ∗ min i ( x )) dx u ( x ) ∈△ + u ∈△ + Ω i =1 Theorem Let p ∗ be optimal to the dual model. If the n values ( f 1 ( x ) + div p ∗ 1 ( x ) , ..., f n ( x ) + div p ∗ n ( x )) have at most two minimizers for each x ∈ Ω, there exists optimal binary primal variables u ∗ such that ( u ∗ , p ∗ ) is a saddle point, i.e. E P ( u ∗ ) = E ( u ∗ , p ∗ ) = E D ( p ∗ )

  41. Convex relaxation over unit simplex a e b a c e b d f Top: (a) Input, (b) alpha expansion (c) dual model. Bottom: (d) Input, (e) ground truth, (f) alpha expansion, (g) alpha-beta swap, (h) Lellmann et. al., (i) dual model.

  42. Multiple Phases: Convex Relaxed Potts Model (CR-PM) –Yuan-Bae-T.-Boykov (ECCV’10) Continuous Max-Flow Model (CMF-PM) 1. n copies Ω i , i = 1 , . . . , n , of Ω; 2. For ∀ x ∈ Ω, the same source flow p s ( x ) from the source s to x ∈ Ω i , i = 1 , . . . , n , simultaneously; 3. For ∀ x ∈ Ω, the sink flow p i ( x ) from x at Ω i , i = 1 , . . . , n , of Ω to the sink t . p i ( x ), i = 1 , . . . , n , may be different one by one; 4. The spatial flow q i ( x ), i = 1 , . . . , n defined within each Ω i .

  43. Max-flow on this graph Max-Flow : � p s , p , q { P ( p s , p , q ) = p s dx } max Ω | q i ( x ) | ≤ g ( x ) , p i ( x ) ≤ f i ( x ) , (div q i − p s + p i )( x ) = 0 , i = 1 , 2 , · · · n . Note that p s ( x ) = div q i ( x ) + p i ( x ) , i = 1 , 2 · · · n . Thus p s ( x ) = min( f 1 + div p 1 , . . . , f n + div p n ) . � Therefore, the maximum of Ω p s ( x ) is: � max min( f 1 + div p 1 , . . . , f n + div p n ) dx | q i ( x ) |≤ g ( x ) Ω

  44. (Convex) min-cut on this graph m � � p s , p , q min max u { E ( p s , p , q , u ) = p s dx + u i (div q i − p s + p i ) dx } Ω i =1 s.t. p i ( x ) ≤ f i ( x ) , | q i ( x ) | ≤ g ( x ) . Rearranging the energy functional E ( · ), we that m m m � � � � E ( p s , p , q , u ) = (1 − u i ) p s + u i p i + u i div q i . dx . Ω i =1 i =1 i =1 The following constraint are automatically satisfied from the optimization: m � u i ( x ) ≥ 0 , u i = 1 . i =1

  45. (Convex) min-cut: Dual formulation It gives the convex min-cut from the dual formulation: � min u i ( x ) f i ( x ) + g ( x ) |∇ u i ( x ) | u i Ω n � u i ( x ) = 1 , u i ( x ) ≥ 0 . s.t i =1

  46. Algorithms Augmented Lagrangian functional n n � � u i , div q i − p s + p i � − c � � � div q i − p s + p i � 2 p s dx + 2 Ω i =1 i =1 Augmented Lagrangian Method (ADMM): i , q 0 and φ 0 . For k = 0 , 1 , ... Initialize p 0 s , p 0 2 � q i � ∞ ≤ α − c � � q k +1 � div q i + p k i − p k s − u k := arg max i / c , i = 1 , ..., n � � i 2 � 2 p i ( x ) ≤ ρ ( l i , x ) − c � � p k +1 � p i + div q k +1 − p k s − u k := arg max i / c , i = 1 , ..., n � � i i 2 � n 2 � p s dx − c � � p k +1 � � p s − ( p k +1 + div q k +1 ) + u k := arg max i / c , � � s i i 2 � p s Ω i =1 u k +1 = u k i − c (div q k +1 − p k +1 + p k +1 ) , i = 1 , ..., n s i i i

  47. Algorithms Comparisons between algorithms: Zach et al 08, Lellmann et al. 09 and the proposed max-flow algorithm: for three images, different precision ǫ are used and the total number of iterations to reach convergence is evaluated. Brain ǫ ≤ 10 − 5 Flower ǫ ≤ 10 − 4 Bear ǫ ≤ 10 − 4 Zach et al 08 fail to reach such a precision Lellmann et al. 09 421 iter. 580 iter. 535 iter. Proposed algorithm 88 iter. 147 iter. 133 iter.

  48. Outline of this presentation First part: Exact optimization ◮ Will focus on two approaches for multiphase problems with global optimality guarantee. ◮ Both can be formulated as max-flow/min-cut problems on a graph in discrete setting. ◮ Both can be exactly formulated as convex problems on continuous setting. Dual problems can be formulated as continuous max-flow problems. Second part: Approximate optimization ◮ Convex relaxations for broader set of non-convex problems. ◮ Includes Potts’ model and joint optimization of regions and region parameters in image segmentation. ◮ Dual problems can be formulated as max-flow, but now there may be a duality gap to original problems

  49. Problem formulations Image partition problems with multiple regions Given input image I 0 defined over Ω. Find partition { Ω i } n i =1 of Ω by solving n � � f i ( I 0 ( x )) dx + α R ( { ∂ Ω i } n min i =1 ) { Ω i } n Ω i i =1 i =1 ∪ n ∩ n such that i =1 Ω i = Ω , i =1 Ω i = ∅ n is known or unknown in advance. Example (Potts’ model): n n � � � f i ( I 0 ( x )) dx + � min α ds , { Ω i } n Ω i ∂ Ω i i =1 i =1 i =1 Discretized problem is NP-hard for n > 2

  50. Problem formulations Image partition problems with multiple regions Given input image I 0 defined over Ω. Find partition { Ω i } n i =1 of Ω by solving n � � f i ( I 0 ( x )) dx + α R ( { ∂ Ω i } n min i =1 ) { Ω i } n Ω i i =1 i =1 ∪ n ∩ n such that i =1 Ω i = Ω , i =1 Ω i = ∅ n is known or unknown in advance. Example (Potts’ model): n n � � | I 0 ( x ) − ξ i | β dx + � � min α ds , { Ω i } n Ω i ∂ Ω i i =1 i =1 i =1 Discretized problem is NP-hard for n > 2

  51. Problem formulations Image partition problems with multiple regions Given input image I 0 defined over Ω. Find partition { Ω i } n i =1 of Ω by solving n � � f i ( ξ i , I 0 ( x )) dx + α R ( { ∂ Ω i } n min i =1 ) { Ω i } n i =1 , { ξ i } n i =1 ∈ X Ω i i =1 ∪ n ∩ n such that i =1 Ω i = Ω , i =1 Ω i = ∅ n is known or unknown in advance. Example: n n � � | I 0 ( x ) − ξ i | β dx + � � min α ds , { Ω i } n i =1 , { ξ i } n i =1 ∈ R Ω i ∂ Ω i i =1 i =1 If regularization α = 0: ”k-mean” problem, which is known to be NP-hard.

  52. Different representations of partitions in terms of functions ◮ 1) Vector function: u ( x ) = ( u 1 ( x ) , ..., u n ( x )) = e i for x ∈ Ω i ◮ 2) Labeling function: ℓ ( x ) = i for all x ∈ Ω i ◮ 3) log representation by m = log 2 ( n ) binary functions φ 1 , ...φ m φ 1 ( x ) = 1 , φ 2 ( x ) = 0 x ∈ Ω 1 iff u ( x ) = e 1 ℓ ( x ) = 1 φ 1 ( x ) = 1 , φ 2 ( x ) = 1 x ∈ Ω 2 iff u ( x ) = e 2 ℓ ( x ) = 2 φ 1 ( x ) = 0 , φ 2 ( x ) = 0 x ∈ Ω 3 iff u ( x ) = e 3 ℓ ( x ) = 3 φ 1 ( x ) = 0 , φ 2 ( x ) = 1 x ∈ Ω 4 iff u ( x ) = e 4 ℓ ( x ) = 4 Table: Representation of 4 regions.

  53. Log representation by two binary functions Ω 1 = { x ∈ Ω s.t. φ 1 ( x ) > 0 , φ 2 ( x ) < 0 } Ω 2 = { x ∈ Ω s.t. φ 1 ( x ) > 0 , φ 2 ( x ) > 0 } Ω 3 = { x ∈ Ω s.t. φ 1 ( x ) < 0 , φ 2 ( x ) < 0 } Ω 4 = { x ∈ Ω s.t. φ 1 ( x ) < 0 , φ 2 ( x ) > 1 } Vese and Chan 2002, A Multiphase Level Set Framework for Image Segmentation Using the Mumford and Shah Model, International Journal of Computer Vision 50(3), 271–293

  54. Log representation by two binary functions Ω 1 = { x ∈ Ω s.t. φ 1 ( x ) = 1 , φ 2 ( x ) = 0 } Ω 2 = { x ∈ Ω s.t. φ 1 ( x ) = 1 , φ 2 ( x ) = 1 } Ω 3 = { x ∈ Ω s.t. φ 1 ( x ) = 0 , φ 2 ( x ) = 0 } Ω 4 = { x ∈ Ω s.t. φ 1 ( x ) = 0 , φ 2 ( x ) = 1 } Lie et al. 2006, A Binary Level Set Model and Some Applications to Mumford–Shah Image Segmentation, IEEE transactions on image processing, 15(5), pg. 1171 - 1181

  55. Log representation by two binary functions 4 regions as intersection of 2 level set functions � � |∇ H ( φ 1 ) | + α |∇ H ( φ 2 ) | φ 1 ,φ 2 α min Ω Ω � { H ( φ 1 ) H ( φ 2 ) f 2 + H ( φ 1 )(1 − H ( φ 2 )) f 1 + Ω +(1 − H ( φ 1 )) H ( φ 2 ) f 4 + (1 − H ( φ 1 ))(1 − H ( φ 2 )) f 3 } dx . ◮ Heaviside function H ( φ ) = 1 if φ > 0 and H ( φ ) = 0 if φ < 0 ◮ Interpretation of regions: Ω 1 = { x ∈ Ω s.t. φ 1 ( x ) > 0 , φ 2 ( x ) < 0 } Ω 2 = { x ∈ Ω s.t. φ 1 ( x ) > 0 , φ 2 ( x ) > 0 } Ω 3 = { x ∈ Ω s.t. φ 1 ( x ) < 0 , φ 2 ( x ) < 0 } Ω 4 = { x ∈ Ω s.t. φ 1 ( x ) < 0 , φ 2 ( x ) < 0 } Vese and Chan 2002, A Multiphase Level Set Framework for Image Segmentation Using the Mumford and Shah Model, International Journal of Computer Vision 50(3), 271–293

  56. Log representation by two binary functions 4 regions as intersection of 2 binary functions � � |∇ φ 1 | + α |∇ φ 2 | + φ 1 ,φ 2 α min Ω Ω � { φ 1 φ 2 f 2 + φ 1 (1 − φ 2 ) f 1 + Ω +(1 − φ 1 ) φ 2 f 4 + (1 − φ 1 )(1 − φ 2 ) f 3 } dx . ◮ Minimize over constraint φ 1 ( x ) , φ 2 ( x ) ∈ { 0 , 1 } ∀ x ∈ Ω. ◮ Interpretation of regions: Ω 1 = { x ∈ Ω s.t. φ 1 ( x ) = 1 , φ 2 ( x ) = 0 } Ω 2 = { x ∈ Ω s.t. φ 1 ( x ) = 1 , φ 2 ( x ) = 1 } Ω 3 = { x ∈ Ω s.t. φ 1 ( x ) = 0 , φ 2 ( x ) = 0 } Ω 4 = { x ∈ Ω s.t. φ 1 ( x ) = 0 , φ 2 ( x ) = 1 } Lie et al. 2006, A Binary Level Set Model and Some Applications to Mumford–Shah Image Segmentation, IEEE transactions on image processing, 15(5), pg. 1171 - 1181

  57. Convex formulation log representation � � |∇ φ 1 | + α |∇ φ 2 | φ 1 ,φ 2 ∈{ 0 , 1 } α min Ω Ω � (1 − φ 1 ( x )) C ( x )+(1 − φ 2 ( x )) D ( x )+ φ 1 ( x ) A ( x )+ φ 2 ( x ) B ( x ) dx Ω � max { φ 1 ( x ) − φ 2 ( x ) , 0 } E ( x ) − min { φ 1 ( x ) − φ 2 ( x ) , 0 } F ( x ) dx + Ω  A ( x ) + B ( x ) = f 2 ( x )   C ( x ) + D ( x ) = f 3 ( x )  A ( x ) + E ( x ) + D ( x ) = f 1 ( x )   B ( x ) + F ( x ) + C ( x ) = f 4 ( x )  ◮ Energy is convex provided E ( x ) , F ( x ) ≥ 0 for all x ∈ Ω. ◮ Discrete counterpart is submodular iff ∃ E ( x ) , F ( x ) ≥ 0 for all x ∈ Ω (otherwise NP-hard) Bae and Tai, Efficient Global Minimization Methods for Image Segmentation Models with Four Regions, Journal of Mathematical Imaging and Vision, 2014

  58. Convex formulation log representation � � |∇ φ 1 | + α |∇ φ 2 | φ 1 ( x ) ,φ 2 ( x ) ∈ [0 , 1] α min Ω Ω � (1 − φ 1 ( x )) C ( x )+(1 − φ 2 ( x )) D ( x )+ φ 1 ( x ) A ( x )+ φ 2 ( x ) B ( x ) dx Ω � max { φ 1 ( x ) − φ 2 ( x ) , 0 } E ( x ) − min { φ 1 ( x ) − φ 2 ( x ) , 0 } F ( x ) dx + Ω  A ( x ) + B ( x ) = f 2 ( x )   C ( x ) + D ( x ) = f 3 ( x )  A ( x ) + E ( x ) + D ( x ) = f 1 ( x )   B ( x ) + F ( x ) + C ( x ) = f 4 ( x )  ◮ Minimize over convex constraint φ 1 ( x ) , φ 2 ( x ) ∈ [0 , 1] ∀ x ∈ Ω. ◮ Theorem : Binary functions obtained by thresholding solution of convex problem φ 1 , φ 2 at any level t ∈ (0 , 1] is a global minimizer to the original problem.

  59. Convex formulation log representation ◮ Exists E ( x ) , F ( x ) ≥ 0 if f 2 ( x ) + f 3 ( x ) ≤ f 1 ( x ) + f 4 ( x ) . ◮ In case of f i = | I 0 − c i | β , a sufficient condition is | c 2 − I | β + | c 3 − I | β ≤ | c 1 − I | β + | c 4 − I | β , ∀ I ∈ [0 , L ] , ◮ Proposition 1: Let 0 ≤ c 1 < c 2 < c 3 < c 4 . Condition is satisfied for all I ∈ [ c 2 − c 1 , c 4 − c 3 ] for any β ≥ 1. 2 2 ◮ Proposition 2: Let 0 ≤ c 1 < c 2 < c 3 < c 4 . There exists a B ∈ N such that condition is satisfied for any β ≥ B .

  60. Convex formulation log representation a c b Figure: L 2 data fidelity: (a) input, (b) level set method gradient descent, (c) New convex formulation of Chan-Vese model (global minimum). e d Figure: Level set method: (d) bad initialization, (e) result.

  61. Convex formulation log representation a c b e d f Figure: (a) Input image, (b) ground truth, (c) level set method gradient descent, (d) global minimum computed by new graph cut approach in discrete setting, (e) New convex optimization approach in continuous setting before threshold, (f) convex minimization approach after threshold (global optimum).

  62. Convex formulation log representation a b c d Figure: L 2 data fidelity: (a) Input, (b) global minimum discrete Chan-Vese model 4 neighbors, (c) convex formulation before threshold, (d) convex formulation after threshold (global minimum).

  63. Convex formulation log representation a b c d Figure: Segmentation with L 2 data term: (a) Input, (b) graph cut 4 neighbors (c) convex formulation before threshold, (d) convex formulation after threshold (global minimum).

  64. Convex formulation log representation a b c d Figure: Segmentation with L 2 data term: (a) Input, (b) result graph cut 8 neighbors in discrete setting (c) result convex formulation before threshold, (d) result convex formulation after threshold (global optimum).

  65. Convex formulation log representation a c b e d f Figure: (a) Input image, (b) ground truth, (c) gradient descent, (d) alpha expansion, (e) alpha-beta swap, (f) convex model.

  66. Log representation - minimization by graph cuts Discrete energy, anisotropic TV � φ 1 ,φ 2 ∈B E d ( φ 1 , φ 2 ) = E data ( φ 1 p , φ 2 min p ) p p ∈P � � � � w pq | φ 1 p − φ 1 w pq | φ 2 p − φ 2 + α q | + α q | p ∈P p ∈P q ∈N k q ∈N k p p ( φ 1 p , φ 2 p ) = { φ 1 p φ 2 p f 2 ( p ) + φ 1 p (1 − φ 2 E data p ) f 1 ( p )) p +(1 − φ 1 p ) φ 2 p f 4 ( p ) + (1 − φ 1 p )(1 − φ 2 p ) f 3 ( p ) } .

  67. Log representation - minimization by graph cuts Graph construction 1 grid point 2 grid points ◮ Associate two vertices to each grid point ( v p , 1 and v p , 2 ) ◮ For any cut ( V s , V t ) ◮ If v p , i ∈ V s then φ i = 1 for i = 1 , 2 ◮ If v p , i ∈ V t then φ i = 0 for i = 1 , 2 ◮ Figure left: graph corresponding to one grid point p ◮ Figure right: graph corresponding to two grid points p and q ◮ Red: Data edges, constituting E data ( φ 1 , φ 2 ) ◮ Blue: Regularization edges with weight w pq Bae and Tai EMMCVPR 2009, Kolmogorov PAMI 2004

  68. Log representation - minimization by graph cuts Graph construction ◮ Linear system for finding edge weights  A ( p ) + B ( p ) = f 2 ( p )   C ( p ) + D ( p ) = f 3 ( p )  A ( p ) + E ( p ) + D ( p ) = f 1 ( p )   B ( p ) + F ( p ) + C ( p ) = f 4 ( p )  such that E ( p ) , F ( p ) ≥ 0 ◮ For each p , E data ( φ 1 p , φ 2 p ) interaction between two binary p variables. Linear system has solution iff E data ( φ 1 p , φ 2 p ) is p submodular.

  69. Log representation - minimization by graph cuts Graph construction ◮ Linear system for finding edge weights  A ( p ) + B ( p ) = f 2 ( p )   C ( p ) + D ( p ) = f 3 ( p )  A ( p ) + E ( p ) + D ( p ) = f 1 ( p )   B ( p ) + F ( p ) + C ( p ) = f 4 ( p )  such that E ( p ) , F ( p ) ≥ 0 ◮ For each p , E data ( φ 1 p , φ 2 p ) interaction between two binary p variables. Linear system has solution iff E data ( φ 1 p , φ 2 p ) is p submodular.

  70. Log representation - minimization by graph cuts Graph construction ◮ Linear system for finding edge weights  A ( p ) + B ( p ) = f 2 ( p )   C ( p ) + D ( p ) = f 3 ( p )  A ( p ) + E ( p ) + D ( p ) = f 1 ( p )   B ( p ) + F ( p ) + C ( p ) = f 4 ( p )  such that E ( p ) , F ( p ) ≥ 0 ◮ For each p , E data ( φ 1 p , φ 2 p ) interaction between two binary p variables. Linear system has solution iff E data ( φ 1 p , φ 2 p ) is p submodular.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend