Multi-agent constrained optimization of a strongly convex function - - PowerPoint PPT Presentation

multi agent constrained optimization of a strongly convex
SMART_READER_LITE
LIVE PREVIEW

Multi-agent constrained optimization of a strongly convex function - - PowerPoint PPT Presentation

Multi-agent constrained optimization of a strongly convex function Necdet Serhat Aybat Industrial & Manufacturing Engineering Penn State University Joint work with Erfan Yazdandoost Hamedani Research supported by NSF CMMI-1635106 and ARO


slide-1
SLIDE 1

Multi-agent constrained optimization

  • f a strongly convex function

Necdet Serhat Aybat Industrial & Manufacturing Engineering Penn State University Joint work with Erfan Yazdandoost Hamedani

Research supported by NSF CMMI-1635106 and ARO W911NF-17-1-0298

slide-2
SLIDE 2

Motivation

Many real-life networks are large-scale composed of agents with local information agents willing to collaborate without sharing their private data This motivated huge interest in designing decentralized methods for

  • ptimization of multi-agent systems

2

slide-3
SLIDE 3

Motivation

Many real-life networks are large-scale composed of agents with local information agents willing to collaborate without sharing their private data This motivated huge interest in designing decentralized methods for

  • ptimization of multi-agent systems

Examples: Routing and congestion control in wired and wireless networks parameter estimation in sensor networks multi-agent cooperative control and coordination processing distributed big-data in (online) machine learning

2

slide-4
SLIDE 4

Decentralized Consensus Optimization

Compute an optimal solution for (P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi

3

slide-5
SLIDE 5

Decentralized Consensus Optimization

Compute an optimal solution for (P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi N = {1, . . . , N} processing nodes

3

slide-6
SLIDE 6

Decentralized Consensus Optimization

Compute an optimal solution for (P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi

ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et)

3

slide-7
SLIDE 7

Decentralized Consensus Optimization

Compute an optimal solution for (P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi

ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et)

Node i can transmit data to j at time t only if (i, j) ∈ Et

3

slide-8
SLIDE 8

Decentralized Consensus Optimization

Compute an optimal solution for (P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi

ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et)

Node i can transmit data to j at time t only if (i, j) ∈ Et

¯ ϕ(x) : strongly convex (¯ µ > 0) ϕi(x) ρi(x) + fi(x) locally known ( ¯ µ mini∈N µi ≥ 0) Xi {x : Gi(x) ∈ −Ki} locally known, Ki closed convex cone.

3

slide-9
SLIDE 9

Decentralized Consensus Optimization

Compute an optimal solution for (P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi

ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et)

Node i can transmit data to j at time t only if (i, j) ∈ Et

¯ ϕ(x) : strongly convex (¯ µ > 0) ϕi(x) ρi(x) + fi(x) locally known ( ¯ µ mini∈N µi ≥ 0)

fi : convex + Lipschitz continuous gradient (constant Li) ρi : convex (possibly non-smooth) + efficient prox-map proxρi(x) argminξ∈Rn

  • ρi(ξ) + 1

2ξ − x2 2

  • Xi {x : Gi(x) ∈ −Ki} locally known, Ki closed convex cone.

3

slide-10
SLIDE 10

Decentralized Consensus Optimization

Compute an optimal solution for (P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi

ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et)

Node i can transmit data to j at time t only if (i, j) ∈ Et

¯ ϕ(x) : strongly convex (¯ µ > 0) ϕi(x) ρi(x) + fi(x) locally known ( ¯ µ mini∈N µi ≥ 0)

fi : convex + Lipschitz continuous gradient (constant Li) ρi : convex (possibly non-smooth) + efficient prox-map proxρi(x) argminξ∈Rn

  • ρi(ξ) + 1

2ξ − x2 2

  • Xi {x : Gi(x) ∈ −Ki} locally known, Ki closed convex cone.

Gi : Ki-convex + Lip. cont. (CG) + Lip. cont. Jacobian ∇Gi (LG)

3

slide-11
SLIDE 11

Examples

Constrained Lasso min

x∈Rn{λx1 + Cx − d2 2 : Ax ≤ b},

K = −Rp

+

distributed data: Ci ∈ Rmi×n and di ∈ Rmi for i ∈ N C = [Ci]i∈N ∈ Rm×n, d = [di]i∈N ∈ Rm, m =

i∈N mi

ϕi(x) =

λ |N |x1 + Cix − di2 2 merely convex (mi < n)

¯ ϕ(x) =

i∈N ϕi(x) strongly convex when rank(C) = n (m ≥ n)

4

slide-12
SLIDE 12

Examples

Constrained Lasso min

x∈Rn{λx1 + Cx − d2 2 : Ax ≤ b},

K = −Rp

+

distributed data: Ci ∈ Rmi×n and di ∈ Rmi for i ∈ N C = [Ci]i∈N ∈ Rm×n, d = [di]i∈N ∈ Rm, m =

i∈N mi

ϕi(x) =

λ |N |x1 + Cix − di2 2 merely convex (mi < n)

¯ ϕ(x) =

i∈N ϕi(x) strongly convex when rank(C) = n (m ≥ n)

Closest point in the intersection min

x∈∩i∈N Xi

  • i∈N

x − ¯ x2

2

s.t. Gi(x) ∈ −Ki, i ∈ N.

4

slide-13
SLIDE 13

Related Work - Constrained

Chang, Nedich, Scaglione’14: primal-dual method – minx∈∩i∈N Xi F (

i∈N fi(x)) s.t. i∈N gi(x) ≤ 0

– F and fi smooth, Xi compact, and time-varying directed G – no rate result, can handle non-smooth constraints

5

slide-14
SLIDE 14

Related Work - Constrained

Chang, Nedich, Scaglione’14: primal-dual method – minx∈∩i∈N Xi F (

i∈N fi(x)) s.t. i∈N gi(x) ≤ 0

– F and fi smooth, Xi compact, and time-varying directed G – no rate result, can handle non-smooth constraints Núñez, Cortés’15: min

i∈N ϕi(ξi, x) s.t. i∈N gi(ξi, x) ≤ 0

– ϕi and gi convex; time-varying directed G – O(1/ √ k) rate for L(¯ ξk, ¯ xk, ¯ yk) − L(ξ∗, x∗, y∗) – no error bounds on infeasibility, and suboptimality

5

slide-15
SLIDE 15

Related Work - Constrained

Aybat, Yazdandoost Hamedani’16: primal-dual method – min

i∈N ϕi(x) s.t. Aix − bi ∈ Ki, i ∈ N

– time-varying undirected and directed G – O(1/k) ergodic rate for infeasibility, and suboptimality – Convergence of the primal-dual iterate sequence (without rate)

6

slide-16
SLIDE 16

Related Work - Constrained

Aybat, Yazdandoost Hamedani’16: primal-dual method – min

i∈N ϕi(x) s.t. Aix − bi ∈ Ki, i ∈ N

– time-varying undirected and directed G – O(1/k) ergodic rate for infeasibility, and suboptimality – Convergence of the primal-dual iterate sequence (without rate) Chang’16: (primal-dual method) – minxi∈Xi

  • i∈N ρi(xi) + fi(Cixi) s.t.

i∈N Aixi = b

– fi smooth and strongly convex; time-varying undirected G – O(1/k) ergodic rate

6

slide-17
SLIDE 17

Notation

.: Euclidean norm σS(.): Support function of set S, σS(θ) supw∈S θ, w

7

slide-18
SLIDE 18

Notation

.: Euclidean norm σS(.): Support function of set S, σS(θ) supw∈S θ, w PS(w) argmin{v − w : v ∈ S}: Projection onto S dS(w) PS(w) − w: Distance function K∗: Dual cone of K, K◦: Polar cone of K, K◦ {θ ∈ Rm : θ, w ≤ 0 ∀w ∈ K},

7

slide-19
SLIDE 19

Notation

.: Euclidean norm σS(.): Support function of set S, σS(θ) supw∈S θ, w PS(w) argmin{v − w : v ∈ S}: Projection onto S dS(w) PS(w) − w: Distance function K∗: Dual cone of K, K◦: Polar cone of K, K◦ {θ ∈ Rm : θ, w ≤ 0 ∀w ∈ K}, ⊗: Kronecker product Π: Cartesian product

7

slide-20
SLIDE 20

Preliminaries: Primal-dual Algorithm (PDA)

PDA for convex-concave saddle-point problem by Chambolle and Pock’16

min

x∈X max y∈Y L(x, y) Φ(x) + Tx, y − h(y),

  • Φ(x) ρ(x) + f(x) strongly convex (µ > 0), h convex, T linear map

8

slide-21
SLIDE 21

Preliminaries: Primal-dual Algorithm (PDA)

PDA for convex-concave saddle-point problem by Chambolle and Pock’16

min

x∈X max y∈Y L(x, y) Φ(x) + Tx, y − h(y),

  • Φ(x) ρ(x) + f(x) strongly convex (µ > 0), h convex, T linear map

PDA:

yk+1 ← argmin

y

h(y) − T xk + ηk(xk − xk−1) , y + Dk(y, yk), xk+1 ← argmin

x

ρ(x) + f(xk) + ∇f(xk), x − xk + Tx, yk+1 + 1 2τ k x − xk,

  • Dk is Bregman distance function Dk(y, ¯

y) ≥

1 2κk y − ¯

y2

8

slide-22
SLIDE 22

Preliminaries: Primal-dual Algorithm (PDA)

PDA for convex-concave saddle-point problem by Chambolle and Pock’16

min

x∈X max y∈Y L(x, y) Φ(x) + Tx, y − h(y),

  • Φ(x) ρ(x) + f(x) strongly convex (µ > 0), h convex, T linear map

PDA:

yk+1 ← argmin

y

h(y) − T xk + ηk(xk − xk−1) , y + Dk(y, yk), xk+1 ← argmin

x

ρ(x) + f(xk) + ∇f(xk), x − xk + Tx, yk+1 + 1 2τ k x − xk,

  • Dk is Bregman distance function Dk(y, ¯

y) ≥

1 2κk y − ¯

y2 Theorem: If τ k, κk, ηk > 0 such that

1 τ k − µ ≥ L + T2κk,

κk = κk+1ηk+1 and ηk+1 ≥ τ k

1 τ k+1 − µ

for all k ≥ 0 then L(¯ xK, y) − L(x, ¯ yK) ≤ 1 NK 1 2τ 0 x − x02 + D0(y, y0)

  • ∀x, y

where NK = K

k=1 κk κ0 = O(1/K2) and (¯

xK, ¯ yK) N−1

K

K

k=1 κk κ0 (xk, yk).

8

slide-23
SLIDE 23

Extension to nonlinear constraints

Consider a more general convex-concave saddle-point problem

min

x∈X max y∈Y L(x, y) Φ(x) + G(x), y − h(y),

  • G is K-convex, Lipschitz continuous (CG)
  • Jacobian ∇G is Lipschitz continuous (LG)

9

slide-24
SLIDE 24

Extension to nonlinear constraints

Consider a more general convex-concave saddle-point problem

min

x∈X max y∈Y L(x, y) Φ(x) + G(x), y − h(y),

  • G is K-convex, Lipschitz continuous (CG)
  • Jacobian ∇G is Lipschitz continuous (LG)

General PDA

yk+1 ← argmin

y

h(y) − G(xk) + ηk G(xk) − G(xk−1) , y + Dk(y, yk) xk+1 ← argmin

x

ρ(x) + f(xk) + ∇f(xk), x − xk + ∇G(xk)x, yk+1 + 1 2τ k x − xk2

9

slide-25
SLIDE 25

Extension to nonlinear constraints

Consider a more general convex-concave saddle-point problem

min

x∈X max y∈Y L(x, y) Φ(x) + G(x), y − h(y),

  • G is K-convex, Lipschitz continuous (CG)
  • Jacobian ∇G is Lipschitz continuous (LG)

General PDA

yk+1 ← argmin

y

h(y) − G(xk) + ηk G(xk) − G(xk−1) , y + Dk(y, yk) xk+1 ← argmin

x

ρ(x) + f(xk) + ∇f(xk), x − xk + ∇G(xk)x, yk+1 + 1 2τ k x − xk2

Theorem: If τ k, κk, ηk > 0 such that

1 τ k − µ ≥ L + 2C2 Gκk + LGyk+1,

κk = κk+1ηk+1 and ηk+1 ≥ τ k(

1 τ k+1 − µ) for all k ≥ 0 then

L(¯ xK, y) − L(x, ¯ yK) ≤ 1 NK 1 2τ 0 x − x02 + D0(y, y0)

  • ∀x, y

where (¯ xK, ¯ yK) N−1

K

K

k=1 κk κ0 (xk, yk), and NK = K k=1 κk κ0 = O(1/K2).

9

slide-26
SLIDE 26

Extension to nonlinear constraints

We also obtain the following bounds:

1 2x∗ − xK2 ≤ κ0 τ K κK 1 2τ 0 x∗ − x02 + D0(y∗, y0)

  • yK ≤ y∗ +
  • 4κ0

1 2τ 0 x∗ − x02 + D0(y∗, y0)

  • and τ K/κK = O(1/K2).

10

slide-27
SLIDE 27

Extension to nonlinear constraints

We also obtain the following bounds:

1 2x∗ − xK2 ≤ κ0 τ K κK 1 2τ 0 x∗ − x02 + D0(y∗, y0)

  • yK ≤ y∗ +
  • 4κ0

1 2τ 0 x∗ − x02 + D0(y∗, y0)

  • and τ K/κK = O(1/K2).

Specific stepsize sequence:

Initialization: η0 = 0, κ0 = 1 For k ≥ 0 : τ k = 1 2C2

Gκk + LGyk+1 + L + µ

ηk+1 =

  • 1 − µτ k,

κk+1 = κk/ηk+1

10

slide-28
SLIDE 28

Comparison with related works

Our proximal gradient primal-dual Alg.

minx∈X maxy∈Y Φ(x) + G(x), y − h(y) Φ composite strongly convex, h convex, G is K-convex, Lipschitz, such that ∇G is Lipschitz continuous L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2) x∗ − xK2 ≤ O(1/K2) and yK ≤ O(1)

11

slide-29
SLIDE 29

Comparison with related works

Our proximal gradient primal-dual Alg.

minx∈X maxy∈Y Φ(x) + G(x), y − h(y) Φ composite strongly convex, h convex, G is K-convex, Lipschitz, such that ∇G is Lipschitz continuous L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2) x∗ − xK2 ≤ O(1/K2) and yK ≤ O(1)

Proximal gradient primal-dual alg. by Chambolle and Pock’16

minx∈X maxy∈Y Φ(x) + Tx, y − h(y) Φ composite strongly convex, h convex, and T linear map L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2) x∗ − xK2 ≤ O(1/K2)

11

slide-30
SLIDE 30

Comparison with related works

Our proximal gradient primal-dual Alg.

minx∈X maxy∈Y Φ(x) + G(x), y − h(y) Φ composite strongly convex, h convex, G is K-convex, Lipschitz, such that ∇G is Lipschitz continuous L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2) x∗ − xK2 ≤ O(1/K2) and yK ≤ O(1)

Proximal gradient primal-dual alg. by Chambolle and Pock’16

minx∈X maxy∈Y Φ(x) + Tx, y − h(y) Φ composite strongly convex, h convex, and T linear map L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2) x∗ − xK2 ≤ O(1/K2)

Proximal gradient primal-dual alg. by Yu and Neely’17

minx f(x) s.t. G(x) ≤ 0 f composite convex, G composite convex and Lipschitz continuous f(¯ xK) − f(x∗) ≤ O(1/K) and G(¯ xK) ≤ O(1/K)

11

slide-31
SLIDE 31

Comparison with related works

mirror-prox for minx maxy φ(x, y) (Nemirovski’04)

φ(x, y) is convex in x and concave in y, ∇φ is Lipschitz continuous in (x, y) φ(¯ xK, y) − φ(x, ¯ yK) ≤ O(1/K)

12

slide-32
SLIDE 32

Comparison with related works

mirror-prox for minx maxy φ(x, y) (Nemirovski’04)

φ(x, y) is convex in x and concave in y, ∇φ is Lipschitz continuous in (x, y) φ(¯ xK, y) − φ(x, ¯ yK) ≤ O(1/K)

primal-dual algorithm with linesearch (Malitsky and Pock’16)

minx∈X maxy∈Y Φ(x) + Tx, y − h(y) Φ and h∗ are convex and at least one is strongly convex Proximal primal and dual steps with linesearch determining stepsizes and accelerating parameter L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2)

12

slide-33
SLIDE 33

Comparison with related works

mirror-prox for minx maxy φ(x, y) (Nemirovski’04)

φ(x, y) is convex in x and concave in y, ∇φ is Lipschitz continuous in (x, y) φ(¯ xK, y) − φ(x, ¯ yK) ≤ O(1/K)

primal-dual algorithm with linesearch (Malitsky and Pock’16)

minx∈X maxy∈Y Φ(x) + Tx, y − h(y) Φ and h∗ are convex and at least one is strongly convex Proximal primal and dual steps with linesearch determining stepsizes and accelerating parameter L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2)

Accelerated primal-dual algorithm (Chen, Lan and Ouyang’13)

minx∈X maxy∈Y Φ(x) + Tx, y − h(y) Φ convex function with Lipshcitz continuous gradient, h convex and T linear map Proximal-gradient steps with several accelerating steps L(¯ xK, y) − L(x, ¯ yK) ≤ O(1/K2 + 1/K)

12

slide-34
SLIDE 34

Decentralized Consensus Optimization

Compute an optimal solution for

(P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et) ¯ ϕ(x) : strongly convex (¯ µ > 0) ϕi(x) ρi(x) + fi(x) locally known ( ¯ µ mini∈N µi ≥ 0) Xi {x : Gi(x) ∈ −Ki} locally known, Ki closed convex cone.

13

slide-35
SLIDE 35

Decentralized Consensus Optimization

Compute an optimal solution for

(P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et) ¯ ϕ(x) : strongly convex (¯ µ > 0) ϕi(x) ρi(x) + fi(x) locally known ( ¯ µ mini∈N µi ≥ 0)

fi : convex + Lipschitz continuous gradient (constant Li) ρi : convex (possibly non-smooth) + efficient prox-map proxρi(x) argminξ∈Rn

  • ρi(ξ) + 1

2ξ − x2 2

  • Xi {x : Gi(x) ∈ −Ki} locally known, Ki closed convex cone.

13

slide-36
SLIDE 36

Decentralized Consensus Optimization

Compute an optimal solution for

(P) : min

x

¯ ϕ(x)

  • i∈N

ϕi(x) s.t. x ∈

  • i∈N

Xi ϕ1(x) ϕ2(x) ϕ5(x) ϕ4(x) ϕ3(x)

N = {1, . . . , N} processing nodes on a time-varying Gt = (N, Et) ¯ ϕ(x) : strongly convex (¯ µ > 0) ϕi(x) ρi(x) + fi(x) locally known ( ¯ µ mini∈N µi ≥ 0)

fi : convex + Lipschitz continuous gradient (constant Li) ρi : convex (possibly non-smooth) + efficient prox-map proxρi(x) argminξ∈Rn

  • ρi(ξ) + 1

2ξ − x2 2

  • Xi {x : Gi(x) ∈ −Ki} locally known, Ki closed convex cone.

Gi : Ki-convex + Lip. cont. (CG) + Lip. cont. Jacobian ∇Gi (LG)

13

slide-37
SLIDE 37

Methodology: Time-varying Topology

Suppose ¯ f strongly convex, and fi’s merely convex, i.e., ¯ f(x) =

i∈N fi(x) strongly convex (¯

µ > 0) ¯ µ = mini∈N µi = 0

14

slide-38
SLIDE 38

Methodology: Time-varying Topology

Suppose ¯ f strongly convex, and fi’s merely convex, i.e., ¯ f(x) =

i∈N fi(x) strongly convex (¯

µ > 0) ¯ µ = mini∈N µi = 0 Define the consensus cone C: C {x = [xi]i∈N ∈ Rn|N| : ∃¯ x ∈ Rn s.t. xi = ¯ x ∀ i ∈ N} f(x)

i∈N fi(xi) is strongly convex on C

14

slide-39
SLIDE 39

Methodology: Time-varying Topology

Suppose ¯ f strongly convex, and fi’s merely convex, i.e., ¯ f(x) =

i∈N fi(x) strongly convex (¯

µ > 0) ¯ µ = mini∈N µi = 0 Define the consensus cone C: C {x = [xi]i∈N ∈ Rn|N| : ∃¯ x ∈ Rn s.t. xi = ¯ x ∀ i ∈ N} f(x)

i∈N fi(xi) is strongly convex on C

Lemma: Let fα(x) f(x) + α

2 d2 C(x). Then fα is strongly convex with

µα ¯ µ/|N| + α 2 − ¯ µ/|N| − α 2 2 + 4¯ L2 > 0

for any α > 4

¯ µ|N|¯

L2, where ¯ L =

  • i∈N L2

i

|N|

.

14

slide-40
SLIDE 40

Methodology: Time-varying Topology

Suppose ¯ f strongly convex, and fi’s merely convex, i.e., ¯ f(x) =

i∈N fi(x) strongly convex (¯

µ > 0) ¯ µ = mini∈N µi = 0 Define the consensus cone C: C {x = [xi]i∈N ∈ Rn|N| : ∃¯ x ∈ Rn s.t. xi = ¯ x ∀ i ∈ N} f(x)

i∈N fi(xi) is strongly convex on C

Lemma: Let fα(x) f(x) + α

2 d2 C(x). Then fα is strongly convex with

µα ¯ µ/|N| + α 2 − ¯ µ/|N| − α 2 2 + 4¯ L2 > 0

for any α > 4

¯ µ|N|¯

L2, where ¯ L =

  • i∈N L2

i

|N|

. Note: The result in (Shi et al.’15) uses mixing matrices (static G)

14

slide-41
SLIDE 41

Methodology: Time-varying Topology

Let ρ(x)

i∈N ρi(xi),

f(x)

i∈N fi(xi),

fα(x) f(x) + α

2 d2 C(x)

15

slide-42
SLIDE 42

Methodology: Time-varying Topology

Let ρ(x)

i∈N ρi(xi),

f(x)

i∈N fi(xi),

fα(x) f(x) + α

2 d2 C(x)

xi ∈ Rn: local copy of x at i ∈ N, and x [xi]i∈N ∈ Rn|N|

15

slide-43
SLIDE 43

Methodology: Time-varying Topology

Let ρ(x)

i∈N ρi(xi),

f(x)

i∈N fi(xi),

fα(x) f(x) + α

2 d2 C(x)

xi ∈ Rn: local copy of x at i ∈ N, and x [xi]i∈N ∈ Rn|N| ∆i maxxi,x′

i∈dom ϕi x − x′,

∆ maxi∈N ∆i < ∞

15

slide-44
SLIDE 44

Methodology: Time-varying Topology

Let ρ(x)

i∈N ρi(xi),

f(x)

i∈N fi(xi),

fα(x) f(x) + α

2 d2 C(x)

xi ∈ Rn: local copy of x at i ∈ N, and x [xi]i∈N ∈ Rn|N| ∆i maxxi,x′

i∈dom ϕi x − x′,

∆ maxi∈N ∆i < ∞ B0 {x ∈ Rm : x ≤ 2∆}, B Πi∈N B0,

  • C C ∩ B

15

slide-45
SLIDE 45

Methodology: Time-varying Topology

Let ρ(x)

i∈N ρi(xi),

f(x)

i∈N fi(xi),

fα(x) f(x) + α

2 d2 C(x)

xi ∈ Rn: local copy of x at i ∈ N, and x [xi]i∈N ∈ Rn|N| ∆i maxxi,x′

i∈dom ϕi x − x′,

∆ maxi∈N ∆i < ∞ B0 {x ∈ Rm : x ≤ 2∆}, B Πi∈N B0,

  • C C ∩ B

For any α ≥ 0, an equivalent formulation:

min

x∈Rn

  • i∈N ρi(x) + fi(x)

s.t. Gi(x) ∈ −Ki, i ∈ N ≡ min x ∈ C ρ(x) + fα(x) s.t. Gi(xi) ∈ −Ki, i ∈ N

15

slide-46
SLIDE 46

Methodology: Time-varying Topology

Let ρ(x)

i∈N ρi(xi),

f(x)

i∈N fi(xi),

fα(x) f(x) + α

2 d2 C(x)

xi ∈ Rn: local copy of x at i ∈ N, and x [xi]i∈N ∈ Rn|N| ∆i maxxi,x′

i∈dom ϕi x − x′,

∆ maxi∈N ∆i < ∞ B0 {x ∈ Rm : x ≤ 2∆}, B Πi∈N B0,

  • C C ∩ B

For any α ≥ 0, an equivalent formulation:

min

x∈Rn

  • i∈N ρi(x) + fi(x)

s.t. Gi(x) ∈ −Ki, i ∈ N ≡ min x ∈ C ρ(x) + fα(x) s.t. Gi(xi) ∈ −Ki, i ∈ N

Saddle Point Formulation: min

x max y

L(x, y) ρ(x) + fα(x) + λ, x − σ

C (λ)

+

  • i∈N

θi, Gi(xi) − σ−Ki(θi) y = [θ⊤ λ⊤]⊤, θ = [θi]i∈N ∈ Rm and λ = [λi]i∈N ∈ Rn|N|

15

slide-47
SLIDE 47

Methodology: Time-varying Topology

  • Implementing PDA on the saddle-point formulation:

θk+1

i

← argmin

θi

σ−Ki(θi) − Gi(xk

i ) + ηk(Gi(xk i ) − Gi(xk−1 i

)), θi + 1 2κk θi − θk

i 2 2

λk+1 ← argmin

λ

σ

C (λ) − xk + ηk(xk − xk−1), λ +

1 2γk λ − λk2

2,

xk+1 ← argmin

x

ρ(x) + ∇fα(xk), x + ∇G(xk)x, θk+1 + x, λk+1 + 1 2τ k x − xk2

16

slide-48
SLIDE 48

Methodology: Time-varying Topology

  • Implementing PDA on the saddle-point formulation:

θk+1

i

← argmin

θi

σ−Ki(θi) − Gi(xk

i ) + ηk(Gi(xk i ) − Gi(xk−1 i

)), θi + 1 2κk θi − θk

i 2 2

λk+1 ← argmin

λ

σ

C (λ) − xk + ηk(xk − xk−1), λ +

1 2γk λ − λk2

2,

xk+1 ← argmin

x

ρ(x) + ∇fα(xk), x + ∇G(xk)x, θk+1 + x, λk+1 + 1 2τ k x − xk2 Note: λ, x updates require PC(ω) = 1|N | ⊗

1 |N |

  • i∈N ωi and P

C(ω) = PB(PC(ω)) 16

slide-49
SLIDE 49

Methodology: Time-varying Topology

  • Implementing PDA on the saddle-point formulation:

θk+1

i

← argmin

θi

σ−Ki(θi) − Gi(xk

i ) + ηk(Gi(xk i ) − Gi(xk−1 i

)), θi + 1 2κk θi − θk

i 2 2

λk+1 ← argmin

λ

σ

C (λ) − xk + ηk(xk − xk−1), λ +

1 2γk λ − λk2

2,

xk+1 ← argmin

x

ρ(x) + ∇fα(xk), x + ∇G(xk)x, θk+1 + x, λk+1 + 1 2τ k x − xk2 Note: λ, x updates require PC(ω) = 1|N | ⊗

1 |N |

  • i∈N ωi and P

C(ω) = PB(PC(ω))

  • λ update: λk+1 = γk

ωk − P

C(ωk)

  • , ωk

1 γk λk + xk + ηk(xk − xk−1)

  • x update: ∇fα(xk) = f(xk) + α(xk − PC(xk))

16

slide-50
SLIDE 50

Methodology: Time-varying Topology

  • Implementing PDA on the saddle-point formulation:

θk+1

i

← argmin

θi

σ−Ki(θi) − Gi(xk

i ) + ηk(Gi(xk i ) − Gi(xk−1 i

)), θi + 1 2κk θi − θk

i 2 2

λk+1 ← argmin

λ

σ

C (λ) − xk + ηk(xk − xk−1), λ +

1 2γk λ − λk2

2,

xk+1 ← argmin

x

ρ(x) + ∇fα(xk), x + ∇G(xk)x, θk+1 + x, λk+1 + 1 2τ k x − xk2 Note: λ, x updates require PC(ω) = 1|N | ⊗

1 |N |

  • i∈N ωi and P

C(ω) = PB(PC(ω))

  • λ update: λk+1 = γk

ωk − P

C(ωk)

  • , ωk

1 γk λk + xk + ηk(xk − xk−1)

  • x update: ∇fα(xk) = f(xk) + α(xk − PC(xk))
  • Approximate averaging operator Rk(ω) = [Rk

i (ω)]i∈N s.t.

PC(ω) − Rk(ω) = O(βqkω) ∀ω for some β ∈ (0, 1), increasing {qk}k≥0

16

slide-51
SLIDE 51

Methodology: Time-varying Topology

  • Implementing PDA on the saddle-point formulation:

θk+1

i

← argmin

θi

σ−Ki(θi) − Gi(xk

i ) + ηk(Gi(xk i ) − Gi(xk−1 i

)), θi + 1 2κk θi − θk

i 2 2

λk+1 ← argmin

λ

σ

C (λ) − xk + ηk(xk − xk−1), λ +

1 2γk λ − λk2

2,

xk+1 ← argmin

x

ρ(x) + ∇fα(xk), x + ∇G(xk)x, θk+1 + x, λk+1 + 1 2τ k x − xk2 Note: λ, x updates require PC(ω) = 1|N | ⊗

1 |N |

  • i∈N ωi and P

C(ω) = PB(PC(ω))

  • λ update: λk+1 = γk

ωk − P

C(ωk)

  • , ωk

1 γk λk + xk + ηk(xk − xk−1)

  • x update: ∇fα(xk) = f(xk) + α(xk − PC(xk))
  • Approximate averaging operator Rk(ω) = [Rk

i (ω)]i∈N s.t.

PC(ω) − Rk(ω) = O(βqkω) ∀ω for some β ∈ (0, 1), increasing {qk}k≥0 λk+1 ← γk ωk − PB

  • Rk(ωk)
  • ,

xk+1 ← proxτ kρ

  • xk − τ ksk

, sk ← ∇f(xk) + ∇G(xk)⊤θk+1 + λk+1 + α

  • xk − Rk(xk)
  • 16
slide-52
SLIDE 52

Methodology: Time-varying Topology

If ¯ µ > 0, then α ← 0 and µ ← ¯ µ; else, α > 4

¯ µ

  • i∈N L2

i and µ ← µα

Algorithm DPDA-TV ( x0, θ0, α, δ, γ, µ ) Initialization: x−1 ← x0, s0 ← 0, δ, γ > 0, γ0 ← γ, µ ∈ (0, max{ ¯ µ, µα}], η0 ← 0, κ0 ← γ

δ 2C2

G

i ∈ N Step k: (k ≥ 0, i ∈ N)

  • 1. θk+1

i

← PK∗

i

  • θk

i + κk

Gi(xk

i ) + ηk(Gi(xk i ) − Gi(xk−1 i

))

  • ,
  • 2. ωk

i ← 1 γk λk i + xk i + ηk(xk i − xk−1 i

),

  • 3. λk+1

i

← γkωk

i − γkPB0

  • Rk

i

  • ωk

,

  • 5. sk

i ← ∇fi(xk i ) + ∇Gi(xk i )⊤θk+1 i

+ λk+1

i

+ α

  • xk

i − Rk i (xk)

  • ,
  • 4. xk+1

i

← proxτ kρi

  • xk

i − τ ksk i

  • ,
  • 5. τ k+1, ηk+1, γk+1, κk+1 update by step-size condition rule

17

slide-53
SLIDE 53

Methodology: Time-varying Topology

If ¯ µ > 0, then α ← 0 and µ ← ¯ µ; else, α > 4

¯ µ

  • i∈N L2

i and µ ← µα

  • Step-size condition: given δ > 0, choose τ k, ηk, κk, γk > 0 such that

ηk+1 ≥ τ k

  • 1

τ k+1 − µ

  • ,

1 τ k − (Li + α + µ + LGθk+1

i

) ≥ γkηk+1(2 + δ) 2κkC2

G

γk ≤ δ, κk+1 ≥ κk ηk+1 , γk+1 = γk ηk+1

  • A possible way of choosing:

Initialization: η0 ← 0, γ0 ← γ, κ0 ← γ δ 2C2

G

For k ≥ 0 : ˜ τ k ← 1 γk(2 + δ) + Lmax + α + LG maxi∈N θk+1

i

  • ,

τ k ← ( 1 ˜ τ k + µ)−1 γk+1 ← γk 1 + µ˜ τ k, ηk+1 ← γk/γk+1, κk+1 ← γk+1 δ 2C2

G

We have ηk ∈ (0, 1), τ k = Θ(1/k), γk = Θ(k), κk = Θ(k)

18

slide-54
SLIDE 54

Main result

Theorem: Let λ0 = 0, θ0 = 0. Suppose step-size condition holds. qk ≥ ⌈(5 + c) log1/β(k + 1)⌉ communication rounds at iteration-k. Then xk → x∗ such that x∗ = 1|N| ⊗ x∗, i.e., x∗

i = x∗ for i ∈ N.

Within O(K log1/β(K)) communication rounds,

19

slide-55
SLIDE 55

Main result

Theorem: Let λ0 = 0, θ0 = 0. Suppose step-size condition holds. qk ≥ ⌈(5 + c) log1/β(k + 1)⌉ communication rounds at iteration-k. Then xk → x∗ such that x∗ = 1|N| ⊗ x∗, i.e., x∗

i = x∗ for i ∈ N.

Within O(K log1/β(K)) communication rounds, Dual iterate bound: θK

i ≤ θ∗ i +

  • δγ0

C2

G Λ(K),

∀i ∈ N

19

slide-56
SLIDE 56

Main result

Theorem: Let λ0 = 0, θ0 = 0. Suppose step-size condition holds. qk ≥ ⌈(5 + c) log1/β(k + 1)⌉ communication rounds at iteration-k. Then xk → x∗ such that x∗ = 1|N| ⊗ x∗, i.e., x∗

i = x∗ for i ∈ N.

Within O(K log1/β(K)) communication rounds, Dual iterate bound: θK

i ≤ θ∗ i +

  • δγ0

C2

G Λ(K),

∀i ∈ N Suboptimality: |ϕ(¯ xK) − ϕ(x∗)| ≤ Λ(K)

NK , where ϕ(x) = ρ(x) + f(x)

19

slide-57
SLIDE 57

Main result

Theorem: Let λ0 = 0, θ0 = 0. Suppose step-size condition holds. qk ≥ ⌈(5 + c) log1/β(k + 1)⌉ communication rounds at iteration-k. Then xk → x∗ such that x∗ = 1|N| ⊗ x∗, i.e., x∗

i = x∗ for i ∈ N.

Within O(K log1/β(K)) communication rounds, Dual iterate bound: θK

i ≤ θ∗ i +

  • δγ0

C2

G Λ(K),

∀i ∈ N Suboptimality: |ϕ(¯ xK) − ϕ(x∗)| ≤ Λ(K)

NK , where ϕ(x) = ρ(x) + f(x)

Infeasiblity: dC(¯ xK) +

i∈N θ∗ i d−Ki(Gi(¯

xK

i )) ≤ Λ(K) NK

19

slide-58
SLIDE 58

Main result

Theorem: Let λ0 = 0, θ0 = 0. Suppose step-size condition holds. qk ≥ ⌈(5 + c) log1/β(k + 1)⌉ communication rounds at iteration-k. Then xk → x∗ such that x∗ = 1|N| ⊗ x∗, i.e., x∗

i = x∗ for i ∈ N.

Within O(K log1/β(K)) communication rounds, Dual iterate bound: θK

i ≤ θ∗ i +

  • δγ0

C2

G Λ(K),

∀i ∈ N Suboptimality: |ϕ(¯ xK) − ϕ(x∗)| ≤ Λ(K)

NK , where ϕ(x) = ρ(x) + f(x)

Infeasiblity: dC(¯ xK) +

i∈N θ∗ i d−Ki(Gi(¯

xK

i )) ≤ Λ(K) NK

Solution error: xK − x∗2 ≤ 2γ0 ˜

τ K γK Λ(K)

19

slide-59
SLIDE 59

Main result

Theorem: Let λ0 = 0, θ0 = 0. Suppose step-size condition holds. qk ≥ ⌈(5 + c) log1/β(k + 1)⌉ communication rounds at iteration-k. Then xk → x∗ such that x∗ = 1|N| ⊗ x∗, i.e., x∗

i = x∗ for i ∈ N.

Within O(K log1/β(K)) communication rounds, Dual iterate bound: θK

i ≤ θ∗ i +

  • δγ0

C2

G Λ(K),

∀i ∈ N Suboptimality: |ϕ(¯ xK) − ϕ(x∗)| ≤ Λ(K)

NK , where ϕ(x) = ρ(x) + f(x)

Infeasiblity: dC(¯ xK) +

i∈N θ∗ i d−Ki(Gi(¯

xK

i )) ≤ Λ(K) NK

Solution error: xK − x∗2 ≤ 2γ0 ˜

τ K γK Λ(K)

Λ(K) = O(K

k=1 βqk−1k4) and supK≥1 Λ(K) < ∞

NK = O(K2), ˜ τ K/γK = O(1/K2) and κK/γK = O(1)

19

slide-60
SLIDE 60

Main result

Theorem: Let λ0 = 0, θ0 = 0. Suppose step-size condition holds. qk ≥ ⌈(5 + c) log1/β(k + 1)⌉ communication rounds at iteration-k. Then xk → x∗ such that x∗ = 1|N| ⊗ x∗, i.e., x∗

i = x∗ for i ∈ N.

Within O(K log1/β(K)) communication rounds, Dual iterate bound: θK

i ≤ θ∗ i +

  • δγ0

C2

G Λ(K),

∀i ∈ N Suboptimality: |ϕ(¯ xK) − ϕ(x∗)| ≤ Λ(K)

NK , where ϕ(x) = ρ(x) + f(x)

Infeasiblity: dC(¯ xK) +

i∈N θ∗ i d−Ki(Gi(¯

xK

i )) ≤ Λ(K) NK

Solution error: xK − x∗2 ≤ 2γ0 ˜

τ K γK Λ(K)

Λ(K) = O(K

k=1 βqk−1k4) and supK≥1 Λ(K) < ∞

NK = O(K2), ˜ τ K/γK = O(1/K2) and κK/γK = O(1) Note: For static undirected G, qk = 1, Λ(K) ≤ |N |∆

2τ 0 + 1 2κ0

  • i∈N θ0

i − θ∗ i 2

19

slide-61
SLIDE 61

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15

20

slide-62
SLIDE 62

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors

20

slide-63
SLIDE 63

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors communication among neighbors occurs instantaneously

20

slide-64
SLIDE 64

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors communication among neighbors occurs instantaneously Synchronous nodes: one communication round in every unit time

20

slide-65
SLIDE 65

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors communication among neighbors occurs instantaneously Synchronous nodes: one communication round in every unit time qk: # of communication rounds within the k-th DPDA iteration

20

slide-66
SLIDE 66

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors communication among neighbors occurs instantaneously Synchronous nodes: one communication round in every unit time qk: # of communication rounds within the k-th DPDA iteration tk: # of communication rounds before the k-th DPDA iteration

20

slide-67
SLIDE 67

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors communication among neighbors occurs instantaneously Synchronous nodes: one communication round in every unit time qk: # of communication rounds within the k-th DPDA iteration tk: # of communication rounds before the k-th DPDA iteration V t ∈ R|N|×|N| be a mixing matrix for Gt = (N, Et):

20

slide-68
SLIDE 68

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors communication among neighbors occurs instantaneously Synchronous nodes: one communication round in every unit time qk: # of communication rounds within the k-th DPDA iteration tk: # of communication rounds before the k-th DPDA iteration V t ∈ R|N|×|N| be a mixing matrix for Gt = (N, Et):

∀ i ∈ N, V t

ij = 0

⇔ j ∈ N t

i

  • j∈N t

i V t

ijxt j can be computed at i ∈ N with local communication

20

slide-69
SLIDE 69

Approximate Averaging Operator Rk(·)

We adopt the following information exchange models Undirected Gt = (N, Et): Nedich & Ozdaglar’09 and Chen & Ozdaglar’12 Directed Gt = (N, Et): Nedich & Olshevsky’15 In one communication round, every node talks to its neighbors communication among neighbors occurs instantaneously Synchronous nodes: one communication round in every unit time qk: # of communication rounds within the k-th DPDA iteration tk: # of communication rounds before the k-th DPDA iteration V t ∈ R|N|×|N| be a mixing matrix for Gt = (N, Et):

∀ i ∈ N, V t

ij = 0

⇔ j ∈ N t

i

  • j∈N t

i V t

ijxt j can be computed at i ∈ N with local communication

W t,s V tV t−1...V s+1 for t ≥ s + 1

20

slide-70
SLIDE 70

Rk(·) for directed communication networks

Rk(w) = [Rk

i (w)]i∈N s.t. PC(w) − Rk(w) = O(βqkw)

∀w, k ≥ 0 Definition: N t,out

i

{j ∈ N : (i, j) ∈ Et} ∪ {i} and dt

i |N t,out i

| N t,in

i

{j ∈ N : (j, i) ∈ Et} ∪ {i} V t ∈ R|N|×|N| s.t. V t

i,j = 1/dt i for j ∈ N t,in i

and 0 otherwise.

21

slide-71
SLIDE 71

Rk(·) for directed communication networks

Rk(w) = [Rk

i (w)]i∈N s.t. PC(w) − Rk(w) = O(βqkw)

∀w, k ≥ 0 Definition: N t,out

i

{j ∈ N : (i, j) ∈ Et} ∪ {i} and dt

i |N t,out i

| N t,in

i

{j ∈ N : (j, i) ∈ Et} ∪ {i} V t ∈ R|N|×|N| s.t. V t

i,j = 1/dt i for j ∈ N t,in i

and 0 otherwise. Assumption: ∃ M > 1 s.t. (N, Ek,M) is strongly connected for k ≥ 1, Ek,M (k+1)M−1

t=kM

Et.

21

slide-72
SLIDE 72

Rk(·) for directed communication networks

Rk(w) = [Rk

i (w)]i∈N s.t. PC(w) − Rk(w) = O(βqkw)

∀w, k ≥ 0 Definition: N t,out

i

{j ∈ N : (i, j) ∈ Et} ∪ {i} and dt

i |N t,out i

| N t,in

i

{j ∈ N : (j, i) ∈ Et} ∪ {i} V t ∈ R|N|×|N| s.t. V t

i,j = 1/dt i for j ∈ N t,in i

and 0 otherwise. Assumption: ∃ M > 1 s.t. (N, Ek,M) is strongly connected for k ≥ 1, Ek,M (k+1)M−1

t=kM

Et. Lemma: ∃β ∈ (0, 1) and Γ > 0 s.t. for any t > s ≥ 0 and w = [wi]i∈N

  • (diag

W t,s1 −1W t,s ⊗ Im) w −

1 |N|

  • i∈N

wi

  • ≤ Γβt−sw

21

slide-73
SLIDE 73

Rk(·) for directed communication networks

Rk(w) = [Rk

i (w)]i∈N s.t. PC(w) − Rk(w) = O(βqkw)

∀w, k ≥ 0 Definition: N t,out

i

{j ∈ N : (i, j) ∈ Et} ∪ {i} and dt

i |N t,out i

| N t,in

i

{j ∈ N : (j, i) ∈ Et} ∪ {i} V t ∈ R|N|×|N| s.t. V t

i,j = 1/dt i for j ∈ N t,in i

and 0 otherwise. Assumption: ∃ M > 1 s.t. (N, Ek,M) is strongly connected for k ≥ 1, Ek,M (k+1)M−1

t=kM

Et. Lemma: ∃β ∈ (0, 1) and Γ > 0 s.t. for any t > s ≥ 0 and w = [wi]i∈N

  • (diag

W t,s1 −1W t,s ⊗ Im) w −

1 |N|

  • i∈N

wi

  • ≤ Γβt−sw

Hence, Rk(w) (diag

  • W tk+qk,tk1|N |

−1W tk+qk,tk ⊗ Im) w

21

slide-74
SLIDE 74

Numerical Experiments

Distributed Isotonic LASSO: x ∈ Rn|N|, C = [Ci]i∈N ∈ Rm|N|×n, d = [di]i∈N ∈ Rm|N|, and A ∈ Rn−1×n

min

x=[xi]i∈N ∈C, Axi≤0 i∈N

1 2

  • i∈N

Cixi − di2 + λ |N|

  • i∈N

xi1,

n = 20, m = n + 2 Random Ci with standard Gaussian entries, di = Ci(x◦ + ǫ) ǫ ∈ Rn random with Gaussian of zero mean and std. deviation 10−3 Random x◦ ∈ Rn−1: x◦ =

   U[−10, 0]5

  • first 5 components

, 0, 0, ..., 0,

  • n−11

, U[0, 10]5

  • last 5 components

  

  • ascending order

22

slide-75
SLIDE 75

Numerical Experiments

G0 = (N, E0) small-world network Gt = (N, Et) generated by sampling 80% of E0 s.t. M = 5 Effect of Network Topology (time-varying undirected):

23

slide-76
SLIDE 76

Numerical Experiments

Compared against DPDA-D (Aybat et al.’16) – O(1/k) ergodic rate Time-varying undirected Network: Gu = (N, Eu), |N| = 10, |Eu| = 45 Gt

u = (N, Et u) generated by sampling 80% of Eu s.t. M = 5

24

slide-77
SLIDE 77

Numerical Experiments

Compared against DPDA-D (Aybat et al.’16) – O(1/k) ergodic rate Time-varying directed Network: Gd = (N, Ed), |N| = 12, |Ed| = 24 Gt

d = (N, Et d) generated by sampling 80% of Ed s.t. M = 5

(Nedich et al.’17)

10 6 1 8 3 11 5 9 4 12 2 7

25