Multi-agent distributed optimization over networks and its - - PowerPoint PPT Presentation

multi agent distributed optimization over networks and
SMART_READER_LITE
LIVE PREVIEW

Multi-agent distributed optimization over networks and its - - PowerPoint PPT Presentation

Vistas in Control | ETH Zurich | September 10 - 11 2018 Multi-agent distributed optimization over networks and its application to energy systems Maria Prandini Introduction Robotic networks Social networks taken from AJGpr.com Transportation


slide-1
SLIDE 1

Vistas in Control | ETH Zurich | September 10 - 11 2018

Multi-agent distributed optimization over networks and its application to energy systems Maria Prandini

slide-2
SLIDE 2

Introduction

taken from AJGpr.com

Social networks Transportation systems Robotic networks Energy systems

2

slide-3
SLIDE 3

Introduction

Goal

Optimize the performance of the network

3

slide-4
SLIDE 4

Introduction

Goal

Optimize the performance of the network

Characteristics of the network

  • Large scale – System with multiple interacting components
  • Multi-agent – Components can perform computations, communicate with

each other, and cooperate to reach a common goal

  • Heterogeneous – Different physical or technological constraints per agent;

different objectives per agent

  • Uncertain – Endogenous and/or exogenous uncertainty affects the system

globally and/or locally

  • Combinatorial – Discrete and continuous decision variables

3

slide-5
SLIDE 5

Introduction

Challenges

  • Computation: Problem size too big, even combinatorial!
  • Communication: Not all communication links at place; link failures
  • Information privacy: Agents may not want to share information with

everyone

  • Uncertainty: Neglecting uncertainty may lead to an infeasible solution;

uncertainty often known through data

4

slide-6
SLIDE 6

Introduction

Challenges

  • Computation: Problem size too big, even combinatorial!
  • Communication: Not all communication links at place; link failures
  • Information privacy: Agents may not want to share information with

everyone

  • Uncertainty: Neglecting uncertainty may lead to an infeasible solution;

uncertainty often known through data

Distributed data-based optimization

Find an optimal solution by solving in parallel smaller optimization problems local to each agent while accounting for uncertainty known locally to each agent through data

4

slide-7
SLIDE 7

Introduction

Why go distributed?

  • 1. Scalable methodology
  • Communication: Only between neighbors, limited amount of info

exchanged

  • Computation: Only local; in parallel for all agents on a smaller

problem

5

slide-8
SLIDE 8

Introduction

Why go distributed?

  • 1. Scalable methodology
  • Communication: Only between neighbors, limited amount of info

exchanged

  • Computation: Only local; in parallel for all agents on a smaller

problem

  • 2. Resilience to communication failures

5

slide-9
SLIDE 9

Introduction

Why go distributed?

  • 1. Scalable methodology
  • Communication: Only between neighbors, limited amount of info

exchanged

  • Computation: Only local; in parallel for all agents on a smaller

problem

  • 2. Resilience to communication failures
  • 3. Information privacy
  • Agents do not reveal information about their preferences (encoded by
  • bjective and constraint functions) to each other

5

slide-10
SLIDE 10

Outline

  • 1. The deterministic case
  • Problem set-up
  • Distributed proximal algorithm
  • Analysis (assumptions + convergence)
  • Connection with other methods

6

slide-11
SLIDE 11

Outline

  • 1. The deterministic case
  • Problem set-up
  • Distributed proximal algorithm
  • Analysis (assumptions + convergence)
  • Connection with other methods
  • 2. The stochastic case
  • Problem set-up
  • Data-based approach
  • Distributed data-based implementation

6

slide-12
SLIDE 12

Outline

  • 1. The deterministic case
  • Problem set-up
  • Distributed proximal algorithm
  • Analysis (assumptions + convergence)
  • Connection with other methods
  • 2. The stochastic case
  • Problem set-up
  • Data-based approach
  • Distributed data-based implementation
  • 3. Constraint-coupled problem set-up
  • Distributed dual decomposition algorithm
  • Discrete case

6

slide-13
SLIDE 13

Outline

  • 1. The deterministic case
  • Problem set-up
  • Distributed proximal algorithm
  • Analysis (assumptions + convergence)
  • Connection with other methods
  • 2. The stochastic case
  • Problem set-up
  • Data-based approach
  • Distributed data-based implementation
  • 3. Constraint-coupled problem set-up
  • Distributed dual decomposition algorithm
  • Discrete case
  • 4. Summary & Future work

6

slide-14
SLIDE 14

Building district energy management

building

Set-up

  • Each building equipped with a chiller plant
  • Shared cooling network that acts as a thermal storage device

Goal Determine use of storage + zones temperature set-points to minimize the cost of the electrical energy consumption of the chillers in the district

7

slide-15
SLIDE 15

Building district energy management

  • 1. Chiller plant
  • Convert electrical energy into cooling energy
  • Characterized via COP (ratio between cooling energy and electrical

energy)

5 10 15 20 25 30 35 40

Echiller,c [MJ]

0.5 1 1.5 2

COP

Medium Small Large

8

slide-16
SLIDE 16

Building district energy management

  • 2. Building energy contribution
  • Walls-zones energy exchange – building thermal dynamics
  • Energy due to people occupancy
  • Zone thermal inertia
  • Other internal energy contribution, e.g. internal lighting, radiation

through windows

9

slide-17
SLIDE 17

Building district energy management

  • 2. Building energy contribution
  • Walls-zones energy exchange – building thermal dynamics
  • Energy due to people occupancy
  • Zone thermal inertia
  • Other internal energy contribution, e.g. internal lighting, radiation

through windows

  • 3. Thermal storage

S(k + 1) = αS(k) −

  • i

si(k)

  • S(k): Energy stored
  • si(k): Energy exchange between building i and storage

> 0: discharging the storage; < 0: charging

  • α: Energy losses coefficient

9

slide-18
SLIDE 18

Building district energy management

Optimization problem

minimize Sum of costs of chillers electrical energy consumption subject to

  • 1. Chiller thermal energy request = Buildings energy request – Storage energy
  • 2. Storage dynamics
  • 3. Storage limits, chillers limits, comfort constraints

9

slide-19
SLIDE 19

Building district energy management

Optimization problem

minimize Sum of costs of chillers electrical energy consumption subject to

  • 1. Chiller thermal energy request = Buildings energy request – Storage energy
  • 2. Storage dynamics
  • 3. Storage limits, chillers limits, comfort constraints

Compact form – x: temperature set-points, storage usage minimize

  • i

fi(x) subject to x ∈

  • i

Xi

9

slide-20
SLIDE 20

Problem set-up

Decision-coupled problem

minimize

  • i

fi(x) subject to x ∈

  • i

Xi

10

slide-21
SLIDE 21

Problem set-up

Decision-coupled problem

minimize

  • i

fi(x) subject to x ∈

  • i

Xi

  • local objectives fi

10

slide-22
SLIDE 22

Problem set-up

Decision-coupled problem

minimize

  • i

fi(x) subject to x ∈

  • i

Xi

  • local objectives fi
  • local constraints Xi

10

slide-23
SLIDE 23

Problem set-up

Decision-coupled problem

minimize

  • i

fi(x) subject to x ∈

  • i

Xi

  • local objectives fi
  • local constraints Xi
  • coupled decision x

10

slide-24
SLIDE 24

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

11

slide-25
SLIDE 25

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

  • xi: “copy” of x maintained by agent i

11

slide-26
SLIDE 26

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

  • xi: “copy” of x maintained by agent i
  • Xi: local constraint set of agent i

11

slide-27
SLIDE 27

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

  • xi: “copy” of x maintained by agent i
  • Xi: local constraint set of agent i
  • zi: information vector – constructed based on the info of agent’s i neighbors

11

slide-28
SLIDE 28

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

  • xi: “copy” of x maintained by agent i
  • Xi: local constraint set of agent i
  • zi: information vector – constructed based on the info of agent’s i neighbors
  • Objective function

fi(xi): local cost/utility of agent i g(xi, zi): Proxy term, penalizing disagreement with other agents

11

slide-29
SLIDE 29

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

12

slide-30
SLIDE 30

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

Step 2a: Broadcast x∗

i (zi) to

neighbors Step 2b: Receive neighbors’ solutions

12

slide-31
SLIDE 31

Proposed distributed algorithm

Step 1: Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

Step 2a: Broadcast x∗

i (zi) to

neighbors Step 2b: Receive neighbors’ solutions Step 3: Update zi on the basis of information received Go to Step 1

12

slide-32
SLIDE 32

Proposed distributed algorithm

Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

12

slide-33
SLIDE 33

Proposed distributed algorithm

Local problem of agent i minimize fi(xi) + g(xi, zi) subject to xi ∈ Xi

  • ⇒ x∗

i (zi)

  • Specify
  • Information vector zi
  • Proxy term term g(xi, zi)
  • Note that these terms change across algorithm iterations

12

slide-34
SLIDE 34

Proposed distributed algorithm

Local problem of agent i at iteration k + 1 zi(k) =

  • j

ai

j(k)xj(k)

xi(k + 1) = arg min

xi ∈Xi fi(xi) +

1 c(k)xi − zi(k)2

13

slide-35
SLIDE 35

Proposed distributed algorithm

Local problem of agent i at iteration k + 1 zi(k) =

  • j

ai

j(k)xj(k)

xi(k + 1) = arg min

xi ∈Xi fi(xi) +

1 c(k)xi − zi(k)2

  • Information vector
  • zi(k) =

j ai j(k)xj(k)

  • ai

j(k): how agent i weights info of agent j

  • Proxy term
  • 1

c(k)xi − zi(k)2: deviation from (weighted) average

  • c(k): trade-off between optimality and agents’ disagreement

13

slide-36
SLIDE 36

Proposed distributed algorithm

Local problem of agent i at iteration k + 1 zi(k) =

  • j

ai

j(k)xj(k)

xi(k + 1) = arg min

xi ∈Xi fi(xi) +

1 c(k)xi − zi(k)2

  • Does this algorithm converge?
  • If yes, does it provide the same solution with the centralized problem (had

we been able to solve it)?

14

slide-37
SLIDE 37

Algorithm analysis

  • 1. Convexity and compactness
  • fi(·): convex for all i
  • Xi: compact, convex, non-empty interior for all i

⇒ fi(·): Lipschitz continuous on Xi

15

slide-38
SLIDE 38

Algorithm analysis

  • 1. Convexity and compactness
  • fi(·): convex for all i
  • Xi: compact, convex, non-empty interior for all i

⇒ fi(·): Lipschitz continuous on Xi

  • 2. Choice of the proxy term
  • c(k)
  • k: non-increasing
  • Should not decrease too fast
  • k

c(k) = ∞

  • k

c(k)2 < ∞

  • E.g., harmonic series

15

slide-39
SLIDE 39

Algorithm analysis

  • 3. Information mix
  • Weights ai

j(k): non-zero lower bound if link between i − j present

⇒ Info mixing at a non-diminishing rate

  • Weights ai

j(k): form a doubly stochastic matrix

⇒ Agents influence each other equally in the long run

16

slide-40
SLIDE 40

Algorithm analysis

  • 3. Information mix
  • Weights ai

j(k): non-zero lower bound if link between i − j present

⇒ Info mixing at a non-diminishing rate

  • Weights ai

j(k): form a doubly stochastic matrix

⇒ Agents influence each other equally in the long run

  • 4. Network connectivity – All information flows (eventually)
  • Any pair of agents communicates infinitely often
  • Bounded intercommunication time

16

slide-41
SLIDE 41

Algorithm analysis

  • 3. Information mix
  • Weights ai

j(k): non-zero lower bound if link between i − j present

⇒ Info mixing at a non-diminishing rate

  • Weights ai

j(k): form a doubly stochastic matrix

⇒ Agents influence each other equally in the long run

  • 4. Network connectivity – All information flows (eventually)
  • Any pair of agents communicates infinitely often
  • Bounded intercommunication time

16

slide-42
SLIDE 42

Algorithm analysis

  • 3. Information mix
  • Weights ai

j(k): non-zero lower bound if link between i − j present

⇒ Info mixing at a non-diminishing rate

  • Weights ai

j(k): form a doubly stochastic matrix

⇒ Agents influence each other equally in the long run

  • 4. Network connectivity – All information flows (eventually)
  • Any pair of agents communicates infinitely often
  • Bounded intercommunication time

16

slide-43
SLIDE 43

Algorithm analysis

  • 3. Information mix
  • Weights ai

j(k): non-zero lower bound if link between i − j present

⇒ Info mixing at a non-diminishing rate

  • Weights ai

j(k): form a doubly stochastic matrix

⇒ Agents influence each other equally in the long run

  • 4. Network connectivity – All information flows (eventually)
  • Any pair of agents communicates infinitely often
  • Bounded intercommunication time

16

slide-44
SLIDE 44

Algorithm analysis

Main result Under the structural + network assumptions, the proposed proximal algorithm converges to some minimizer x∗ of the centralized problem, i.e., lim

k→∞ xi(k) − x∗ = 0, for all i

  • Asymptotic agreement and optimality
  • Rate no faster than c(k) – “slow enough” to trade agreement and
  • ptimality

17

slide-45
SLIDE 45

Comparison with other methods

  • Proximal algorithms vs. gradient/subgradient methods

18

slide-46
SLIDE 46

Comparison with other methods

  • Proximal algorithms

xi(k + 1) = arg min

xi ∈Xi fi(xi) +

1 c(k)xi − zi(k)2

  • Gradient algorithms

xi(k + 1) = PXi

  • zi(k) − c(k)∇fi(zi(k))
  • Proximal algorithms allow for
  • No gradient/subgradient calculation – user can feed problem data in

any solver

  • Heterogeneous constraint sets
  • No differentiability assumptions

19

slide-47
SLIDE 47

Comparison with subgradient

Optimal power allocation in cellular networks (non-differentiable objective) proposed solution vs. gradient-based approach

20 40 60 80 100

Iteration

5 10 15 20 25

Cost function

Proximal (Sub)gradient 20

slide-48
SLIDE 48

Building district problem revisited – Simulation results

Set-up

  • 3 buildings - 3 zones each (different chiller per building)
  • Pair-wise communication (gossip)

19

slide-49
SLIDE 49

Building district problem revisited – Simulation results

Set-up

  • 3 buildings - 3 zones each (different chiller per building)
  • Pair-wise communication (gossip)

19

slide-50
SLIDE 50

Building district problem revisited – Simulation results

Set-up

  • 3 buildings - 3 zones each (different chiller per building)
  • Pair-wise communication (gossip)

19

slide-51
SLIDE 51

Building district problem revisited – Simulation results

Set-up

  • 3 buildings - 3 zones each (different chiller per building)
  • Pair-wise communication (gossip)

19

slide-52
SLIDE 52

Building district problem revisited – Simulation results

Set-up

  • 3 buildings - 3 zones each (different chiller per building)
  • Pair-wise communication (gossip)

Implementation

  • Simulation in MATLAB
  • Optimization solver SEDUMI via the MATLAB interface YALMIP

19

slide-53
SLIDE 53

Simulation results – Temperature set-points

Optimal zone temperature profiles of building 1 (consensus solution).

2 4 6 8 10 12 14 16 18 20 22 24 Time [h] 19 20 21 22 23 24 25 26 Zones Temperature [oC]

Zone 1 Zone 2 Zone 3 Constr.

Temperature of zone 2 (middle one) is always the lowest, it acts as a passive thermal storage draining heat of the other zones through floor/ceiling.

20

slide-54
SLIDE 54

Simulation results – Storage usage

10 12 14 16 18 20 22 24 Time [h] 25 50 75 100 125 150 175 200 Stored energy [MJ]

  • 10
  • 7.5
  • 5
  • 2.5

2.5 5 7.5 10

Energy exchange [MJ] 2 4 6 8

E es

s 1

es

2

es

3

Solution computed at iteration k = 1 by the middle-chiller building (“blue”). The middle-chiller building uses the storage charged by the others

21

slide-55
SLIDE 55

Simulation results – Storage usage

Time [h] 25 50 75 100 125 150 175 200 Stored energy [MJ]

  • 10
  • 7.5
  • 5
  • 2.5

2.5 5 7.5 10

Energy exchange [MJ]

E es

s 1

es

2

es

3

10 12 14 16 18 20 22 24 2 4 6 8

At consensus, the small-chiller building (“orange”) uses the storage charged by the others

22

slide-56
SLIDE 56

Simulation results – Chillers usage

2 4 6 8 10 12 14 16 18 20 22 24 Time [h] 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 COP

Medium Small Large

COP of the chillers when each building uses a fix fraction of the storage

23

slide-57
SLIDE 57

Simulation results – Chillers usage

2 4 6 8 10 12 14 16 18 20 22 24

Time [h]

0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2

COP

Medium Small Large

COP of the chillers in the optimally shared storage case

24

slide-58
SLIDE 58

Simulation results

Solution computed based on nominal disturbance profiles...

2 4 6 8 10 12 14 16 18 20 22 24 Time [h] 200 400 600 800 1000 20 40 60 LW rad. [W/m2] SW rad. [W/m2] Outside T. [°C] Occupancy [#]

25

slide-59
SLIDE 59

Simulation results

Solution computed based on nominal disturbance profiles...

2 4 6 8 10 12 14 16 18 20 22 24 Time [h] 200 400 600 800 1000 20 40 60 LW rad. [W/m2] SW rad. [W/m2] Outside T. [°C] Occupancy [#]

Courtesy of Istituto di Scienze dell’Atmosfera e del Clima (ISAC) - CNR

25

slide-60
SLIDE 60

Problem set-up

Decision-coupled problem

minimize

  • i

fi(x) subject to x ∈

  • i

Xi

26

slide-61
SLIDE 61

Problem set-up

Decision-coupled problem with uncertainty

minimize

  • i

fi(x) subject to x ∈

  • i

Xi(δ), for all δ ∈ ∆

27

slide-62
SLIDE 62

Problem set-up

Decision-coupled problem with uncertainty

minimize

  • i

fi(x) subject to x ∈

  • i

Xi(δ), for all δ ∈ ∆

  • Stochastic set-up
  • δ: Uncertain parameter δ ∼ P
  • ∆: (Possibly) continuous set
  • Semi-infinite optimization program

28

slide-63
SLIDE 63

Problem set-up

Decision-coupled problem with uncertainty

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈∆

Xi(δ)

  • Stochastic set-up
  • δ: Uncertain parameter δ ∼ P
  • ∆: (Possibly) continuous set
  • Semi-infinite optimization program

29

slide-64
SLIDE 64

Data-based approach

Decision-coupled problem with uncertainty

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈S

Xi(δ)

  • Replace ∆ with S

27

slide-65
SLIDE 65

Data-based approach

Decision-coupled problem with uncertainty

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈S

Xi(δ) Two cases:

  • 1. Agents have the same data set S

28

slide-66
SLIDE 66

Data-based approach

Decision-coupled problem with uncertainty

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈S

Xi(δ) Two cases:

  • 1. Agents have the same data set S
  • 2. Agents have different data sets
  • Si
  • i

28

slide-67
SLIDE 67

Data-based approach

Decision-coupled problem with uncertainty

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈Si

Xi(δ) Two cases:

  • 1. Agents have the same data set S
  • 2. Agents have different data sets
  • Si
  • i

29

slide-68
SLIDE 68

Data-based approach

Common data set – distributed implementation minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈S

Xi(δ)

28

slide-69
SLIDE 69

Data-based approach

Common data set – distributed implementation minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈S

Xi(δ)

  • Apply proximal algorithm with

δ∈S Xi(δ) in place of Xi

28

slide-70
SLIDE 70

Data-based approach

Common data set – distributed implementation minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈S

Xi(δ)

  • Apply proximal algorithm with

δ∈S Xi(δ) in place of Xi

  • Let x∗

S denote the converged solution

28

slide-71
SLIDE 71

Probabilistic feasibility – Common data set

Data-based program PS

minimize

  • i

fi(x) subject to → x∗

S

x ∈

  • i
  • δ∈S

Xi(δ)

Robust program P∆

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈∆

Xi(δ)

  • Is x∗

S feasible for P∆?

29

slide-72
SLIDE 72

Probabilistic feasibility – Common data set

Data-based program PS

minimize

  • i

fi(x) subject to → x∗

S

x ∈

  • i
  • δ∈S

Xi(δ)

Robust program P∆

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈∆

Xi(δ)

  • Is x∗

S feasible for P∆?

  • Is this true for any S?

29

slide-73
SLIDE 73

Probabilistic feasibility – Common data set

Data-based program PS

minimize

  • i

fi(x) subject to → x∗

S

x ∈

  • i
  • δ∈S

Xi(δ)

Robust program P∆

minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈∆

Xi(δ)

Feasibility link [Calafiore & Campi, TAC 2006] Fix β ∈ (0, 1) and S. With confidence ≥ 1 − β, x∗

S is feasible with probability

≥ 1 − ǫ(d, |S|, β), i.e. P

  • δ ∈ ∆ : x∗

S /

  • i

Xi(δ)

  • ≤ ǫ(d, |S|, β) with prob. ≥ 1 − β

30

slide-74
SLIDE 74

Probabilistic feasibility – Common data set

Feasibility link Fix β ∈ (0, 1) and S. With confidence ≥ 1 − β, x∗

S is feasible for P∆ with

probability ≥ 1 − ǫ(d, |S|, β), i.e. P

  • δ ∈ ∆ : x∗

S /

  • i

Xi(δ)

  • ≤ ǫ(d, |S|, β) with prob. ≥ 1 − β
  • On which parameters does ǫ depends on?

ǫ = 2 |S|

  • d + ln 1

β

  • Logarithmic in β: β can be set close to 0
  • Linear in |S|−1: The more data the better the result
  • Linear in d: # decision variables

31

slide-75
SLIDE 75

Data-based approach

Different data set – distributed implementation minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈Si

Xi(δ)

  • Apply proximal algorithm with

δ∈Si Xi(δ) in place of Xi

32

slide-76
SLIDE 76

Data-based approach

Different data set – distributed implementation minimize

  • i

fi(x) subject to x ∈

  • i
  • δ∈Si

Xi(δ)

  • Apply proximal algorithm with

δ∈Si Xi(δ) in place of Xi

  • Let x∗

S denote the converged solution, S = {Si}i

32

slide-77
SLIDE 77

Probabilistic feasibility – Different data sets

Single-agent - a posteriori Fix βi ∈ (0, 1) and Si. With confidence ≥ 1 − βi, P

  • δ ∈ ∆ : x∗

S /

∈ Xi(δ)

  • ≤ ǫi(dSi

i , |Si|, βi)

A posteriori result

  • dSi

i : empirical estimate of “support” samples (wait and see)

Changing Si the result will change

  • Complexity of ǫi(dSi

i , |Si|, βi) as in the previous case

  • Result thanks to [Campi, Garatti & Ramponi, CDC 2015]

33

slide-78
SLIDE 78

Probabilistic feasibility – Different data sets

Single-agent - a posteriori Fix βi ∈ (0, 1) and Si. With confidence ≥ 1 − βi, P

  • δ ∈ ∆ : x∗

S /

∈ Xi(δ)

  • ≤ ǫi(dSi

i )

  • Two-agent example, d = 2

dS1

1 = 1 and dS2 2 = 1

dS1

1 = 0 and dS2 2 = 2

34

slide-79
SLIDE 79

Probabilistic feasibility – Different data sets

Multi-agent - a posteriori Fix β ∈ (0, 1) and

  • Si
  • i. With confidence ≥ 1 − β,

P

  • δ ∈ ∆ : x∗

S /

  • i

Xi(δ)

  • i

ǫi(dSi

i )

A posteriori result

  • Can we turn it into an a priori statement?

35

slide-80
SLIDE 80

Probabilistic feasibility – Different data sets

Multi-agent - a posteriori Fix β ∈ (0, 1) and

  • Si
  • i. With confidence ≥ 1 − β,

P

  • δ ∈ ∆ : x∗

S /

  • i

Xi(δ)

  • i

ǫi(dSi

i )

A posteriori result

  • Can we turn it into an a priori statement?
  • What is the worst-case value for

i ǫi(dSi i ) that we can “observe”?

35

slide-81
SLIDE 81

Probabilistic feasibility – Different data sets

Multi-agent - a posteriori Fix β ∈ (0, 1) and

  • Si
  • i. With confidence ≥ 1 − β,

P

  • δ ∈ ∆ : x∗

S /

  • i

Xi(δ)

  • i

ǫi(dSi

i )

A posteriori result

  • Can we turn it into an a priori statement?
  • What is the worst-case value for

i ǫi(dSi i ) that we can “observe”?

  • Conservative bound: dSi

i

≤ d for all i

35

slide-82
SLIDE 82

Probabilistic feasibility – Different data sets

Multi-agent - a posteriori Fix β ∈ (0, 1) and

  • Si
  • i. With confidence ≥ 1 − β,

P

  • δ ∈ ∆ : x∗

S /

  • i

Xi(δ)

  • i

ǫi(dSi

i )

A posteriori result

  • Can we turn it into an a priori statement?
  • What is the worst-case value for

i ǫi(dSi i ) that we can “observe”?

  • Conservative bound: dSi

i

≤ d for all i

  • Sharper bound:

i dSi i

≤ d (# decision variables)

35

slide-83
SLIDE 83

Probabilistic feasibility – Different data sets

Multi-agent - a priori Fix β ∈ (0, 1) and

  • Si
  • i. With confidence ≥ 1 − β,

P

  • δ ∈ ∆ : x∗

S /

  • i

Xi(δ)

  • ≤ ǫ

where ǫ = maximize

  • i

ǫi(di) subject to

  • i

di ≤ d

36

slide-84
SLIDE 84

Common vs. different data sets

Number of agents - m

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15

Probability of violation

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

" e " "

Approach using different constraint sets

  • Close to the case of common data sets
  • Less conservative than the worst case bound

37

slide-85
SLIDE 85

Literature comparison

Closest approach1

  • almost sure convergence results

(need to sample constraints infinitely many times)

  • 1S. Lee and A. Nedic, Distributed random projection algorithm for convex
  • ptimization, IEEE Journal on Selected Topics in Signal Processing 2013.

38

slide-86
SLIDE 86

Literature comparison

Closest approach1

  • almost sure convergence results

(need to sample constraints infinitely many times)

Proposed solution

  • weaker guarantees but with a finite number of samples
  • 1S. Lee and A. Nedic, Distributed random projection algorithm for convex
  • ptimization, IEEE Journal on Selected Topics in Signal Processing 2013.

38

slide-87
SLIDE 87

Addressed problems

min

x m

  • i=1

fi(x) s.t. x ∈

m

  • i=1

Xi

  • local objectives fi
  • coupled decision x
  • local constraints Xi

Decision-coupled problem

39

slide-88
SLIDE 88

Addressed problems

min

x m

  • i=1

fi(x) s.t. x ∈

m

  • i=1

Xi

  • local objectives fi
  • coupled decision x
  • local constraints Xi

Decision-coupled problem min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

39

slide-89
SLIDE 89

Addressed problems

min

x m

  • i=1

fi(x) s.t. x ∈

m

  • i=1

Xi

  • local objectives fi
  • coupled decision x
  • local constraints Xi

Decision-coupled problem min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi

39

slide-90
SLIDE 90

Addressed problems

min

x m

  • i=1

fi(x) s.t. x ∈

m

  • i=1

Xi

  • local objectives fi
  • coupled decision x
  • local constraints Xi

Decision-coupled problem min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi

39

slide-91
SLIDE 91

Addressed problems

min

x m

  • i=1

fi(x) s.t. x ∈

m

  • i=1

Xi

  • local objectives fi
  • coupled decision x
  • local constraints Xi

Decision-coupled problem min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi
  • coupling constraint

m

i=1 gi(xi) ≤ 0 39

slide-92
SLIDE 92

Addressed problems

min

x m

  • i=1

fi(x) s.t. x ∈

m

  • i=1

Xi

  • local objectives fi
  • coupled decision x
  • local constraints Xi

Decision-coupled problem min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi
  • coupling constraint

m

i=1 gi(xi) ≤ 0

Constraint-coupled problem

39

slide-93
SLIDE 93

Proposed solution for constraint-coupled problems

min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi
  • coupling constraint

m

i=1 gi(xi) ≤ 0 40

slide-94
SLIDE 94

Proposed solution for constraint-coupled problems

At each iteration k, agent i min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi
  • coupling constraint

m

i=1 gi(xi) ≤ 0 40

slide-95
SLIDE 95

Proposed solution for constraint-coupled problems

At each iteration k, agent i ℓi(k) ←

  • j∈Ni

ai

j(k)λj(k)

min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi
  • coupling constraint

m

i=1 gi(xi) ≤ 0 40

slide-96
SLIDE 96

Proposed solution for constraint-coupled problems

At each iteration k, agent i ℓi(k) ←

  • j∈Ni

ai

j(k)λj(k)

λi(k+1) ← arg max

λi≥0 ˜

ϕi(λi) where ˜ ϕi(λi) = λ⊤

i gi(xi(k+1))

1 c(k) λi − ℓi(k)2 2

min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi
  • coupling constraint

m

i=1 gi(xi) ≤ 0 40

slide-97
SLIDE 97

Proposed solution for constraint-coupled problems

At each iteration k, agent i ℓi(k) ←

  • j∈Ni

ai

j(k)λj(k)

xi(k+1) ← arg min

xi∈Xi

˜ fi(xi) λi(k+1) ← arg max

λi≥0 ˜

ϕi(λi) where ˜ fi(xi) = fi(xi) + ℓi(k)⊤gi(xi) ˜ ϕi(λi) = λ⊤

i gi(xi(k+1))

1 c(k) λi − ℓi(k)2 2

min

x1,...,xm m

  • i=1

fi(xi) s.t.

m

  • i=1

gi(xi) ≤ 0 xi ∈ Xi ∀i

  • local objectives fi
  • local decisions xi
  • coupling constraint

m

i=1 gi(xi) ≤ 0 40

slide-98
SLIDE 98

Algorithm analysis

Main result (Convergence & optimality) Under the structural + network assumptions, the proposed algorithm combining dual decomposition and proximal minimization converges to the set of minimizers

  • f the centralized problem.

41

slide-99
SLIDE 99

Algorithm analysis

Main result (Convergence & optimality) Under the structural + network assumptions, the proposed algorithm combining dual decomposition and proximal minimization converges to the set of minimizers

  • f the centralized problem.

Probabilistic feasibility results for the stochastic case have been developed.

41

slide-100
SLIDE 100

Problem set-up – discrete decision variables

P : min

x1,...,xm m

  • i=1

c⊤

i xi

subject to:

m

  • i=1

Aixi ≤ b xi ∈ Xi ∀i = 1, . . . , m Features

  • local decision vectors xi
  • local linear objectives c⊤

i xi

  • p coupling linear constraints m

i=1 Aixi ≤ b

  • local mixed-integer polyhedral constraint sets Xi

42

slide-101
SLIDE 101

Problem set-up – discrete decision variables

P : min

x1,...,xm m

  • i=1

c⊤

i xi

subject to:

m

  • i=1

Aixi ≤ b xi ∈ Xi ∀i = 1, . . . , m Features

  • local decision vectors xi
  • local linear objectives c⊤

i xi

  • p coupling linear constraints m

i=1 Aixi ≤ b

  • local mixed-integer polyhedral constraint sets Xi ⇒ combinatorial

complexity

43

slide-102
SLIDE 102

Constraint-coupled MILPs

The problem fits the structure of a constraint-coupled problem

44

slide-103
SLIDE 103

Constraint-coupled MILPs

The problem fits the structure of a constraint-coupled problem

but...

It is non-convex, hence the distributed algorithms developed for convex problems have no guarantees

44

slide-104
SLIDE 104

Constraint-coupled MILPs

The problem fits the structure of a constraint-coupled problem

but...

It is non-convex, hence the distributed algorithms developed for convex problems have no guarantees

we then aim at

  • 1. providing a feasible (possibly sub-optimal) solution
  • 2. quantifying the quality of the solution

44

slide-105
SLIDE 105

Constraint-coupled MILPs

Goal

  • 1. provide a feasible (possibly sub-optimal) solution
  • 2. quantifying the quality of the solution

Literature

Some problem-specific approaches to recover a feasible solution

[Vujanic et al., 2016]2

More general duality-based approach to recover a feasible solution with sub-optimality guarantees

  • 2R. Vujanic, P. M. Esfahani, P. J. Goulart, S. Mariethoz, and M. Morari, A

decomposition method for large scale MILPs, with performance guarantees and a power system application, Automatica, 2016

45

slide-106
SLIDE 106

Proposed solution

Main idea of [Vujanic et al., 2016]

  • 1. tighten the coupling constraint by a specific amount ˜

ρ ≥ 0

  • 2. obtain the dual optimal solution λ⋆

˜ ρ

  • 3. recover a feasible primal solution using λ⋆

˜ ρ

46

slide-107
SLIDE 107

Proposed solution

Main idea of [Vujanic et al., 2016]

  • 1. tighten the coupling constraint by a specific amount ˜

ρ ≥ 0

  • 2. obtain the dual optimal solution λ⋆

˜ ρ

  • 3. recover a feasible primal solution using λ⋆

˜ ρ

Proposed solution

A distributed iterative algorithm that merges:

  • the procedure for solving constraint-coupled problems
  • adaptive tightening of coefficient ρ based on the same idea of

[Vujanic et al., 2016]

46

slide-108
SLIDE 108

Theoretical results

Fact

By construction, ρ → ¯ ρ ≤ ˜ ρ (often <)

47

slide-109
SLIDE 109

Theoretical results

Fact

By construction, ρ → ¯ ρ ≤ ˜ ρ (often <)

Theorem (Feasibility)

After a finite number of iterations, the algorithm provides a solution that is feasible for P

47

slide-110
SLIDE 110

Theoretical results

Fact

By construction, ρ → ¯ ρ ≤ ˜ ρ (often <)

Theorem (Feasibility)

After a finite number of iterations, the algorithm provides a solution that is feasible for P

Theorem (Performance)

The performance are no-worse than that of [Vujanic et al., 2016] (often better)

47

slide-111
SLIDE 111

Summary & Future work

Performance optimization of a network

  • General distributed optimization framework accounting for different

complexity features , i.e., heterogeneity of the agents, privacy of their local info, uncertainty, combinatorial complexity

48

slide-112
SLIDE 112

Summary & Future work

Performance optimization of a network

  • General distributed optimization framework accounting for different

complexity features , i.e., heterogeneity of the agents, privacy of their local info, uncertainty, combinatorial complexity Applications

  • energy management of a building district
  • power allocation in cellular networks
  • plug-in electric vehicles charging scheduling

48

slide-113
SLIDE 113

Summary & Future work

Performance optimization of a network

  • General distributed optimization framework accounting for different

complexity features , i.e., heterogeneity of the agents, privacy of their local info, uncertainty, combinatorial complexity Applications

  • energy management of a building district
  • power allocation in cellular networks
  • plug-in electric vehicles charging scheduling

What comes next?

  • Convergence rate analysis
  • Rolling horizon implementations
  • Uncertain constraint-coupled MILP
  • Application to Mixed Logical Dynamical (MLD) systems
  • More applications

48

slide-114
SLIDE 114

Main references

Margellos, Falsone, Garatti, Prandini (2018) Distributed constrained optimization and consensus in uncertain networks via proximal minimization. IEEE Transactions on Automatic Control, 63(5):1372–1387. Falsone, Margellos, Garatti, Prandini (2017) Dual decomposition for multi-agent distributed optimization with coupling constraints. Automatica, 84:149–158. Falsone, Margellos, Prandini (2018) A distributed iterative algorithm for multi-agent MILPs: Finite-time feasibility and performance characterization. IEEE Control Systems Letters, 2(4):563–568.

34

slide-115
SLIDE 115

Credit

Alessandro Falsone Simone Garatti Kostas Margellos

35

slide-116
SLIDE 116

Credit

Work supported by the European Commission under the UnCoVerCPS project Unifying Control and Verification of Cyber-Physical Systems, H2020, 2015-2018

36