A Dynamic Near-Optimal Algorithm for Online Linear Programming - - PowerPoint PPT Presentation

a dynamic near optimal algorithm for online linear
SMART_READER_LITE
LIVE PREVIEW

A Dynamic Near-Optimal Algorithm for Online Linear Programming - - PowerPoint PPT Presentation

Online Linear Programming Main Results and Key Ideas Related and More Recent Work A Dynamic Near-Optimal Algorithm for Online Linear Programming Yinyu Ye Department of Management Science and Engineering and Institute of Computational and


slide-1
SLIDE 1

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

A Dynamic Near-Optimal Algorithm for Online Linear Programming

Yinyu Ye

Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford University Joint work with Shipra Agrawal and Zizhuo Wang

Information-Based Complexity and Stochastic Computation

September 17, 2014

Yinyu Ye Online LP, ICERM 2014

slide-2
SLIDE 2

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Outline

◮ Online Linear Programming ◮ Main Results and Key Ideas ◮ Related and More Recent Work

Yinyu Ye Online LP, ICERM 2014

slide-3
SLIDE 3

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Background

Consider a store that sells a number of goods/products

◮ There is a fixed selling period or number of buyers

Yinyu Ye Online LP, ICERM 2014

slide-4
SLIDE 4

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Background

Consider a store that sells a number of goods/products

◮ There is a fixed selling period or number of buyers ◮ There is a fixed inventory of goods

Yinyu Ye Online LP, ICERM 2014

slide-5
SLIDE 5

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Background

Consider a store that sells a number of goods/products

◮ There is a fixed selling period or number of buyers ◮ There is a fixed inventory of goods ◮ Customers come and require a bundle of goods and bid for

certain prices

Yinyu Ye Online LP, ICERM 2014

slide-6
SLIDE 6

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Background

Consider a store that sells a number of goods/products

◮ There is a fixed selling period or number of buyers ◮ There is a fixed inventory of goods ◮ Customers come and require a bundle of goods and bid for

certain prices

◮ Decision: To sell or not to sell to each individual customer?

Yinyu Ye Online LP, ICERM 2014

slide-7
SLIDE 7

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Background

Consider a store that sells a number of goods/products

◮ There is a fixed selling period or number of buyers ◮ There is a fixed inventory of goods ◮ Customers come and require a bundle of goods and bid for

certain prices

◮ Decision: To sell or not to sell to each individual customer? ◮ Objective: Maximize the revenue.

Yinyu Ye Online LP, ICERM 2014

slide-8
SLIDE 8

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

An Example

Bid 1(t = 1) Bid 2(t = 2) ..... Inventory(b) Price(πt) $100 $30 ... Decision x1 x2 ... Pants 1 ... 100 Shoes 1 ... 50 T-shirts 1 ... 500 Jackets ... 200 Hats 1 1 ... 1000

Yinyu Ye Online LP, ICERM 2014

slide-9
SLIDE 9

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Online Linear Programming Model

The classical offline version of the above program can be formulated as a linear (integer) program as all information data would have arrived: compute xt, t = 1, ..., n and maximizex n

t=1 πtxt

subject to n

t=1 aitxt ≤ bi,

∀i = 1, ..., m xt ∈ {0, 1} (0 ≤ xt ≤ 1), ∀t = 1, ..., n.

Yinyu Ye Online LP, ICERM 2014

slide-10
SLIDE 10

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Online Linear Programming Model

The classical offline version of the above program can be formulated as a linear (integer) program as all information data would have arrived: compute xt, t = 1, ..., n and maximizex n

t=1 πtxt

subject to n

t=1 aitxt ≤ bi,

∀i = 1, ..., m xt ∈ {0, 1} (0 ≤ xt ≤ 1), ∀t = 1, ..., n. Now we consider the online or streamline and data-driven version

  • f this problem:

◮ We only know b and n at the start

Yinyu Ye Online LP, ICERM 2014

slide-11
SLIDE 11

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Online Linear Programming Model

The classical offline version of the above program can be formulated as a linear (integer) program as all information data would have arrived: compute xt, t = 1, ..., n and maximizex n

t=1 πtxt

subject to n

t=1 aitxt ≤ bi,

∀i = 1, ..., m xt ∈ {0, 1} (0 ≤ xt ≤ 1), ∀t = 1, ..., n. Now we consider the online or streamline and data-driven version

  • f this problem:

◮ We only know b and n at the start ◮ the bidder information is revealed sequentially along with the

corresponding objective coefficient.

Yinyu Ye Online LP, ICERM 2014

slide-12
SLIDE 12

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Online Linear Programming Model

The classical offline version of the above program can be formulated as a linear (integer) program as all information data would have arrived: compute xt, t = 1, ..., n and maximizex n

t=1 πtxt

subject to n

t=1 aitxt ≤ bi,

∀i = 1, ..., m xt ∈ {0, 1} (0 ≤ xt ≤ 1), ∀t = 1, ..., n. Now we consider the online or streamline and data-driven version

  • f this problem:

◮ We only know b and n at the start ◮ the bidder information is revealed sequentially along with the

corresponding objective coefficient.

◮ an irrevocable decision must be made as soon as an order

arrives without observing or knowing the future data.

Yinyu Ye Online LP, ICERM 2014

slide-13
SLIDE 13

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Application Overview

◮ Revenue management problems: Airline tickets booking, hotel

booking;

◮ Online network routing on an edge-capacitated network; ◮ Online combinatorial auction; ◮ Online adwords allocation

Yinyu Ye Online LP, ICERM 2014

slide-14
SLIDE 14

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Model Assumptions

Main Assumptions

◮ 0 ≤ ait ≤ 1, for all (i, t); ◮ πt ≥ 0 for all t ◮ The data (at, πt) arrive in a random order.

Yinyu Ye Online LP, ICERM 2014

slide-15
SLIDE 15

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Model Assumptions

Main Assumptions

◮ 0 ≤ ait ≤ 1, for all (i, t); ◮ πt ≥ 0 for all t ◮ The data (at, πt) arrive in a random order.

Denote the offline LP maximal value by OPT(A, π). We call an

  • nline algorithm A to be c-competitive if and only if

Eσ n

  • t=1

πtxt(σ, A)

  • ≥ c · OPT(A, π) ∀(A, π),

where σ is the permutation of arriving orders. In what follows, we let B = min

i {bi}(> 0).

Yinyu Ye Online LP, ICERM 2014

slide-16
SLIDE 16

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Main Results: Necessary and Sufficient Conditions

Theorem

For any fixed 0 < ǫ < 1, there is no online algorithm for solving the linear program with competitive ratio 1 − ǫ if B < log(m) ǫ2 .

Yinyu Ye Online LP, ICERM 2014

slide-17
SLIDE 17

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Main Results: Necessary and Sufficient Conditions

Theorem

For any fixed 0 < ǫ < 1, there is no online algorithm for solving the linear program with competitive ratio 1 − ǫ if B < log(m) ǫ2 .

Theorem

For any fixed 0 < ǫ < 1, there is a 1 − ǫ competitive online algorithm for solving the linear program if B ≥ Ω m log (n/ǫ) ǫ2

  • .

Yinyu Ye Online LP, ICERM 2014

slide-18
SLIDE 18

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Main Results: Necessary and Sufficient Conditions

Theorem

For any fixed 0 < ǫ < 1, there is no online algorithm for solving the linear program with competitive ratio 1 − ǫ if B < log(m) ǫ2 .

Theorem

For any fixed 0 < ǫ < 1, there is a 1 − ǫ competitive online algorithm for solving the linear program if B ≥ Ω m log (n/ǫ) ǫ2

  • .

Agrawal, Wang and Y [Operations Research 2014]

Yinyu Ye Online LP, ICERM 2014

slide-19
SLIDE 19

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Key Ideas: A Worst-Case Distribution Example

The proof of the negative result is based on a distribution of instances (the number of each types of columns is chosen according to certain distribution) with m = 2k, and then show that no allocation rule can achieve (1 − ǫ)-optimality in expectation under randomized permutation.

Yinyu Ye Online LP, ICERM 2014

slide-20
SLIDE 20

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Key Ideas: A Learning Algorithm is Needed

The proof of the positive result is constructive and based on a learning policy.

Yinyu Ye Online LP, ICERM 2014

slide-21
SLIDE 21

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Key Ideas: A Learning Algorithm is Needed

The proof of the positive result is constructive and based on a learning policy.

◮ There is no distribution known so that any type of stochastic

  • ptimization models is not applicable.

Yinyu Ye Online LP, ICERM 2014

slide-22
SLIDE 22

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Key Ideas: A Learning Algorithm is Needed

The proof of the positive result is constructive and based on a learning policy.

◮ There is no distribution known so that any type of stochastic

  • ptimization models is not applicable.

◮ Unlike dynamic programming, the decision maker does not

have full information/data so that a backward recursion can not be carried out to find an optimal sequential decision policy.

Yinyu Ye Online LP, ICERM 2014

slide-23
SLIDE 23

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Key Ideas: A Learning Algorithm is Needed

The proof of the positive result is constructive and based on a learning policy.

◮ There is no distribution known so that any type of stochastic

  • ptimization models is not applicable.

◮ Unlike dynamic programming, the decision maker does not

have full information/data so that a backward recursion can not be carried out to find an optimal sequential decision policy.

◮ Thus, the online algorithm needs to be learning-based, in

particular, learning-while-doing.

Yinyu Ye Online LP, ICERM 2014

slide-24
SLIDE 24

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Price Observation of Online Learning I

The problem would be easy if there is an ”ideal price” vector: Bid 1(t = 1) Bid 2(t = 2) ..... Inventory(b) p∗ Bid(πt) $100 $30 ... Decision x1 x2 ... Pants 1 ... 100 $45 Shoes 1 ... 50 $45 T-shirts 1 ... 500 $10 Jackets ... 200 $55 Hats 1 1 ... 1000 $15

Yinyu Ye Online LP, ICERM 2014

slide-25
SLIDE 25

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Price Observation of Online Learning II

◮ Pricing the bid: The optimal dual price vector p∗ of the offline

LP problem can play such a role, that is x∗

t = 1 if πt > aT t p∗

and x∗

t = 0 otherwise, yields a near-optimal solution.

Yinyu Ye Online LP, ICERM 2014

slide-26
SLIDE 26

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Price Observation of Online Learning II

◮ Pricing the bid: The optimal dual price vector p∗ of the offline

LP problem can play such a role, that is x∗

t = 1 if πt > aT t p∗

and x∗

t = 0 otherwise, yields a near-optimal solution. ◮ Based on this observation, our online algorithm works by

learning a threshold price vector ˆ p and using ˆ p to price the bids.

Yinyu Ye Online LP, ICERM 2014

slide-27
SLIDE 27

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Price Observation of Online Learning II

◮ Pricing the bid: The optimal dual price vector p∗ of the offline

LP problem can play such a role, that is x∗

t = 1 if πt > aT t p∗

and x∗

t = 0 otherwise, yields a near-optimal solution. ◮ Based on this observation, our online algorithm works by

learning a threshold price vector ˆ p and using ˆ p to price the bids.

◮ One-time learning algorithm: learn the price vector once using

the initial ǫn input.

Yinyu Ye Online LP, ICERM 2014

slide-28
SLIDE 28

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Price Observation of Online Learning II

◮ Pricing the bid: The optimal dual price vector p∗ of the offline

LP problem can play such a role, that is x∗

t = 1 if πt > aT t p∗

and x∗

t = 0 otherwise, yields a near-optimal solution. ◮ Based on this observation, our online algorithm works by

learning a threshold price vector ˆ p and using ˆ p to price the bids.

◮ One-time learning algorithm: learn the price vector once using

the initial ǫn input.

◮ Dynamic learning algorithm: dynamically update the price

vector at a carefully chosen pace.

Yinyu Ye Online LP, ICERM 2014

slide-29
SLIDE 29

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

One-Time Learning Algorithm

We illustrate a simple One-Time Learning Algorithm:

◮ Set xt = 0 for all 1 ≤ t ≤ ǫn;

Yinyu Ye Online LP, ICERM 2014

slide-30
SLIDE 30

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

One-Time Learning Algorithm

We illustrate a simple One-Time Learning Algorithm:

◮ Set xt = 0 for all 1 ≤ t ≤ ǫn; ◮ Solve the ǫ portion of the problem

maximizex ǫn

t=1 πtxt

subject to ǫn

t=1 aitxt ≤ (1 − ǫ)ǫbi

i = 1, ..., m 0 ≤ xt ≤ 1 t = 1, ..., ǫn and get the optimal dual solution ˆ p;

Yinyu Ye Online LP, ICERM 2014

slide-31
SLIDE 31

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

One-Time Learning Algorithm

We illustrate a simple One-Time Learning Algorithm:

◮ Set xt = 0 for all 1 ≤ t ≤ ǫn; ◮ Solve the ǫ portion of the problem

maximizex ǫn

t=1 πtxt

subject to ǫn

t=1 aitxt ≤ (1 − ǫ)ǫbi

i = 1, ..., m 0 ≤ xt ≤ 1 t = 1, ..., ǫn and get the optimal dual solution ˆ p;

◮ Determine the future allocation xt as:

xt = if πt ≤ ˆ pTat 1 if πt > ˆ pTat as long as aitxt ≤ bi − t−1

j=1 aijxj for all i; otherwise, set

xt = 0.

Yinyu Ye Online LP, ICERM 2014

slide-32
SLIDE 32

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

One-Time Learning Algorithm Result

Theorem

For any fixed ǫ > 0, the one-time learning algorithm is (1 − ǫ) competitive for solving the linear program when B ≥ Ω

  • m log (n/ǫ)

ǫ3

  • Yinyu Ye

Online LP, ICERM 2014

slide-33
SLIDE 33

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Outline of the Proof

◮ With high probability, we clear the market; ◮ With high probability, the revenue is near-optimal if we

include the initial ǫ portion revenue;

◮ With high probability, the first ǫ portion revenue, a learning

cost, doesn’t contribute too much. Then, we prove that the one-time learning algorithm is (1 − ǫ) competitive under condition B ≥ 6m log(n/ǫ)

ǫ3

.

Yinyu Ye Online LP, ICERM 2014

slide-34
SLIDE 34

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Outline of the Proof

◮ With high probability, we clear the market; ◮ With high probability, the revenue is near-optimal if we

include the initial ǫ portion revenue;

◮ With high probability, the first ǫ portion revenue, a learning

cost, doesn’t contribute too much. Then, we prove that the one-time learning algorithm is (1 − ǫ) competitive under condition B ≥ 6m log(n/ǫ)

ǫ3

. But this is one ǫ factor higher than the lower bound...

Yinyu Ye Online LP, ICERM 2014

slide-35
SLIDE 35

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Dynamic Learning Algorithm

In the dynamic price learning algorithm, we update the price at time ǫn, 2ǫn, 4ǫn, ..., till 2kǫ ≥ 1.

Yinyu Ye Online LP, ICERM 2014

slide-36
SLIDE 36

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Dynamic Learning Algorithm

In the dynamic price learning algorithm, we update the price at time ǫn, 2ǫn, 4ǫn, ..., till 2kǫ ≥ 1. At time ℓ ∈ {ǫn, 2ǫn, ...}, the price vector is the optimal dual solution to the following linear program: maximizex ℓ

t=1 πtxt

subject to ℓ

t=1 aitxt ≤ (1 − hℓ) ℓ nbi

i = 1, ..., m 0 ≤ xt ≤ 1 t = 1, ..., ℓ where hℓ = ǫ n ℓ ; and this price vector is used to determine the allocation for the next immediate period.

Yinyu Ye Online LP, ICERM 2014

slide-37
SLIDE 37

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Geometric Pace/Grid of Price Updating

Yinyu Ye Online LP, ICERM 2014

slide-38
SLIDE 38

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Comments on Dynamic Learning Algorithm

◮ In the dynamic algorithm, we update the prices log2 (1/ǫ)

times during the entire time horizon.

Yinyu Ye Online LP, ICERM 2014

slide-39
SLIDE 39

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Comments on Dynamic Learning Algorithm

◮ In the dynamic algorithm, we update the prices log2 (1/ǫ)

times during the entire time horizon.

◮ The numbers hℓ play an important role in improving the

condition on B in the main theorem. It basically balances the probability that the inventory ever gets violated and the lost

  • f revenue due to the factor 1 − hℓ.

Yinyu Ye Online LP, ICERM 2014

slide-40
SLIDE 40

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Comments on Dynamic Learning Algorithm

◮ In the dynamic algorithm, we update the prices log2 (1/ǫ)

times during the entire time horizon.

◮ The numbers hℓ play an important role in improving the

condition on B in the main theorem. It basically balances the probability that the inventory ever gets violated and the lost

  • f revenue due to the factor 1 − hℓ.

◮ Choosing large hℓ (more conservative) at the beginning

periods and smaller hℓ (more aggressive) at the later periods,

  • ne can now control the loss of revenue by an ǫ order while

the required size of B can be weakened by an ǫ factor.

Yinyu Ye Online LP, ICERM 2014

slide-41
SLIDE 41

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Related Work on Random-Permutation

Sufficient Condition Learning Kleinberg [2005] B ≥ 1

ǫ2 , for m = 1

Dynamic

Yinyu Ye Online LP, ICERM 2014

slide-42
SLIDE 42

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Related Work on Random-Permutation

Sufficient Condition Learning Kleinberg [2005] B ≥ 1

ǫ2 , for m = 1

Dynamic Devanur et al [2009] OPT ≥ m2 log(n)

ǫ3

One-time

Yinyu Ye Online LP, ICERM 2014

slide-43
SLIDE 43

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Related Work on Random-Permutation

Sufficient Condition Learning Kleinberg [2005] B ≥ 1

ǫ2 , for m = 1

Dynamic Devanur et al [2009] OPT ≥ m2 log(n)

ǫ3

One-time Feldman et al [2010] B ≥ m log n

ǫ3

and OPT ≥ m log n

ǫ

One-time

Yinyu Ye Online LP, ICERM 2014

slide-44
SLIDE 44

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Related Work on Random-Permutation

Sufficient Condition Learning Kleinberg [2005] B ≥ 1

ǫ2 , for m = 1

Dynamic Devanur et al [2009] OPT ≥ m2 log(n)

ǫ3

One-time Feldman et al [2010] B ≥ m log n

ǫ3

and OPT ≥ m log n

ǫ

One-time Agrawal et al [2010] B ≥ m log n

ǫ2

  • r OPT ≥ m2 log n

ǫ2

Dynamic

Yinyu Ye Online LP, ICERM 2014

slide-45
SLIDE 45

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Related Work on Random-Permutation

Sufficient Condition Learning Kleinberg [2005] B ≥ 1

ǫ2 , for m = 1

Dynamic Devanur et al [2009] OPT ≥ m2 log(n)

ǫ3

One-time Feldman et al [2010] B ≥ m log n

ǫ3

and OPT ≥ m log n

ǫ

One-time Agrawal et al [2010] B ≥ m log n

ǫ2

  • r OPT ≥ m2 log n

ǫ2

Dynamic Molinaro and Ravi [2013] B ≥ m2 log m

ǫ2

Dynamic

Yinyu Ye Online LP, ICERM 2014

slide-46
SLIDE 46

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Related Work on Random-Permutation

Sufficient Condition Learning Kleinberg [2005] B ≥ 1

ǫ2 , for m = 1

Dynamic Devanur et al [2009] OPT ≥ m2 log(n)

ǫ3

One-time Feldman et al [2010] B ≥ m log n

ǫ3

and OPT ≥ m log n

ǫ

One-time Agrawal et al [2010] B ≥ m log n

ǫ2

  • r OPT ≥ m2 log n

ǫ2

Dynamic Molinaro and Ravi [2013] B ≥ m2 log m

ǫ2

Dynamic Kesselheim et al [2014] B ≥ log m

ǫ2

Dynamic* Gupta and Molinaro [2014] B ≥ log m

ǫ2

Dynamic* Table: Comparison of several existing results

Yinyu Ye Online LP, ICERM 2014

slide-47
SLIDE 47

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Summary and Future Questions on OLP

◮ B = log m ǫ2

is now a necessary and sufficient condition (differing by a constant factor).

Yinyu Ye Online LP, ICERM 2014

slide-48
SLIDE 48

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Summary and Future Questions on OLP

◮ B = log m ǫ2

is now a necessary and sufficient condition (differing by a constant factor).

◮ Thus, they are near-optimal online algorithms for a very

general class of online linear programs.

Yinyu Ye Online LP, ICERM 2014

slide-49
SLIDE 49

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Summary and Future Questions on OLP

◮ B = log m ǫ2

is now a necessary and sufficient condition (differing by a constant factor).

◮ Thus, they are near-optimal online algorithms for a very

general class of online linear programs.

◮ The algorithms are distribution-free and/or non-parametric,

thereby robust to distribution/data uncertainty.

Yinyu Ye Online LP, ICERM 2014

slide-50
SLIDE 50

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Summary and Future Questions on OLP

◮ B = log m ǫ2

is now a necessary and sufficient condition (differing by a constant factor).

◮ Thus, they are near-optimal online algorithms for a very

general class of online linear programs.

◮ The algorithms are distribution-free and/or non-parametric,

thereby robust to distribution/data uncertainty.

◮ The dynamic learning has the feature of

“learning-while-doing”, and is provably better than one-time learning by a factor.

Yinyu Ye Online LP, ICERM 2014

slide-51
SLIDE 51

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Summary and Future Questions on OLP

◮ B = log m ǫ2

is now a necessary and sufficient condition (differing by a constant factor).

◮ Thus, they are near-optimal online algorithms for a very

general class of online linear programs.

◮ The algorithms are distribution-free and/or non-parametric,

thereby robust to distribution/data uncertainty.

◮ The dynamic learning has the feature of

“learning-while-doing”, and is provably better than one-time learning by a factor.

◮ Buy-and-sell or double market?

Yinyu Ye Online LP, ICERM 2014

slide-52
SLIDE 52

Online Linear Programming Main Results and Key Ideas Related and More Recent Work

Summary and Future Questions on OLP

◮ B = log m ǫ2

is now a necessary and sufficient condition (differing by a constant factor).

◮ Thus, they are near-optimal online algorithms for a very

general class of online linear programs.

◮ The algorithms are distribution-free and/or non-parametric,

thereby robust to distribution/data uncertainty.

◮ The dynamic learning has the feature of

“learning-while-doing”, and is provably better than one-time learning by a factor.

◮ Buy-and-sell or double market? ◮ price-posting market?

Yinyu Ye Online LP, ICERM 2014