Adaptive Approximation for Multivariate Linear Problems with Inputs - - PowerPoint PPT Presentation

adaptive approximation for multivariate linear problems
SMART_READER_LITE
LIVE PREVIEW

Adaptive Approximation for Multivariate Linear Problems with Inputs - - PowerPoint PPT Presentation

Adaptive Approximation for Multivariate Linear Problems with Inputs Lying in a Cone Fred J. Hickernell Department of Applied Mathematics Center for Interdisciplinary Scientific Computation Illinois Institute of Technology hickernell@iit.edu


slide-1
SLIDE 1

Adaptive Approximation for Multivariate Linear Problems with Inputs Lying in a Cone

Fred J. Hickernell

Department of Applied Mathematics Center for Interdisciplinary Scientific Computation Illinois Institute of Technology

hickernell@iit.edu mypages.iit.edu/~hickernell

Joint work with Yuhan Ding, Peter Kritzer, and Simon Mak

This work partially supported by NSF-DMS-1522687 and NSF-DMS-1638521 (SAMSI) RICAM Workshop on Multivariate Algorithms and Information-Based Complexity, November 9, 2018

slide-2
SLIDE 2

Thank you

Thank you all for your participation

slide-3
SLIDE 3

Thank you

Thank you all for your participation Thanks to Peter Kritzer and Annette Weihs for doing all the work of organizing

slide-4
SLIDE 4

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Context

Tidbits from talks this week My reponse Henryk, Klaus, Greg, Stefan, Erich, ... Many results for f ∈ Let’s obtain analogous re- sults for f ∈ ? Houman Let’s learn the appropriate ker- nel from the function data Will only work for nice functions in a Klaus “This adaptive algorithm has no theory” We want to construct adaptive algorithms with theory Henryk Tractability Yes! Greg POD weights Yes! Mac Function values are expensive Yes!

3/21

slide-5
SLIDE 5

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Multivariate Linear Problems

Given f ∈ F find S(f) ∈ G, where

S : F → G is linear, e.g., S(f) =

  • Rd f(x) ̺(x) dx

S(f) = f S(f) = ∂f ∂x1 −∇2S(f) = f, S(f) = 0 on boundary

Successful algorithms

A(C, Λ) := {A : C × (0, ∞) → G such that S(f) − A(f, ε)G ε ∀f ∈ C ⊆ F, ε > 0}

where A(f, ε) depends on function values, Λstd, Fourier coefficients, Λser, or any linear function- als, Λall, e.g.,

Sapp(f, n) =

n

  • i=1

f(xi) gi, gi ∈ G Sapp(f, n) =

n

  • i=1
  • fi gi,

gi ∈ G Sapp(f, n) =

n

  • i=1

Li(f) gi, gi ∈ G A(f, ε) = Sapp(f, n)+ stopping criterion C is a

4/21

slide-6
SLIDE 6

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Issues

Given f ∈ F find S(f) ∈ G, where

S : F → G is linear

Successful algorithms

A(C, Λ) := {A : C×(0, ∞) → G such that S(f) − A(f, ε)G ε ∀f ∈ C ⊆ F, ε > 0}

where A(f, ε) depends on function values,

Λstd, Fourier coefficients, Λser, or any linear

functionals, Λall Solvability1: A(C, Λ) = ∅ Construction: Find A ∈ A(C, Λ)

1Kunsch, R. J., Novak, E. & Rudolf, D. Solvable Integration Problems and Optimal Sample Size Selection.

Journal of Complexity. To appear (2018).

2Traub, J. F., Wasilkowski, G. W. & Woźniakowski, H. Information-Based Complexity. (Academic Press, Boston,

1988).

3Novak, E. & Woźniakowski, H. Tractability of Multivariate Problems Volume I: Linear Information. EMS Tracts

in Mathematics 6 (European Mathematical Society, Zürich, 2008).

5/21

slide-7
SLIDE 7

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Issues

Given f ∈ F find S(f) ∈ G, where

S : F → G is linear

Successful algorithms

A(C, Λ) := {A : C×(0, ∞) → G such that S(f) − A(f, ε)G ε ∀f ∈ C ⊆ F, ε > 0}

where A(f, ε) depends on function values,

Λstd, Fourier coefficients, Λser, or any linear

functionals, Λall Solvability1: A(C, Λ) = ∅ Construction: Find A ∈ A(C, Λ) Cost: cost(A, f, ε) = # of function data cost(A, C, ε, ρ) = max

f∈C∩Bρ

cost(A, f, ε)

Bρ := {f ∈ F : fF ρ}

Complexity2: comp(A(C, Λ), ε, ρ)

=

min

A∈A(C,Λ) cost(A, C, ε, ρ)

Optimality: cost(A, C, ε, ρ) comp(A(C, Λ), ωε, ρ)

1Kunsch, R. J., Novak, E. & Rudolf, D. Solvable Integration Problems and Optimal Sample Size Selection.

Journal of Complexity. To appear (2018).

2Traub, J. F., Wasilkowski, G. W. & Woźniakowski, H. Information-Based Complexity. (Academic Press, Boston,

1988).

3Novak, E. & Woźniakowski, H. Tractability of Multivariate Problems Volume I: Linear Information. EMS Tracts

in Mathematics 6 (European Mathematical Society, Zürich, 2008).

5/21

slide-8
SLIDE 8

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Issues

Given f ∈ F find S(f) ∈ G, where

S : F → G is linear

Successful algorithms

A(C, Λ) := {A : C×(0, ∞) → G such that S(f) − A(f, ε)G ε ∀f ∈ C ⊆ F, ε > 0}

where A(f, ε) depends on function values,

Λstd, Fourier coefficients, Λser, or any linear

functionals, Λall Solvability1: A(C, Λ) = ∅ Construction: Find A ∈ A(C, Λ) Cost: cost(A, f, ε) = # of function data cost(A, C, ε, ρ) = max

f∈C∩Bρ

cost(A, f, ε)

Bρ := {f ∈ F : fF ρ}

Complexity2: comp(A(C, Λ), ε, ρ)

=

min

A∈A(C,Λ) cost(A, C, ε, ρ)

Optimality: cost(A, C, ε, ρ) comp(A(C, Λ), ωε, ρ) Tractability3: comp(A(C, Λ), ε, ρ) Cρpε−pdq

1Kunsch, R. J., Novak, E. & Rudolf, D. Solvable Integration Problems and Optimal Sample Size Selection.

Journal of Complexity. To appear (2018).

2Traub, J. F., Wasilkowski, G. W. & Woźniakowski, H. Information-Based Complexity. (Academic Press, Boston,

1988).

3Novak, E. & Woźniakowski, H. Tractability of Multivariate Problems Volume I: Linear Information. EMS Tracts

in Mathematics 6 (European Mathematical Society, Zürich, 2008).

5/21

slide-9
SLIDE 9

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Issues

Given f ∈ F find S(f) ∈ G, where

S : F → G is linear

Successful algorithms

A(C, Λ) := {A : C×(0, ∞) → G such that S(f) − A(f, ε)G ε ∀f ∈ C ⊆ F, ε > 0}

where A(f, ε) depends on function values,

Λstd, Fourier coefficients, Λser, or any linear

functionals, Λall Solvability1: A(C, Λ) = ∅ Construction: Find A ∈ A(C, Λ) Cost: cost(A, f, ε) = # of function data cost(A, C, ε, ρ) = max

f∈C∩Bρ

cost(A, f, ε)

Bρ := {f ∈ F : fF ρ}

Complexity2: comp(A(C, Λ), ε, ρ)

=

min

A∈A(C,Λ) cost(A, C, ε, ρ)

Optimality: cost(A, C, ε, ρ) comp(A(C, Λ), ωε, ρ) Tractability3: comp(A(C, Λ), ε, ρ) Cρpε−pdq Implementation in open source software

1Kunsch, R. J., Novak, E. & Rudolf, D. Solvable Integration Problems and Optimal Sample Size Selection.

Journal of Complexity. To appear (2018).

2Traub, J. F., Wasilkowski, G. W. & Woźniakowski, H. Information-Based Complexity. (Academic Press, Boston,

1988).

3Novak, E. & Woźniakowski, H. Tractability of Multivariate Problems Volume I: Linear Information. EMS Tracts

in Mathematics 6 (European Mathematical Society, Zürich, 2008).

5/21

slide-10
SLIDE 10

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Cones ×

Ball Bρ := {f ∈ F : fF ρ} (Non-Convex) Cone C Assume set of inputs, C ⊆ F, is a cone, not a ball Cone means f ∈ C =

⇒ af ∈ C ∀a ∈ R

Cones are unbounded If we can bound the

  • S(f) − Sapp(f, n)
  • G for f ∈ cone, then we can typically also bound

the error for af Philosophy: What we cannot observe about f is not much worse than what we can

  • bserve about f

6/21

slide-11
SLIDE 11

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

“But I Like !”

How might you construct an algorithm if you insist on using ? Step 1 Pick with a default radius ρ, and assume input f ∈ Bρ Step 2 Choose n large enough so that

  • S(f) − Sapp(f, n)
  • G
  • S − Sapp(·, n)
  • F→G ρ ε

where Sapp(f, n) =

n

  • i=1

Li(f)gi

then return A(f, ε) = Sapp(f, n)

7/21

slide-12
SLIDE 12

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

“But I Like !”

How might you construct anadaptive algorithm if you insist on using ? Step 1 Pick with a default radius ρ, and assume input f ∈ Bρ Step 2 Choose n large enough so that

  • S(f) − Sapp(f, n)
  • G
  • S − Sapp(·, n)
  • F→G ρ ε

where Sapp(f, n) =

n

  • i=1

Li(f)gi

Step 3 Let fmin ∈ F be the minimum norm interpolant of the data L1(f), . . . , Ln(f) Step 4 If C

  • fmin
  • F ρ for some preset inflation factor, C,

then return A(f, ε) = Sapp(f, n);

  • therwise, choose ρ = 2C
  • fmin
  • F, and go to Step 2

This succeeds for the cone defined as those functions in F whose norms are not much larger than their minimum norm interpolants.

7/21

slide-13
SLIDE 13

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

When Is (S : C ⊆ F → G, Λ) Solvable?

A(C, Λ) := {A : C × (0, ∞) → G such that S(f) − A(f, ε)G ε ∀f ∈ C ⊆ F, ε > 0}

where A(f, ε) depends on {Li(f)}n

i=1 ∈ Λn ⊆ F∗n

Definition

(S : C ⊆ F → G, Λ) solvable ⇐ ⇒ A(C, Λ) = ∅

Lemma

f1, f2 ∈ C and A(f1, ε) = A(f2, ε) = ⇒ S(f1 − f2)G 2ε

Corollary

(S : C ⊆ F → G, Λ) solvable and ∃f ∈ C, ε > 0 with A(f, ε) = A(0, ε) = ⇒ S(f) = 0

Theorem

(S : C ⊆ F → G, Λ) solvable and C is a vector space ⇐ ⇒ ∃ n ∈ N, L ∈ Λn, g ∈ Gn such that S(f) =

n

  • i=1

Li(f)gi ∀f ∈ C exactly

Proof

E.g.

  • [0,1]d ·(x) dx : C1,...,1[0, 1]d → R, Λstd/all
  • is unsolvable/solvable

8/21

slide-14
SLIDE 14

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Inputs and Outputs

Input: f =

  • j∈N
  • f(j)uj,

Output: g =

  • j∈N

^ g(j)vj, S(uj) = vj,

E.g., Integration: S(uj) =

  • [0,1]d uj(x) dx = vj,

Function recovery: S(uj) = uj = vj, Fixed sample size algorithm: Sapp(f, n) =

n

  • i=1

Li(f)gi A(f, ε) = Sapp(f, n) + stopping criterion

Three scenarios presented here: Integration use Λstd, but cost, complexity, etc. lacking General Linear Problems use Λser, have cost, complexity, and optimality Function Approximation use Λser, learn coordinate and smoothness weights

9/21

slide-15
SLIDE 15

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Integration Using Function Values, Λstd4

S(f) =

  • [0,1]d f(x) dx

f =

  • j∈N
  • f(j)uj,

uj = Cosine/Sine or Walsh, S(uj) = δj,0 Sapp(f, n) = 1 n

n

  • i=1

f(xi)

  • S(f) − Sapp(f, n)
  • 0=j∈dual set
  • f(j)
  • 4Dick, J., Kuo, F. & Sloan, I. H. High dimensional integration — the Quasi-Monte Carlo way. Acta Numer. 22,

133–288 (2013).

10/21

slide-16
SLIDE 16

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Complex Exponential and Walsh Bases for Cubature

Cosine & Sine Walsh

11/21

slide-17
SLIDE 17

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Cones of Integrands Whose Fourier Series Coefficients Decay Steadily5

n0 = 0, nk = 2k−1, k ∈ N, j ordered,

true coef. =

f(j) ≈ fdisc(j) = discrete coef. σk(f) =

nk

  • i=nk−1+1
  • f(ji)
  • ,

^ σk,ℓ(f) =

nk

  • i=nk−1+1

  • m=1
  • f(ji+mnℓ)
  • ,

σdisc,k(f) =

nk

  • i=nk−1+1
  • fdisc(ji)
  • C :=
  • f ∈ F : σℓ(f) a1bℓ−k

1

σk(f), ^ σk,ℓ(f) a2bℓ−k

2

σℓ(f) ∀k ℓ

  • a1, a2 > 1 > b1, b2

Sapp(f, nk) = 1 nk

nk

  • i=1

f(xi)

  • S(f) − Sapp(f, nk)
  • 0=j∈dual set
  • f(j)
  • C(k)σdisc,k(f)

A(f, ε) = S(f, nk) for the smallest k satisfying C(k)σdisc,k(f) ε

Have cost(f, A, ε); No cost(A, C, ε, ρ), comp(A(C, Λstd), ε, ρ), or tractability results yet

5H., F. J. & Jiménez Rugama, Ll. A. Reliable Adaptive Cubature Using Digital Sequences. in Monte Carlo and

Quasi-Monte Carlo Methods: MCQMC, Leuven, Belgium, April 2014 (eds Cools, R. & Nuyens, D.) 163. arXiv:1410.8615 [math.NA] (Springer-Verlag, Berlin, 2016), 367–383, Jiménez Rugama, Ll. A. & H., F. J. Adaptive Multidimensional Integration Based on Rank-1 Lattices. in Monte Carlo and Quasi-Monte Carlo Methods: MCQMC, Leuven, Belgium, April 2014 (eds Cools, R. & Nuyens, D.) 163. arXiv:1411.1966 (Springer-Verlag, Berlin, 2016), 407–422.

12/21

slide-18
SLIDE 18

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Integrands in C Aren’t Fuzzy

13/21

slide-19
SLIDE 19

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Option Pricing6

fair price =

  • Rd e−rT max

 1 d

d

  • j=1

Sj − K, 0   e−zTz/2 (2π)d/2 dz ≈ $13.12 Sj = S0e(r−σ2/2)jT/d+σxj = stock price at time jT/d, x = Az,

AAT = Σ =

  • min(i, j)T/d

d

i,j=1,

T = 1/4, d = 13 here

Error Median Worst 10% Worst 10% Tolerance Method Error Accuracy

n

Time (s)

1E−2

IID diff

2E−3 100% 6.1E7 33.000 1E−2

  • Sh. Latt.

PCA

1E−3 100% 1.6E4 0.041 1E−2

  • Scr. Sobol’

PCA

1E−3 100% 1.6E4 0.040 1E−2

  • Scr. Sob. cont. var. PCA

2E−3 100% 4.1E3 0.019

6Choi, S.-C. T., Ding, Y., H., F. J., Jiang, L., Jiménez Rugama, Ll. A., Li, D., Jagadeeswaran, R., Tong, X.,

Zhang, K., et al. GAIL: Guaranteed Automatic Integration Library (Versions 1.0–2.2). MATLAB software. 2013–2017. http://gailgithub.github.io/GAIL_Dev/.

14/21

slide-20
SLIDE 20

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Solving General Linear Problems Using Series Coefficients, Λser7

F :=   f =

  • j∈N
  • f(j)uj : fF :=
  • f(j)
  • λj
  • j∈N
  • 2

   λ affects

convergence rate & tractability

G :=

  • g =
  • j∈N

^ g(j)vj : gG :=

  • ^

g

  • 2
  • ,

vj = S(uj) λj1 λj2 · · · , n0 < n1 < n2 < · · · , σk(f) =

  • nk
  • i=nk−1+1
  • f(ji)
  • 2,

k ∈ N C :=

  • f ∈ F : σℓ(f) abℓ−kσk(f)

∀k ℓ

  • a > 1 > b

series coef. decay steadily

Sapp(f, nk) =

nk

  • i=1
  • f(ji)vji is optimal for fixed nk,
  • S(f) − Sapp(f, nk)
  • G abσk(f)

√ 1 − b2 A(f, ε) = Sapp(f, nk) for the smallest k satisfying σk(f) ε √ 1 − b2 ab

7Ding, Y., H., F. J. & Jiménez Rugama, Ll. A. An Adaptive Algorithm Employing Continuous Linear Functionals.

2018+.

15/21

slide-21
SLIDE 21

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Solving General Linear Problems Using Series Coefficients, Λser7

λj1 λj2 · · · , n0 < n1 < n2 < · · · , σk(f) =

  • nk
  • i=nk−1+1
  • f(ji)
  • 2,

k ∈ N C :=

  • f ∈ F : σℓ(f) abℓ−kσk(f)

∀k ℓ

  • a > 1 > b

series coef. decay steadily

Sapp(f, nk) =

nk

  • i=1
  • f(ji)vji is optimal for fixed nk,
  • S(f) − Sapp(f, nk)
  • G abσk(f)

√ 1 − b2 A(f, ε) = Sapp(f, nk) for the smallest k satisfying σk(f) ε √ 1 − b2 ab

cost(A, C, ε, ρ) = nℓ†,

ℓ† min

  • ℓ ∈ N : ρ2

ε2 (1 − b2) a2b2 ℓ−1

  • k=1

b2(k−ℓ) a2λ2

nk−1+1

+ 1 λ2

nℓ−1+1

  • cost(A, C, ε, ρ) essentially no worse than comp(A(C, Λall), ε, ρ)

No tractability results

7Ding, Y., H., F. J. & Jiménez Rugama, Ll. A. An Adaptive Algorithm Employing Continuous Linear Functionals.

2018+.

15/21

slide-22
SLIDE 22

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Function Approximation when Function Values Are Expensive8

F :=   f =

  • j∈Nd
  • f(j)uj : fF :=
  • f(j)
  • λj
  • j∈N

   λ affects

convergence rate & tractability

G :=

  • g =
  • j∈Nd

^ g(j)vj : gG :=

  • ^

g

  • 1
  • ,

vj = S(uj) = uj λj = Γj0

d

  • ℓ=1

jℓ>0

γℓsjℓ    γℓ = coordinate importance Γr = order size sj = smoothness degree

POSD weights reflect effect sparsity, effect hierarchy, effect heredity, and effect smoothness

λj1 λj2 · · · , Sapp(f, n) =

n

  • i=1
  • f(ji)uji,
  • S(f) − Sapp(f, nk)
  • G fF
  • i=n+1

λji C :=

  • f ∈ F : those functions for which fF can be inferred from a pilot sample
  • 8Wu, C. F. J. & Hamada, M. Experiments: Planning, Analysis, and Parameter Design Optimization. (John Wiley

& Sons, Inc., New York, 2000), Kuo, F. Y., Schwab, C. & Sloan, I. H. Quasi-Monte Carlo Finite Element Methods for a Class of Elliptic Partial Differential Equations with Random Coefficients. SIAM J. Numer. Anal. 50, 3351–3374 (2012).

16/21

slide-23
SLIDE 23

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Legendre and Chebyshev Bases for Function Approximation

Legendre Chebyshev

17/21

slide-24
SLIDE 24

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Cheng and Sandu Function9

Chebyshev polynomials, Order weights Γk = 1, Coordinate weights γℓ inferred, Smoothness weights sj inferred,

Λstd

9Bingham, D. The Virtual Library of Simulation Experiments: Test Functions and Data Sets. 2017. https://www.sfu.ca/~ssurjano/index.html (2017).

18/21

slide-25
SLIDE 25

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Take Home Messages

Assuming that the input functions lie in convex cones allow us to construct adaptive algorithms Cone definitions reflect prior beliefs and/or practical considerations Demonstration of concept

Integration using Λstd

Constructed A ∈ A(C, Λstd) No cost, complexity, or tractability yet

General linear problems

Constructed A ∈ A(C, Λser), upper bound on cost(A, C, ε, ρ), lower bound on comp(A(C, Λall), ε, ρ), optimality No tractability yet

Function approximation (recovery)

Constructed A ∈ A(C, Λser), algorithm learns weights, works in practice for Λstd Remaining theory in progress

Much to be done

19/21

slide-26
SLIDE 26

Thank you

These slides are available at

speakerdeck.com/fjhickernell/ricam-2018-nov

slide-27
SLIDE 27

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Kunsch, R. J., Novak, E. & Rudolf, D. Solvable Integration Problems and Optimal Sample Size Selection. Journal of Complexity. To appear (2018). Traub, J. F., Wasilkowski, G. W. & Woźniakowski, H. Information-Based Complexity. (Academic Press, Boston, 1988). Novak, E. & Woźniakowski, H. Tractability of Multivariate Problems Volume I: Linear

  • Information. EMS Tracts in Mathematics 6 (European Mathematical Society, Zürich,

2008). Dick, J., Kuo, F. & Sloan, I. H. High dimensional integration — the Quasi-Monte Carlo

  • way. Acta Numer. 22, 133–288 (2013).

H., F. J. & Jiménez Rugama, Ll. A. Reliable Adaptive Cubature Using Digital Sequences. in Monte Carlo and Quasi-Monte Carlo Methods: MCQMC, Leuven, Belgium, April 2014 (eds Cools, R. & Nuyens, D.) 163. arXiv:1410.8615 [math.NA] (Springer-Verlag, Berlin, 2016), 367–383.

20/21

slide-28
SLIDE 28

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Jiménez Rugama, Ll. A. & H., F. J. Adaptive Multidimensional Integration Based on Rank-1 Lattices. in Monte Carlo and Quasi-Monte Carlo Methods: MCQMC, Leuven, Belgium, April 2014 (eds Cools, R. & Nuyens, D.) 163. arXiv:1411.1966 (Springer-Verlag, Berlin, 2016), 407–422. Choi, S.-C. T. et al. GAIL: Guaranteed Automatic Integration Library (Versions 1.0–2.2). MATLAB software. 2013–2017. http://gailgithub.github.io/GAIL_Dev/. Ding, Y., H., F. J. & Jiménez Rugama, Ll. A. An Adaptive Algorithm Employing Continuous Linear Functionals. 2018+. Wu, C. F. J. & Hamada, M. Experiments: Planning, Analysis, and Parameter Design

  • Optimization. (John Wiley & Sons, Inc., New York, 2000).

Kuo, F. Y., Schwab, C. & Sloan, I. H. Quasi-Monte Carlo Finite Element Methods for a Class of Elliptic Partial Differential Equations with Random Coefficients. SIAM J. Numer.

  • Anal. 50, 3351–3374 (2012).

20/21

slide-29
SLIDE 29

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Bingham, D. The Virtual Library of Simulation Experiments: Test Functions and Data

  • Sets. 2017. https://www.sfu.ca/~ssurjano/index.html (2017).

(eds Cools, R. & Nuyens, D.) Monte Carlo and Quasi-Monte Carlo Methods: MCQMC, Leuven, Belgium, April 2014. 163 (Springer-Verlag, Berlin, 2016).

21/21

slide-30
SLIDE 30

Introduction Integration General Linear Problems Function Approximation Summary References Bonus

Proof of Theorem for Solvability on a Vector Space

Let the cone C be a vector space and let

A be a successful algorithm ε > 0 be any positive tolerance {L1, . . . , LM} ⊂ Λ be the linear functionals used by A(0, ε), and {L1, . . . , Lm} be a basis for span({L1, . . . , LM}) n = min(m, dim(C)) {f1, . . . , fn} ⊂ C satisfy Li(fj) = δi,j, i = 1, . . . , n, j = 1, . . . , m

For any f ∈ C, let

f = f −

n

  • i=1

Li(f)fi, and note that Lj( f) = 0 for j = 1, . . . , M. Thus, A( f, ε) = A(0, ε), and so by the

Corollary ,

0 = S( f) = S(f) −

n

  • i=1

Li(f)S(fi),

which implies S(f) =

n

  • i=1

Li(f)S(fi)

Back

21/21