Polynomials Paul Valiant Brown University (Based on joint work - - PowerPoint PPT Presentation

polynomials
SMART_READER_LITE
LIVE PREVIEW

Polynomials Paul Valiant Brown University (Based on joint work - - PowerPoint PPT Presentation

Three Perspectives on Orthogonal Polynomials Paul Valiant Brown University (Based on joint work with Gregory Valiant, mostly STOC11: Estimating the Unseen: An n/log(n)-sample Estimator for Entropy and Support Size, Shown Optimal via New


slide-1
SLIDE 1

Three Perspectives on Orthogonal Polynomials Paul Valiant

Brown University

(Based on joint work with Gregory Valiant, mostly STOC’11: “Estimating the Unseen: An n/log(n)-sample Estimator for Entropy and Support Size, Shown Optimal via New CLTs”)

slide-2
SLIDE 2

Chebyshev Laguerre Hermite “cosine” “cosine times exponential” “cosine over Gaussian”

Seems like:

“Seems,” madam? Nay, it is. I know not “seems.” – Hamlet

Structure of this talk: 3 polynomial challenges… and solutions

slide-3
SLIDE 3

Challenge 1: Poisson bumps  thinnest bumps 𝑞𝑝𝑗 𝜇, 𝑙 = 𝜇𝑙𝑓−𝑙 𝑙!

Linear transform

  • 1. Thin as possible
  • 2. σ

=1

Motivation: Given an event with probability p, 𝑞𝑝𝑗 𝑞𝑜, 𝑙 captures the probability of it

  • ccurring exactly k times in

Poi(n) samples. Let Fk be total number of events that were

  • bserved k times. Fk captures

probabilities from . Is there a linear combination of Fkthat captures ? general log n factor improvement in resolution, #samples

Thm:

x-resolution y-resolution

 ?

(with bounded coeffs)

x x k k+1 k+2 k-2 k-1

slide-4
SLIDE 4

Chebyshev Polynomials 𝑈

𝑘 cos 𝑦 = cos(𝑘𝑦)

Chebyshev is exactly like cosine, except on distorted x-axis

  • 1. Thin as possible
  • 2. σ

=1

x-resolution y-resolution

j Both unchanged under x-axis distortion!

New question: thin cosine bumps

(with bounded coeffs)

slide-5
SLIDE 5

Thinnest Cosine Bumps Thinnest linear combination of cos(𝑘𝑦) for 𝑘 < 𝑐:

≈ 1/𝑐 (Intuition: Fourier transform of degree b gives resolution 1/b)

  • 1. Thin as possible

(with bounded coeffs)

  • 2. σ

=1

x-resolution y-resolution

j

Sum of all possible x- translated bumps is constant (Trig functions are well- behaved under x-translation)

slide-6
SLIDE 6

Chebyshev Takeaways: 𝑞𝑝𝑗 𝜇, 𝑙 = 𝜇𝑙𝑓−𝑙 𝑙!

Linear transform

  • 1. Thin as possible

(with coeffs << n)

  • 2. σ

=1

Motivation: Given an event with probability p, 𝑞𝑝𝑗 𝑞𝑜, 𝑙 captures the probability of it

  • ccurring exactly k times in

Poi(n) samples. Let Fk be total number of events that were

  • bserved k times. Fk captures

probabilities from . Is there a linear combination of Fkthat captures ? general log n factor improvement in resolution, #samples

Thm:

x-resolution y-resolution

exp(b) b times thinner Thus: 𝑐 = 𝜄(log 𝑜)

(Modulo x-axis distortion) “polynomials are cosines”

slide-7
SLIDE 7

Degree j polynomial with roots at , 2; and all remaining roots have much larger derivative, growing exponentially with x

Challenge 2: Exponentially Growing Derivatives

Roots close together have small derivatives Pulling a root farther away increases its derivative, but

  • nly polynomially

Success requires a delicate balancing act! Find:

slide-8
SLIDE 8

Fact: If P is a degree j polynomial with distinct real roots {xi}, then the signed measure hP having point mass 1/𝑄′(𝑦𝑗) at each root xi is orthogonal to all polynomials of degree ≤j-2

Orthogonal to Polynomials

Task: find P such that 𝑄′(𝑦𝑗) grows exponentially in xi

Motivation: Want to construct a pair of distributions g+,g- that are, respectively, close to the uniform distributions on T and 2T elements, but where for each (small) k, the expected number of domain elements seen k times from Poi(n) samples is identical for g+,g-. Essentially: find a signed measure g(x) that is 1) Orthogonal to 𝑞𝑝𝑗 𝑦, 𝑙 =

𝑦𝑙𝑓−𝑙 𝑙!

for each small k, 2) Has most of its positive mass at 1/T and most of its negative mass at 1/(2T) 𝑕 𝑦 ≜ 𝑓𝑦ℎ 𝑦 Essentially: find a signed measure ℎ(𝑦) that is 1) Orthogonal to all degree ≤k polynomials 2) Has most of its positive mass at 1/T and most of its negative mass at 1/(2T)

  • and otherwise decays ≪ 𝑓−𝑦
slide-9
SLIDE 9

Laguerre Polynomials

Defined by 𝑀𝑜 𝑦 = 𝑓𝑦 𝑒𝑜

𝑒𝑦𝑜 𝑓−𝑦𝑦𝑜 𝑜!

and

  • rthogonal as: ׬

∞ 𝑀𝑜 𝑦 𝑀𝑛 𝑦 𝑓−𝑦𝑒𝑦 = [𝑛 = 𝑜]

Why should the derivative be so nicely behaved at its roots, in particular, growing exponentially?

Many differential equations, including 𝑤′′ + 4𝑜 + 2 − 𝑦2 +

1 4𝑦2 𝑤 = 0

Almost harmonic motion, v→sine Nicely spaced zeros, and max derivative at the zeros

Transform the Laguerre: 𝑤 = 𝑓−𝑦2/2 𝑦 ⋅ 𝑀𝑜(𝑦2)

slide-10
SLIDE 10

The Construction

We want a signed measure g on the positive reals that:

  • Is orthogonal to low degree polynomials
  • Decays exponentially fast
  • Its positive portion has most of its mass at 2𝜗
  • Its negative portion has most of its mass at 𝜗

Recall: Theorem: p+ is “close” to Un/2 , and p- is “close” to Un , and p+ and p- are indistinguishable via cn/log n samples p+ p-

(Modulo diff-eq distortion) “polynomials are 𝑓𝑦sin(𝑦)”

slide-11
SLIDE 11

Challenge 3: exponentially good bump approximations

Find a linear combination over j of poi(x,j) that approximates poi(x,k)2 to within , using coefficients ≤1/

These look like Gaussians! 1) What’s the answer for Gaussians? Think of =1/exp(j) 2) Analyze via Hermite polynomials instead Motivation: Previously, constructed lowerbound distributions g+,g- where expectation of every measurement

  • matched. Lower bound? No… until

we show variances match too. Aim: show that variances can be approximated as linear combinations

  • f expectations, with moderate

coefficients; thus matching means implies matching variances. Since means come from poi(j,x), second moments come from poi(j,x)2.

slide-12
SLIDE 12

Approximating “Thin” Gaussians as Linear Combinations of Gaussians

What do we convolve a Gaussian with to approximate a thinner Gaussian? (Other direction is easy, since convolving Gaussians adds their variances) “Blurring is easy, unblurring is hard”  can only do it approximately Now: what do we multiply a Gaussian with to approximate a fatter Gaussian? 𝑓−𝑦2 ⋅ ? ? ? = 𝑓−𝑦2/2 𝑓𝑦2/2 Problem: blows up How to analyze? Fourier transform! Convolution becomes multiplication Result: Can approximate to within 𝜗 using coefficients no bigger than 1/𝜗 Answer: if we want to approximate to within 𝜗, we only need to approximate out to where 𝑓−𝑦2/2 = 𝜗. How big is 𝑓𝑦2/2 here? 1/𝜗

slide-13
SLIDE 13

Hermite Polynomials

Poissons seem a lot like Gaussians (I was stuck here for about a month) Idea: xy2 𝑦𝑘𝑓−𝑦 𝑘! = 𝑧2𝑘𝑓−𝑧2 𝑘!

𝑔𝑝𝑣𝑠𝑗𝑓𝑠

1 𝑘! 𝑒2𝑘 𝑒𝑥2𝑘 𝑓−𝑥2 = 1 𝑘! 𝐼2𝑘 𝑥 𝑓−𝑥2 To express any function in this basis, just compute each coefficient as an inner product Which function? Fourier transform of “thin” Poisson, cut off at 𝜗 Proposition: Can approximate Pr[𝑄𝑝𝑗(2𝜇) = 𝑙] to within 𝜗 as a linear combination σ𝑘 𝛽𝑙,𝑘Pr[𝑄𝑝𝑗 𝜇 = 𝑘] with coefficients that sum to σ𝑘 |𝛽𝑙,𝑘| ≤

1 𝜗 200max{

4 𝑙, 24 log 3 2 1

𝜗}

Hermite Polynomials! Sequence of orthogonal polynomials, from which we take the even ones:

  • rthogonal basis for even functions
slide-14
SLIDE 14

Chebyshev Laguerre Hermite “cosine” “cosine times exponential” “cosine over Gaussian”

Seems like:

“Seems,” madam? Nay, it is. I know not “seems.” – Hamlet

Structure of this talk: 3 polynomial challenges… and solutions