Chapter 3: The Second Moment The Probabilistic Method Summer 2020 - - PowerPoint PPT Presentation

β–Ά
chapter 3 the second moment
SMART_READER_LITE
LIVE PREVIEW

Chapter 3: The Second Moment The Probabilistic Method Summer 2020 - - PowerPoint PPT Presentation

Chapter 3: The Second Moment The Probabilistic Method Summer 2020 Freie Universitt Berlin Chapter Overview Introduce the second moment method Survey applications in graph theory and number theory 1 Concentration Inequalities


slide-1
SLIDE 1

Chapter 3: The Second Moment

The Probabilistic Method Summer 2020 Freie UniversitΓ€t Berlin

slide-2
SLIDE 2

Chapter Overview

  • Introduce the second moment method
  • Survey applications in graph theory and number theory
slide-3
SLIDE 3

Β§1 Concentration Inequalities

Chapter 3: The Second Moment The Probabilistic Method

slide-4
SLIDE 4

What Does the Expectation Mean?

Basic fact

  • π‘Œ ≀ 𝔽 π‘Œ

and {π‘Œ β‰₯ 𝔽 π‘Œ } have positive probability

  • Often want more quantitative information
  • What are these positive probabilities?
  • How much below/above the expectation can the random variable be?

Limit laws

  • Law of large numbers
  • Average of independent trials will tend to the expectation
  • Central limit theorem
  • Average will be normally distributed

Not always applicable

  • We often only have a single instance, or lack independence
  • Can still make use of more general bounds
slide-5
SLIDE 5

Markov’s Inequality

Proof

  • Let 𝑔 be the density function for the distribution of π‘Œ
  • 𝔽 π‘Œ = Χ¬

∞ 𝑦𝑔 𝑦 ⅆ𝑦 = Χ¬ 𝑏 𝑦𝑔 𝑦 ⅆ𝑦 + Χ¬ 𝑏 ∞ 𝑦𝑔 𝑦 ⅆ𝑦

β‰₯ Χ¬

𝑏 ∞ 𝑦𝑔 𝑦 ⅆ𝑦 β‰₯ Χ¬ 𝑏 ∞ 𝑏𝑔 𝑦 ⅆ𝑦 = 𝑏 Χ¬ 𝑏 ∞ 𝑔 𝑦 ⅆ𝑦 = 𝑏ℙ(π‘Œ β‰₯ 𝑏)

∎

Moral: 𝔽 π‘Œ small β‡’ π‘Œ typically small Theorem 3.1.1 (Markov’s Inequality) Let π‘Œ be a non-negative random variable, and let 𝑏 > 0. Then β„™ π‘Œ β‰₯ 𝑏 ≀ 𝔽 π‘Œ 𝑏 .

slide-6
SLIDE 6

Theorem 3.1.2 (Chebyshev’s Inequality) Let π‘Œ be a random variable, and let 𝑏 > 0. Then β„™ π‘Œ βˆ’ 𝔽 π‘Œ β‰₯ 𝑏 ≀ Var π‘Œ 𝑏2 .

Chebyshev’s Inequality

Converse? Does 𝔽 π‘Œ large β‡’ π‘Œ typically large?

  • Not necessarily; e.g. π‘Œ = π‘œ2 with probability π‘œβˆ’1, 0 otherwise
  • But such random variables have large variance…

Proof

  • π‘Œ βˆ’ 𝔽 π‘Œ

β‰₯ 𝑏 = π‘Œ βˆ’ 𝔽 π‘Œ

2 β‰₯ 𝑏2

  • Let 𝑍 = π‘Œ βˆ’ 𝔽 π‘Œ

2

  • Then 𝔽 𝑍 = Var π‘Œ
  • Apply Markov’s Inequality

∎

slide-7
SLIDE 7

Corollary 3.1.3 If Var π‘Œ = 𝑝 𝔽 π‘Œ 2 , then β„™ π‘Œ = 0 = 𝑝 1 .

Using Chebyshev

Moral

  • 𝔽 π‘Œ large and Var π‘Œ small β‡’ π‘Œ typically large
  • Special case: showing π‘Œ nonzero

Proof

  • π‘Œ = 0 βŠ†

π‘Œ βˆ’ 𝔽 π‘Œ β‰₯ 𝔽 π‘Œ

  • Chebyshev β‡’ β„™ π‘Œ βˆ’ 𝔽 π‘Œ

β‰₯ 𝔽 π‘Œ ≀ Var π‘Œ

𝔽 π‘Œ 2 = 𝑝(1)

∎

  • In fact, in this case π‘Œ = 1 + 𝑝 1

𝔽[π‘Œ] with high probability

slide-8
SLIDE 8

Typical application

Set-up

  • 𝐹𝑗 events, occurring with probability π‘žπ‘—
  • π‘Œπ‘— = 1𝐹𝑗 their indicator random variables
  • π‘Œ = σ𝑗 π‘Œπ‘— their sum, the number of occurring events

Goal

  • Show that with high probability, some event occurs

Applying Chebyshev

  • Need to show Var π‘Œ = 𝑝 𝔽 π‘Œ 2

Expand the variance

  • Var π‘Œ = Var σ𝑗 π‘Œπ‘— = σ𝑗 Var(π‘Œπ‘—) + Οƒπ‘—β‰ π‘˜ Cov(π‘Œπ‘—, π‘Œ

π‘˜)

slide-9
SLIDE 9

Corollary 3.1.4 Let 𝐹𝑗 be a sequence of events with probabilities π‘žπ‘—, and let π‘Œ count the number of events that occur. Write 𝑗 ∼ π‘˜ if the events 𝐹𝑗 and 𝐹

π‘˜ are

not independent, and let Ξ” = Οƒπ‘—βˆΌπ‘˜ β„™ 𝐹𝑗 ∧ 𝐹

π‘˜ . If 𝔽 π‘Œ β†’ ∞ and

Ξ” = 𝑝 𝔽 π‘Œ 2 , then 𝑄 π‘Œ = 0 = 𝑝 1 .

Some Simplification

Estimating the summands

  • Var π‘Œ = Var σ𝑗 π‘Œπ‘— = σ𝑗 Var(π‘Œπ‘—) + Οƒπ‘—β‰ π‘˜ Cov(π‘Œπ‘—, π‘Œ

π‘˜)

  • Var π‘Œπ‘— = π‘žπ‘— 1 βˆ’ π‘žπ‘— ≀ π‘žπ‘—
  • ∴ σ𝑗 Var(π‘Œπ‘—) ≀ σ𝑗 π‘žπ‘— = σ𝑗 𝔽 π‘Œπ‘— = 𝔽[π‘Œ]
  • Cov π‘Œ, 𝑍 = 𝔽 π‘Œπ‘ βˆ’ 𝔽 π‘Œ 𝔽 𝑍
  • Cov π‘Œ, 𝑍 = 0 if π‘Œ and 𝑍 are independent
  • Otherwise Cov π‘Œπ‘—, π‘Œ

π‘˜ ≀ 𝔽 π‘Œπ‘—π‘Œ π‘˜ = β„™ 𝐹𝑗 ∧ 𝐹 π‘˜

slide-10
SLIDE 10

Any questions?

slide-11
SLIDE 11

Β§2 Thresholds

Chapter 3: The Second Moment The Probabilistic Method

slide-12
SLIDE 12

Lemma 3.2.1 If 𝒬 is a monotone increasing graph property, then β„™ 𝐻 π‘œ, π‘ž ∈ 𝒬 is monotone increasing in π‘ž.

Monotone properties

Graph properties

  • Say a graph 𝒬 is monotone (increasing) if adding edges preserves 𝒬
  • e.g.: containing a subgraph 𝐼 βŠ† 𝐻, having 𝛽 𝐻 < 𝑙, connectivity, …

Proof (Coupling)

  • Sampling 𝐻 π‘œ, π‘ž
  • Assign to each pair of vertices 𝑣, 𝑀 an independent uniform 𝑍

𝑣,𝑀~Unif 0,1

  • Add edge 𝑣, 𝑀 to 𝐻 iff 𝑍

𝑣,𝑀 ≀ π‘ž

  • Each edge appears independently with probability π‘ž
  • If π‘ž ≀ π‘žβ€², then 𝐻 π‘œ, π‘ž βŠ† 𝐻 π‘œ, π‘žβ€² β‡’ if 𝐻 π‘œ, π‘ž ∈ 𝒬, then 𝐻 π‘œ, π‘žβ€² ∈ 𝒬

∎

slide-13
SLIDE 13

Definition 3.2.2 (Thresholds) Given a nontrivial monotone graph property 𝒬, π‘ž0(π‘œ) is a threshold for 𝒬 if β„™ 𝐻 π‘œ, π‘ž ∈ 𝒬 β†’ α‰Š0 if p β‰ͺ π‘ž0 π‘œ , 1 if π‘ž ≫ π‘ž0 π‘œ .

Thresholds

Transitions

  • A monotone property 𝒬 is nontrivial if it is not satisfied by the edgeless graph,

and is satisfied by the complete graph

  • β‡’ β„™ 𝐻 π‘œ, 0 ∈ 𝒬 = 0 and β„™ 𝐻 π‘œ, 1 ∈ 𝒬 = 1
  • Lemma 3.2.1 β‡’ β„™ 𝐻 π‘œ, π‘ž ∈ 𝒬 increases from 0 to 1 as π‘ž does
  • How quickly does this increase happen?
slide-14
SLIDE 14

Proposition 3.2.3 The threshold for 𝐻(π‘œ, π‘ž) to contain a cycle is π‘ž0 π‘œ =

1 π‘œ.

A Cyclic Example

Proof (lower bound)

  • Let π‘Œ = # cycles in 𝐻(π‘œ, π‘ž)
  • For β„“ β‰₯ 3, let π‘Œβ„“ = # 𝐷ℓ βŠ† 𝐻 π‘œ, π‘ž
  • β‡’ π‘Œ = Οƒβ„“=3

π‘œ

π‘Œβ„“

  • Linearity of expectation: 𝔽 π‘Œβ„“ ≀ π‘œβ„“π‘žβ„“
  • β‡’ 𝔽 π‘Œ ≀ Οƒβ„“=3

π‘œ

π‘œπ‘ž β„“ < π‘œπ‘ž 3 Οƒβ„“=0

∞

π‘œπ‘ž β„“ =

π‘œπ‘ž 3 1βˆ’π‘œπ‘ž

  • β‡’ 𝔽 π‘Œ = 𝑝(1) if π‘ž β‰ͺ

1 π‘œ

  • Markov: β„™ 𝐻 π‘œ, π‘ž has a cycle = β„™ π‘Œ β‰₯ 1 ≀ 𝔽 π‘Œ β†’ 0

∎

slide-15
SLIDE 15

Proposition 3.2.3 The threshold for 𝐻(π‘œ, π‘ž) to contain a cycle is π‘ž0 π‘œ =

1 π‘œ.

Cycles Continued

Proof (upper bound)

  • Let π‘ž =

4 π‘œβˆ’1 and set 𝑍 = 𝑓 𝐻 π‘œ, π‘ž

  • Then 𝑍 ∼ Bin

π‘œ 2 , π‘ž

  • β‡’ 𝔽 𝑍 =

π‘œ 2 π‘ž = 2π‘œ

  • β‡’ Var 𝑍 =

π‘œ 2 π‘ž 1 βˆ’ π‘ž < 2π‘œ

  • ∴ Var 𝑍 = 𝑝 𝔽 𝑍 2
  • Chebyshev: β„™ 𝑍 < π‘œ β†’ 0
  • β„™ 𝐻 π‘œ, π‘ž has a cycle β‰₯ β„™ 𝑓 𝐻 π‘œ, π‘ž

β‰₯ π‘œ β†’ 1 ∎

slide-16
SLIDE 16

Existence of Thresholds

Proof (upper bound)

  • Let π‘ž π‘œ = π‘ž0 be such that β„™ 𝐻 π‘œ, π‘ž0 ∈ 𝒬 =

1 2

  • Let 𝐻 ∼ 𝐻1 βˆͺ 𝐻2 βˆͺ β‹― βˆͺ 𝐻𝑛, where each 𝐻𝑗 ∼ 𝐻 π‘œ, π‘ž0 is independent
  • β‡’ 𝐻 ∼ 𝐻 π‘œ, π‘ž for π‘ž ≔ 1 βˆ’ 1 βˆ’ π‘ž0 𝑛 ≀ π‘›π‘ž0
  • Property is monotone:
  • β„™ 𝐻 ∈ 𝒬 β‰₯ β„™ βˆͺ𝑗 𝐻𝑗 ∈ 𝒬

= 1 βˆ’ β„™ βˆ©π‘— 𝐻𝑗 βˆ‰ 𝒬

  • Graphs are independent:
  • β„™ βˆ©π‘— 𝐻𝑗 βˆ‰ 𝒬

= ς𝑗 β„™ 𝐻𝑗 βˆ‰ 𝒬

  • Since 𝐻𝑗 ∼ 𝐻(π‘œ, π‘ž0), β„™ 𝐻𝑗 βˆ‰ 𝒬 =

1 2

  • ∴ β„™ 𝐻 ∈ 𝒬 β‰₯ 1 βˆ’ 2βˆ’π‘› β†’ 1 if 𝑛 β†’ ∞ (or if π‘ž ≫ π‘ž0)

∎

Theorem 3.2.4 (BollobΓ‘s-Thomason, 1987) Every nontrivial monotone graph property has a threshold.

slide-17
SLIDE 17

Below the Threshold

Proof (lower bound)

  • Let 𝐻 ∼ 𝐻1 βˆͺ 𝐻2 βˆͺ β‹― βˆͺ 𝐻𝑛 as before, but with 𝐻𝑗 ∼ 𝐻 π‘œ, π‘ž for π‘ž =

π‘ž0 𝑛

  • β‡’ 𝐻 ∼ 𝐻(π‘œ, π‘Ÿ) for π‘Ÿ = 1 βˆ’ 1 βˆ’ π‘ž 𝑛 ≀ π‘›π‘ž = π‘ž0
  • β‡’ β„™ 𝐻 βˆ‰ 𝒬 β‰₯

1 2

  • As before, β„™ 𝐻 βˆ‰ 𝒬 ≀ β„™ 𝐻 π‘œ, π‘ž βˆ‰ 𝒬 𝑛
  • β‡’ β„™ 𝐻 π‘œ, π‘ž βˆ‰ 𝒬 β‰₯

1 2 1/m

  • β‡’ β„™ 𝐻 π‘œ, π‘ž ∈ 𝒬 ≀ 1 βˆ’

1 2 1/𝑛

β†’ 0 if 𝑛 β†’ ∞ (or if π‘ž β‰ͺ π‘ž0) ∎

Theorem 3.2.4 (BollobΓ‘s-Thomason, 1987) Every nontrivial monotone graph property has a threshold.

slide-18
SLIDE 18

Definition 3.2.5 (Sharp thresholds) We say π‘ž0(π‘œ) is a sharp threshold for 𝒬 if there are positive constants 𝑑1, 𝑑2 such that β„™ 𝐻 π‘œ, π‘ž ∈ 𝒬 β†’ α‰Š0 if π‘ž ≀ 𝑑1π‘ž0 π‘œ , 1 if π‘ž β‰₯ 𝑑2π‘ž0 π‘œ .

Closing Remarks

Random graph theory

  • Fundamental problem: given a graph property 𝒬, what is its threshold?

At the threshold

  • We showed what happens for probabilities much smaller than the threshold,

and much larger than the threshold

  • What if π‘ž = Θ π‘ž0 π‘œ

? Some properties have a much quicker transition

slide-19
SLIDE 19

Any questions?

slide-20
SLIDE 20

Β§3 Subgraphs of 𝐻 π‘œ, π‘ž

Chapter 3: The Second Moment The Probabilistic Method

slide-21
SLIDE 21

Returning to Ramsey

Choosing parameters

  • Want to choose π‘œ as large as possible
  • Need to avoid large independent sets
  • β‡’ would like to make edge probability π‘ž large
  • Limitation: need to avoid 𝐿ℓ

Question: What is the threshold for Kβ„“ βŠ† 𝐻(π‘œ, π‘ž)? Theorem 1.5.7 Given β„“, 𝑙, π‘œ ∈ β„• and π‘ž ∈ 0,1 , if π‘œ β„“ π‘ž

β„“ 2 + π‘œ

𝑙 1 βˆ’ π‘ž

𝑙 2 < 1,

then 𝑆 β„“, 𝑙 > π‘œ.

slide-22
SLIDE 22

A Lower Bound

Goal

  • Let π‘Œ count the number of 𝐿ℓ in 𝐻(π‘œ, π‘ž)
  • For which π‘ž do we have β„™ π‘Œ β‰₯ 1 = 𝑝(1)?

First moment

  • 𝔽 π‘Œ =

π‘œ β„“ π‘ž

β„“ 2 = Θ π‘œβ„“π‘ž β„“ 2

  • Markov’s Inequality: β„™ π‘Œ β‰₯ 1 ≀ 𝔽 π‘Œ

Threshold bound

  • 𝔽 π‘Œ = Θ π‘œβ„“π‘ž

β„“ 2

β‰ͺ 1

  • ⇔ π‘ž

β„“ 2 β‰ͺ π‘œβˆ’β„“ ⇔ π‘ž β‰ͺ π‘œβˆ’2/(β„“βˆ’1)

  • β‡’ π‘ž0(π‘œ) β‰₯ π‘œβˆ’2/(β„“βˆ’1)
slide-23
SLIDE 23

Corollary 3.1.4 Let 𝐹𝑗 be a sequence of events with probabilities π‘žπ‘—, and let π‘Œ count the number of events that occur. Write 𝑗 ∼ π‘˜ if the events 𝐹𝑗 and 𝐹

π‘˜ are

not independent, and let Ξ” = Οƒπ‘—βˆΌπ‘˜ β„™ 𝐹𝑗 ∧ 𝐹

π‘˜ . If 𝔽 π‘Œ β†’ ∞ and

Ξ” = 𝑝 𝔽 π‘Œ 2 , then 𝑄 π‘Œ = 0 = 𝑝 1 .

An Upper Bound

Goal

  • For which π‘ž do we have β„™ π‘Œ = 0 = 𝑝 1 ?

Our parameters

  • Let 𝐻 ∼ 𝐻(π‘œ, π‘ž) and, for 𝑇 ∈

π‘œ β„“ , let 𝐹𝑇 = 𝐻 𝑇 β‰… 𝐿ℓ

  • 𝔽 π‘Œ =

π‘œ β„“ π‘ž

β„“ 2 β†’ ∞ for π‘ž ≫ π‘œβˆ’2/(β„“βˆ’1)

  • Suffices to show Ξ” = 𝑝 𝔽 π‘Œ 2
slide-24
SLIDE 24

Clique Dependencies

Independent events

  • 𝐹𝑗 occurs ⇔ all edges in 𝑗th clique present
  • Edges appear independently
  • ∴ 𝑇 ∩ π‘ˆ ≀ 1 β‡’ 𝐹𝑇, πΉπ‘ˆ independent

Dependent events

  • Suppose 𝑇 ∩ π‘ˆ = 𝑑 β‰₯ 2
  • β‡’ 𝑇 ∼ π‘ˆ
  • 𝐹𝑇 ∧ πΉπ‘ˆ: 𝐻 𝑇 , 𝐻[π‘ˆ] both β„“-cliques, sharing 𝑑 vertices
  • Number of prescribed edges: 2 β„“

2 βˆ’ 𝑑 2

  • β‡’ β„™ 𝐹𝑇 ∧ πΉπ‘ˆ = π‘ž2 β„“

2 βˆ’ 𝑑 2

slide-25
SLIDE 25

Computing Ξ”

Recall

  • 𝑇 ∼ π‘ˆ ⇔ 𝑑 ≔ 𝑇 ∩ π‘ˆ β‰₯ 2
  • β„™ 𝐹𝑇 ∧ πΉπ‘ˆ = π‘ž2 β„“

2 βˆ’ s 2

Substituting terms

Ξ” = Οƒπ‘‡βˆΌπ‘ˆ β„™ 𝐹𝑇 ∧ πΉπ‘ˆ = Οƒ π‘‡βˆ©π‘ˆ β‰₯2 β„™ 𝐹𝑇 ∧ πΉπ‘ˆ = σ𝑇 Οƒπ‘ˆ: π‘‡βˆ©π‘ˆ β‰₯2 β„™ 𝐹𝑇 ∧ πΉπ‘ˆ = σ𝑇 σ𝑑=2

β„“βˆ’1 Οƒπ‘ˆ: π‘‡βˆ©π‘ˆ =𝑑 β„™ 𝐹𝑇 ∧ πΉπ‘ˆ

= σ𝑇 σ𝑑=2

β„“βˆ’1 Οƒπ‘ˆ: π‘‡βˆ©π‘ˆ =𝑑 π‘ž2 β„“

2 βˆ’ 𝑑 2

β‡’ Ξ” =

π‘œ β„“ σ𝑑=2 β„“βˆ’1 β„“ 𝑑 π‘œβˆ’β„“ β„“βˆ’π‘‘ π‘ž2 β„“

2 βˆ’ 𝑑 2

slide-26
SLIDE 26

Bounding Ξ”

Recall

  • Ξ” =

π‘œ β„“ σ𝑑=2 β„“βˆ’1 β„“ 𝑑 π‘œβˆ’β„“ β„“βˆ’π‘‘ π‘ž2 β„“

2 βˆ’ 𝑑 2

Goal

  • Show Ξ” = 𝑝 𝔽 π‘Œ 2

Estimates

  • β„“

𝑑 ≀ 2β„“

  • π‘œβˆ’β„“

β„“βˆ’π‘‘

≀ π‘œβ„“βˆ’π‘‘ = Θ

π‘œ β„“ π‘œβˆ’π‘‘

Bound

Ξ” ≀

π‘œ β„“ σ𝑑=2 β„“βˆ’1 2β„“Ξ˜ π‘œ β„“ π‘œβˆ’π‘‘ π‘ž2 β„“

2 βˆ’ 𝑑 2

=

π‘œ β„“ 2π‘ž2 β„“

2 σ𝑑=2

β„“βˆ’1 Θ π‘œβˆ’π‘‘π‘žβˆ’ 𝑑

2

= 𝔽 π‘Œ 2 σ𝑑=2

β„“βˆ’1 Θ π‘œβˆ’π‘‘π‘žβˆ’ 𝑑

2

slide-27
SLIDE 27

Theorem 3.3.1 For β„“ β‰₯ 2, the threshold for Kβ„“ βŠ† 𝐻(π‘œ, π‘ž) is π‘ž0 π‘œ = π‘œβˆ’2/(β„“βˆ’1).

Completing the Calculation

Recall

  • Ξ” = 𝔽 π‘Œ 2 σ𝑑=2

β„“βˆ’1 Θ π‘œβˆ’π‘‘π‘žβˆ’ 𝑑

2

Substituting π‘ž

  • π‘œβˆ’π‘‘π‘žβˆ’ 𝑑

2 = π‘œπ‘ž π‘‘βˆ’1 /2 βˆ’π‘‘

  • We took π‘ž ≫ π‘œβˆ’2/(β„“βˆ’1)
  • β‡’ π‘œβˆ’π‘‘π‘žβˆ’ 𝑑

2 β‰ͺ n1βˆ’(sβˆ’1)/(β„“βˆ’1) βˆ’s

  • For 2 ≀ 𝑑 ≀ β„“ βˆ’ 1, this is 𝑝(1)
  • β‡’ Ξ” = 𝑝 1
slide-28
SLIDE 28

An Incomplete Result

Lower bound

  • Let π‘Œ be the number of copies of 𝐼 in 𝐻 π‘œ, π‘ž
  • Markov: 𝔽 π‘Œ = 𝑝 1 β‡’ π‘ž0 π‘œ ≫ π‘ž

Expectation

  • Number of possible copies
  • Specify vertices of 𝐼 – at most π‘œπ‘€ 𝐼 possibilities
  • Probability of appearance
  • Each edge of 𝐼 must be present – probability is π‘žπ‘“ 𝐼
  • β‡’ 𝔽 π‘Œ ≀ π‘œπ‘€ 𝐼 π‘žπ‘“ 𝐼

Conclusion: π‘ž0 π‘œ β‰₯ π‘œβˆ’π‘€ 𝐼 /𝑓(𝐼) Problem Given a graph 𝐼, what is the threshold π‘ž0

𝐼(π‘œ) for 𝐼 βŠ† 𝐻 π‘œ, π‘ž ?

slide-29
SLIDE 29

An Illustrated Example

Graph statistics

  • Let 𝐼 be 𝐿4 with a pendant edge
  • Statistics:
  • 𝑀 𝐼 = 5
  • 𝑓 𝐼 = 7
  • β‡’ π‘ž0

𝐼 π‘œ β‰₯ π‘œβˆ’5/7

An issue

  • 𝐿4 βŠ† 𝐼
  • β‡’ if 𝐼 βŠ† 𝐻(π‘œ, π‘ž), then 𝐿4 βŠ† 𝐻 π‘œ, π‘ž
  • β‡’ π‘ž0

𝐿4 π‘œ ≀ π‘ž0 𝐼 π‘œ

  • But we showed π‘ž0

𝐿4 π‘œ = π‘œβˆ’2/3 ≫ π‘œβˆ’5/7

slide-30
SLIDE 30

Definition 3.3.2 (Maximum density) Given a graph 𝐼, define d 𝐼 =

𝑓 𝐼 𝑀 𝐼 , and let

𝑛 𝐼 = max {β…† 𝐺 : 𝐺 βŠ† 𝐼}.

Monotonicity and Density

General lower bound

  • π‘ž0

𝐼 π‘œ β‰₯ max π‘ž0 𝐺 π‘œ : 𝐺 βŠ† 𝐼

  • Can substitute first moment bound
  • β‡’ π‘ž0

𝐼 π‘œ β‰₯ max π‘œβˆ’π‘€ 𝐺 /𝑓(𝐺): 𝐺 βŠ† 𝐼, 𝑓 𝐺 β‰₯ 1

Remarks

  • We have π‘ž0

𝐼 π‘œ β‰₯ π‘œβˆ’1/𝑛(𝐼)

  • Say 𝐼 is balanced if d 𝐼 = 𝑛 𝐼
  • 𝐼 is strictly balanced if β…† 𝐺 < 𝑛(𝐼) for all 𝐺 βŠ‚ 𝐼
slide-31
SLIDE 31

Expected Subgraph Counts

Boundless expectations

  • Let π‘ŒπΌ be the number of copies of 𝐼 in 𝐻(π‘œ, π‘ž)
  • Total # possible copies = Θ π‘œπ‘€ 𝐼
  • Probability of each copy: π‘žπ‘“ 𝐼
  • β‡’ 𝔽 π‘ŒπΌ = Θ π‘œπ‘€ 𝐼 π‘žπ‘“ 𝐼
  • ∴ 𝔽 π‘ŒπΌ β†’ ∞ when π‘ž ≫ π‘œβˆ’π‘€ 𝐼 /𝑓 𝐼

Guaranteeing subgraph existence

  • Goal: to show β„™ π‘ŒπΌ = 0 = 𝑝(1) for π‘ž ≫ π‘ž0

𝐼(π‘œ)

  • Apply second moment: need to show Ξ” = o 𝔽 π‘ŒπΌ 2
  • Edge-disjoint copies are independent
slide-32
SLIDE 32

Dependent Subgraphs

Common subgraphs

  • Let 𝐼1, 𝐼2 be two copies of 𝐼 sharing an edge
  • 𝐹𝐼1 ∧ 𝐹𝐼2 = 𝐼1 βˆͺ 𝐼2 βŠ† 𝐻 π‘œ, π‘ž
  • Let 𝐺 ≔ 𝐼1 ∩ 𝐼2 be the common subgraph
  • 𝑀 𝐼1 βˆͺ 𝐼2 = 2𝑀 𝐼 βˆ’ 𝑀 𝐺
  • 𝑓 𝐼1 βˆͺ 𝐼2 = 2𝑓 𝐼 βˆ’ 𝑓 𝐺

Counting pairs

  • Group dependent pairs 𝐼1, 𝐼2 by common subgraphs 𝐺 = 𝐼1 ∩ 𝐼2
  • At most 2𝑓 𝐼 possible subgraphs 𝐺
  • For each 𝐾, O π‘œ2𝑀 𝐼 βˆ’π‘€ 𝐺

pairs (𝐼1, H2)

  • For each such pair, β„™ 𝐹𝐼1 ∧ 𝐹𝐼2 = π‘ž2𝑓 𝐼 βˆ’π‘“ 𝐺
slide-33
SLIDE 33

Bounding Ξ”

Recall

Ξ” = Οƒπ‘—βˆΌπ‘˜ β„™ 𝐹𝐼𝑗 ∧ πΉπΌπ‘˜

Group by common subgraph

Ξ” = Οƒπ‘—βˆΌπ‘˜ β„™ 𝐹𝐼𝑗 ∧ πΉπΌπ‘˜ = ΟƒπΊβŠ‚πΌ Οƒ 𝑗,π‘˜ :πΌπ‘—βˆ©πΌπ‘˜=𝐺 β„™ 𝐹𝐼𝑗 ∧ πΉπΌπ‘˜

Substitute estimates

Ξ” = ΟƒπΊβŠ‚πΌ 𝑃 π‘œ2𝑀 𝐼 βˆ’π‘€ 𝐺 π‘ž2𝑓 𝐼 βˆ’π‘“ 𝐺 β‡’ Ξ” = π‘œπ‘€ 𝐼 π‘žπ‘“ 𝐼

2 ΟƒπΊβŠ‚πΌ 𝑃 π‘œβˆ’π‘€ 𝐺 π‘žβˆ’π‘“ 𝐺

β‡’ Ξ” = 𝔽 π‘ŒπΌ 2 ΟƒπΊβŠ‚πΌ 𝑃 π‘œβˆ’π‘€ 𝐺 π‘žβˆ’π‘“ 𝐺

slide-34
SLIDE 34

Theorem 3.3.3 Given a graph 𝐼, the threshold for 𝐼 βŠ† 𝐻(π‘œ, π‘ž) is π‘ž0

𝐼 π‘œ = π‘œβˆ’1/𝑛(𝐼),

where 𝑛 𝐼 = max 𝑓 𝐺 𝑀 𝐺 : 𝐺 βŠ† 𝐼 .

A Complete Solution

Recall

  • Ξ” = 𝔽 π‘ŒπΌ 2 ΟƒπΊβŠ‚πΌ 𝑃 π‘œβˆ’π‘€ 𝐺 π‘žβˆ’π‘“ 𝐺

Choice of π‘ž

  • We have π‘ž ≫ π‘œβˆ’1/𝑛(𝐼)
  • β‡’ π‘ž ≫ π‘œβˆ’π‘€ 𝐺 /𝑓(𝐺) for all nonempty 𝐺 βŠ‚ 𝐼
  • β‡’ π‘œβˆ’π‘€ 𝐺 π‘žβˆ’π‘“ 𝐺 = 𝑝(1)
  • β‡’ Ξ” = 𝑝(1)
slide-35
SLIDE 35

Any questions?

slide-36
SLIDE 36

Β§4 Prime Factors

Chapter 3: The Second Moment The Probabilistic Method

slide-37
SLIDE 37

Theorem 3.4.1 (Hadamard, De la VallΓ©e Poussin, 1896) The number 𝜌(π‘œ) of prime numbers in [π‘œ] satisfies 𝜌 π‘œ = 1 + 𝑝 1 π‘œ ln π‘œ .

Time For Primes

Fun facts

  • There are infinitely many primes (Euclid, -300)
  • The primes contain arbitrarily long arithmetic progressions (Green–Tao, 2004)
  • Infinitely many pairs of primes are at most 70000000 apart (Zhang, 2014)

Central problem

  • How are the primes distributed in β„•?
slide-38
SLIDE 38

Definition 3.4.2 Given 𝑦 ∈ β„•, let πœ‰ 𝑦 denote the number of distinct prime factors of 𝑦.

Prime Factorisation

The funnest of facts

  • Every natural number is the product of primes

Our goal

  • To understand what these factorisations look like

Examples

  • πœ‰ 19 = ?
  • πœ‰ 210 = ?
  • πœ‰ 256 = ?
  • πœ‰ 2020 = ?
slide-39
SLIDE 39

The Average Case

Proof

  • Express πœ‰(𝑦) in terms of indicator random variables:
  • πœ‰ 𝑦 = Οƒπ‘žβ‰€π‘œ 1{π‘ž|𝑦}
  • Exchange order of summation
  • 1

π‘œ Οƒπ‘¦βˆˆ π‘œ πœ‰ 𝑦 = 1 π‘œ Οƒπ‘žβ‰€π‘œ Οƒπ‘¦βˆˆ π‘œ 1{π‘ž|𝑦}

  • Count multiples
  • Οƒπ‘¦βˆˆ π‘œ 1{π‘ž|𝑦} =

π‘œ π‘ž = n p + O 1

  • β‡’

1 π‘œ Οƒπ‘¦βˆˆ π‘œ πœ‰ 𝑦 = Οƒπ‘žβ‰€π‘œ 1 π‘ž + 𝑃 1 = ln ln π‘œ + 𝑃(1)

∎

Proposition 3.4.3 The average number of distinct prime factors of a number 𝑦 ∈ [π‘œ] is ln ln π‘œ + 𝑃(1).

slide-40
SLIDE 40

A Harmonic Digression

β€œProof”

  • Let 𝑛 = 𝜌 π‘œ ∼

π‘œ ln π‘œ

  • Οƒπ‘žβ‰€π‘œ

1 π‘ž = σ𝑙=1 𝑛 1 π‘žπ‘™

  • Prime Number Theorem β‡’ π‘žπ‘™ ∼ 𝑙 ln 𝑙
  • β‡’ Οƒπ‘žβ‰€π‘œ

1 π‘ž ∼ σ𝑙=2 𝑛 1 𝑙 ln 𝑙

  • Approximate by an integral:
  • σ𝑙=2

𝑛 1 𝑙 ln 𝑙 ∼ Χ¬ 𝑦=2 𝑛 1 𝑦 ln 𝑦 ⅆ𝑦 ∼ ln ln 𝑛 ∼ ln ln π‘œ

∎

Theorem 3.4.4 (Mertens, 1874) As π‘œ β†’ ∞, we have Οƒπ‘žβ‰€π‘œ

1 π‘ž = ln ln π‘œ + 𝑃 1 .

slide-41
SLIDE 41

Theorem 3.4.5 (Hardy-Ramanujan, 1920) As π‘œ β†’ ∞, we have πœ‰ 𝑦 = 1 + 𝑝 1 ln ln π‘œ for all but 𝑝 π‘œ integers 𝑦 ∈ π‘œ .

The Typical Case

Variation in πœ‰(𝑦), 𝑦 ∈ π‘œ

  • Minimum:

1

  • Average:

ln ln π‘œ + 𝑃 1

  • Maximum:

1 + 𝑝 1

ln π‘œ ln ln π‘œ

  • Product of first 𝑛 primes ∼ ς𝑙=1

𝑛

𝑙 ln 𝑙 ∼ 𝑛! ln 𝑛 𝑛 ≀ π‘œ for 𝑛 ∼

ln π‘œ ln ln π‘œ

What can we say about the distribution of πœ‰ 𝑦 ?

slide-42
SLIDE 42

Theorem 3.4.5 (Hardy-Ramanujan, 1920) As π‘œ β†’ ∞, we have πœ‰ 𝑦 = 1 + 𝑝 1 ln ln π‘œ for all but 𝑝 π‘œ integers 𝑦 ∈ π‘œ .

The Probabilistic Approach

Probabilistic proof (TurΓ‘n, 1934)

  • Choose 𝑦 ∈ π‘œ uniformly at random
  • Interested in the random variable π‘Œ = πœ‰ 𝑦
  • Proposition 3.4.3 β‡’ 𝔽 π‘Œ = ln ln π‘œ + 𝑃 1

Corollary 3.1.3’ If Var π‘Œ = 𝑝 𝔽 π‘Œ 2 , then π‘Œ = 1 + 𝑝 1 𝔽 π‘Œ with high probability.

slide-43
SLIDE 43

Expressing the Variance

Recall

  • 𝑦 ∈ π‘œ uniformly random
  • π‘Œ = πœ‰(𝑦) number of distinct prime factors
  • Goal: show Var π‘Œ = 𝑝 𝔽 π‘Œ 2

Indicator random variables

  • For a prime π‘ž, let π‘Œπ‘ž = 1{π‘ž|𝑦}, Bernoulli random variable
  • β„™ π‘Œπ‘ž = 1 =

π‘œ/π‘ž n

∈

1 π‘ž βˆ’ 1 π‘œ , 1 π‘ž

  • π‘Œ = Οƒπ‘žβ‰€π‘œ π‘Œπ‘ž

Our friend the variance

  • Var π‘Œ = Οƒπ‘ž Var π‘Œπ‘ž + Οƒ π‘ž,π‘Ÿ :π‘žβ‰ π‘Ÿ Cov π‘Œπ‘ž, π‘Œπ‘Ÿ
  • Οƒπ‘ž Var π‘Œπ‘ž ≀ Οƒπ‘ž 𝔽 π‘Œπ‘ž = 𝔽 π‘Œ
slide-44
SLIDE 44

Computing Covariances

Pairs π‘ž β‰  π‘Ÿ

  • Cov π‘Œπ‘ž, π‘Œπ‘Ÿ = 𝔽 π‘Œπ‘žπ‘Œπ‘Ÿ βˆ’ 𝔽 π‘Œπ‘ž 𝔽 π‘Œπ‘Ÿ
  • 𝔽 π‘Œπ‘ž β‰₯

1 π‘ž βˆ’ 1 π‘œ, 𝔽 π‘Œπ‘Ÿ β‰₯ 1 π‘Ÿ βˆ’ 1 n

  • 𝔽 π‘Œπ‘žπ‘Œπ‘Ÿ = β„™ π‘žπ‘Ÿ 𝑦 ≀

1 π‘žπ‘Ÿ

  • β‡’ Cov π‘Œπ‘ž, π‘Œπ‘Ÿ ≀

1 π‘žπ‘Ÿ βˆ’ 1 π‘ž βˆ’ 1 π‘œ 1 π‘Ÿ βˆ’ 1 π‘œ ≀ 1 π‘œ 1 π‘ž + 1 π‘Ÿ

Bounding the sum

  • β‡’ Οƒ π‘ž,π‘Ÿ :π‘žβ‰ π‘Ÿ Cov π‘Œπ‘ž, π‘Œπ‘Ÿ ≀

1 π‘œ Οƒ π‘ž,π‘Ÿ :π‘žβ‰ π‘Ÿ 1 π‘ž + 1 π‘Ÿ ≀ 2𝜌 π‘œ π‘œ

Οƒπ‘žβ‰€π‘œ

1 π‘ž

  • 𝜌 π‘œ = 1 + 𝑝 1

π‘œ ln π‘œ and Οƒπ‘žβ‰€π‘œ 1 π‘ž = ln ln π‘œ + 𝑃 1 = 𝔽 π‘Œ

  • β‡’ Οƒ π‘ž,π‘Ÿ :π‘žβ‰ π‘Ÿ Cov π‘Œπ‘ž, π‘Œπ‘Ÿ = 𝑝 𝔽 π‘Œ
slide-45
SLIDE 45

A Final Flourish

The variance

  • Var π‘Œ = Οƒπ‘ž Var π‘Œπ‘ž + Οƒπ‘žβ‰ π‘Ÿ Cov π‘Œπ‘ž, π‘Œπ‘Ÿ
  • Οƒπ‘ž Var π‘Œπ‘ž ≀ 𝔽 π‘Œ

and Οƒπ‘žβ‰ π‘Ÿ Cov π‘Œπ‘ž, π‘Œπ‘Ÿ = 𝑝 𝔽 π‘Œ

  • β‡’ Var π‘Œ = 1 + 𝑝 1

𝔽 π‘Œ = 1 + 𝑝 1 ln ln π‘œ

Applying Chebyshev

  • β„™ πœ‰ 𝑦 βˆ’ ln ln π‘œ > πœ‡ ln ln π‘œ ≀ Var π‘Œ

πœ‡2 ln ln π‘œ = 1 πœ‡2 + 𝑝(1)

  • β‡’ β„™ πœ‰ 𝑦 β‰  1 + 𝑝 1

ln ln π‘œ = 𝑝 1

  • 𝑦 uniform in [π‘œ] β‡’ 𝑝(π‘œ) such integers

∎

Remark

  • Most 𝑦 ∈ π‘œ satisfy πœ‰ 𝑦 = ln ln π‘œ + 𝑃

ln ln π‘œ

slide-46
SLIDE 46

Any questions?

slide-47
SLIDE 47

Β§5 Distinct Sums

Chapter 3: The Second Moment The Probabilistic Method

slide-48
SLIDE 48

Mathemagic

An illusion

  • You have a deck of cards, with each card bearing a number
  • You invite your friend to select as many cards from the deck as they like
  • They add the numbers and only tell you the sum
  • The chosen cards are then shuffled back into the deck
  • You then go through the deck, and magically pick out your friend’s cards

The secret

  • Cards labelled with powers of two: 1,2,4,8,16, …
  • Each number 𝑦 ∈ β„• has a unique binary expansion, 𝑦 = Οƒπ‘˜ 2π‘—π‘˜
  • β‡’ given the sum 𝑦, can recover the labels 2π‘—π‘˜ of the chosen cards
slide-49
SLIDE 49

A Little Showmanship

Obstacles

  • Mathematician friends will see through the illusion
  • Non-mathematician friends may not be able to add well
  • Card labels shouldn’t be larger than π‘œ
  • Binary labels β‡’ log π‘œ cards
  • Small deck is not so impressive

Better decks

  • Can we replace the binary labels?
  • Suppose we have labels 𝑇 = 𝑑1, 𝑑2, … , 𝑑𝑙
  • Key property:
  • distinct sums – no two subsets should have the same total
  • Extremal problem
  • How large can a subset 𝑇 βŠ† [π‘œ] with distinct sums be?
slide-50
SLIDE 50

Claim 3.5.1 The greedy algorithm returns the set of powers of two.

The Greedy Magician

Greedy algorithm

  • Start with 𝑇 = βˆ…
  • Go through elements in π‘œ one at a time
  • Add to 𝑇 if they preserve distinct sums property

Proof

  • After the first step, we have 𝑇 = 1
  • Suppose we have 𝑇 = 1,2, … , 2𝑠 at some stage in the algorithm
  • We can write every number up to 2𝑠+1 βˆ’ 1 as a sum of these elements
  • None of these added to 𝑇
  • Next available number to be added: 2𝑠+1

∎

slide-51
SLIDE 51

The Extremal Function

Notation

  • Let 𝑔 π‘œ = max

𝑇 : 𝑇 βŠ† π‘œ has distinct sums

Lower bound

  • Binary set β‡’ 𝑔 π‘œ β‰₯ log π‘œ + 1
  • Is this best possible?

Counterexamples

  • 𝑇 = {11,17,20,22,23,24} has distinct sums
  • β‡’ 𝑔 π‘œ β‰₯ log π‘œ + 2 for 24 ≀ π‘œ ≀ 31
  • If a set 𝑇 has distinct sums, so does 𝑇′ = 2𝑇 βˆͺ {1}
  • Iterating β†’ infinite sequence of counterexamples
slide-52
SLIDE 52

An Upper Bound

Proof

  • Let 𝑙 = 𝑔(π‘œ) and let 𝑇 βŠ† [π‘œ] be a largest set with distinct sums
  • For each π‘ˆ βŠ† 𝑇, we have 0 ≀ Οƒπ‘‘βˆˆπ‘ˆ 𝑑 < π‘™π‘œ
  • Distinct sums β‡’ each of these 2𝑙 sums is distinct
  • β‡’ 2k ≀ π‘™π‘œ

β‡’ 𝑙 ≀ log π‘œ + log 𝑙 β‡’ 𝑙 ≀ log π‘œ + log log π‘œ + log 𝑙 ≀ log π‘œ + log 2 log π‘œ = log π‘œ + log log π‘œ + 1 ∎

Proposition 3.5.2 As π‘œ β†’ ∞, we have 𝑔 π‘œ ≀ log π‘œ + log log π‘œ + 1.

slide-53
SLIDE 53

Theorem 3.5.3 As π‘œ β†’ ∞, 𝑔 π‘œ ≀ log π‘œ +

1 2 log log π‘œ + 𝑃(1).

An Improved Upper Bound

Flawed argument

  • Wasteful in estimating range of sums
  • Max sum ∼ π‘™π‘œ β‡’ all members of 𝑇 ∼ π‘œ
  • In that case, few small numbers will be sums

Fix

  • Try to find a smaller interval still containing many sums
  • Chebyshev β‡’ sums may concentrate around the average
slide-54
SLIDE 54

Probabilistic Framework

Random variables

  • Let 𝑔 π‘œ = 𝑙, let 𝑇 = 𝑑1, 𝑑2, … , 𝑑𝑙 βŠ† [π‘œ] be a largest set with distinct sums
  • Let π‘Œ be a uniformly random sum from 𝑇
  • β‡’ π‘Œ = σ𝑗=1

𝑙

πœπ‘—π‘‘π‘—, where each πœπ‘— is independent, uniform on 0,1

Expectation

  • Let 𝜈 ≔ 𝔽 π‘Œ = σ𝑗=1

𝑙

𝔽 πœπ‘—π‘‘π‘— =

1 2 σ𝑗=1 𝑙

𝑑𝑗

  • Actual value is unimportant

Variance

  • Variables πœπ‘— are independent
  • β‡’ Var π‘Œ = Var σ𝑗=1

𝑙

πœπ‘—π‘‘π‘— = σ𝑗=1

𝑙

Var πœπ‘— 𝑑𝑗

2 = 1 4 σ𝑗=1 𝑙

𝑑𝑗

2 ≀ 1 4 π‘œ2𝑙

slide-55
SLIDE 55

Concentrated Sums

Recall

  • Var π‘Œ ≀

1 4 π‘œ2𝑙

Applying Chebyshev

  • β„™ π‘Œ βˆ’ 𝜈 β‰₯ π‘œ 𝑙 ≀ Var π‘Œ

π‘œ2𝑙

≀

1 4

  • β‡’ β„™ π‘Œ βˆ’ 𝜈 < π‘œ 𝑙 β‰₯

3 4

Distinct sums

  • Each value comes from at most one sum β‡’ β„™ π‘Œ = 𝑦 ∈ 0, 2βˆ’π‘™
  • ∴ β„™ π‘Œ βˆ’ 𝜈 < π‘œ 𝑙 = β„™ 𝜈 βˆ’ π‘œ 𝑙 < π‘Œ < 𝜈 + π‘œ 𝑙 ≀ 2π‘œ 𝑙 β‹… 2βˆ’π‘™

Bounding 𝑙

  • 2𝑙 ≀

8 3 π‘œ 𝑙 β‡’ 𝑙 ≀ log π‘œ + 1 2 log 𝑙 + log 8 3 ≀ log π‘œ + 1 2 log log π‘œ + 𝑃(1)

∎

slide-56
SLIDE 56

Any questions?