A Mathematical Theory of Communication (after C. E. Shannon) Alex - - PowerPoint PPT Presentation

a mathematical theory of communication
SMART_READER_LITE
LIVE PREVIEW

A Mathematical Theory of Communication (after C. E. Shannon) Alex - - PowerPoint PPT Presentation

A Mathematical Theory of Communication (after C. E. Shannon) Alex Vlasiuk Alex A Mathematical Theory of Communication 1 / 17 Image: IEEE Information Theory Society Why Shannon? Alex A Mathematical Theory of Communication 3 / 17 Why


slide-1
SLIDE 1

A Mathematical Theory of Communication

(after C. E. Shannon) Alex Vlasiuk

Alex A Mathematical Theory of Communication 1 / 17

slide-2
SLIDE 2

Image: IEEE Information Theory Society

slide-3
SLIDE 3

Why Shannon?

Alex A Mathematical Theory of Communication 3 / 17

slide-4
SLIDE 4

Why Shannon?

◮ ”the father of information theory”

Alex A Mathematical Theory of Communication 3 / 17

slide-5
SLIDE 5

Why Shannon?

◮ ”the father of information theory” ◮ ideas from the 1948 paper are ubiquitous

Alex A Mathematical Theory of Communication 3 / 17

slide-6
SLIDE 6

Why Shannon?

◮ ”the father of information theory” ◮ ideas from the 1948 paper are ubiquitous ◮ (hopefully) some can be explained through handwaving

Alex A Mathematical Theory of Communication 3 / 17

slide-7
SLIDE 7

Why Shannon?

◮ ”the father of information theory” ◮ ideas from the 1948 paper are ubiquitous ◮ (hopefully) some can be explained through handwaving:

c Jeff Portaro, Noun Project Alex A Mathematical Theory of Communication 3 / 17

slide-8
SLIDE 8

Why Shannon?

◮ ”the father of information theory” ◮ ideas from the 1948 paper are ubiquitous ◮ (hopefully) some can be explained through handwaving:

c Jeff Portaro, Noun Project

◮ was on my desktop

Alex A Mathematical Theory of Communication 3 / 17

slide-9
SLIDE 9

Why Shannon?

◮ ”the father of information theory” ◮ ideas from the 1948 paper are ubiquitous ◮ (hopefully) some can be explained through handwaving:

c Jeff Portaro, Noun Project

◮ was on my desktop

Shannon, Claude Elwood. ”A mathematical theory of communication.” ACM SIGMOBILE Mobile Computing and Communications Review 5.1 (2001): 3-55.

Alex A Mathematical Theory of Communication 3 / 17

slide-10
SLIDE 10

Setting

Alex A Mathematical Theory of Communication 4 / 17

slide-11
SLIDE 11

Capacity and states of a channel

Symbols: S1, . . . , Sn with certain durations t1, . . . , tn.

Alex A Mathematical Theory of Communication 5 / 17

slide-12
SLIDE 12

Capacity and states of a channel

Symbols: S1, . . . , Sn with certain durations t1, . . . , tn. Allowed combinations of symbols are signals.

Alex A Mathematical Theory of Communication 5 / 17

slide-13
SLIDE 13

Capacity and states of a channel

Symbols: S1, . . . , Sn with certain durations t1, . . . , tn. Allowed combinations of symbols are signals. Capacity of a channel: C = lim

T→∞

log N(T) T , N(T) is the number of allowed signals of duration T.

Alex A Mathematical Theory of Communication 5 / 17

slide-14
SLIDE 14

Capacity and states of a channel

Symbols: S1, . . . , Sn with certain durations t1, . . . , tn. Allowed combinations of symbols are signals. Capacity of a channel: C = lim

T→∞

log N(T) T , N(T) is the number of allowed signals of duration T. Units: bits per second.

Alex A Mathematical Theory of Communication 5 / 17

slide-15
SLIDE 15

Capacity and states of a channel

Symbols: S1, . . . , Sn with certain durations t1, . . . , tn. Allowed combinations of symbols are signals. Capacity of a channel: C = lim

T→∞

log N(T) T , N(T) is the number of allowed signals of duration T. Units: bits per second.

Alex A Mathematical Theory of Communication 5 / 17

slide-16
SLIDE 16

Graphical representation of a Markov process

Source is a stochastic (random) process.

Alex A Mathematical Theory of Communication 6 / 17

slide-17
SLIDE 17

Graphical representation of a Markov process

Source is a stochastic (random) process.

  • Example. Alphabet: A, B, C.

Alex A Mathematical Theory of Communication 6 / 17

slide-18
SLIDE 18

Graphical representation of a Markov process

Source is a stochastic (random) process.

  • Example. Alphabet: A, B, C. Transition probabilities:

Alex A Mathematical Theory of Communication 6 / 17

slide-19
SLIDE 19

Graphical representation of a Markov process

Source is a stochastic (random) process.

  • Example. Alphabet: A, B, C. Transition probabilities:

Alex A Mathematical Theory of Communication 6 / 17

slide-20
SLIDE 20

Example: approximations to English

Using 27 (26+space) alphabet.

Alex A Mathematical Theory of Communication 7 / 17

slide-21
SLIDE 21

Example: approximations to English

Using 27 (26+space) alphabet. ◮ symbols independent and equiprobable: XFOML RXKHRJFFJUJ ZLPWCFWKCYJ FFJEYVKCQSGHYD QPAAMKBZAACIBZLHJQD

Alex A Mathematical Theory of Communication 7 / 17

slide-22
SLIDE 22

Example: approximations to English

Using 27 (26+space) alphabet. ◮ symbols independent and equiprobable: XFOML RXKHRJFFJUJ ZLPWCFWKCYJ FFJEYVKCQSGHYD QPAAMKBZAACIBZLHJQD ◮ symbols independent but with frequencies of English text: OCRO HLI RGWR NMIELWIS EU LL NBNESEBYA TH EEI ALHENHTTPA OOBTTVA NAH BRL

Alex A Mathematical Theory of Communication 7 / 17

slide-23
SLIDE 23

Example: approximations to English

Using 27 (26+space) alphabet. ◮ symbols independent and equiprobable: XFOML RXKHRJFFJUJ ZLPWCFWKCYJ FFJEYVKCQSGHYD QPAAMKBZAACIBZLHJQD ◮ symbols independent but with frequencies of English text: OCRO HLI RGWR NMIELWIS EU LL NBNESEBYA TH EEI ALHENHTTPA OOBTTVA NAH BRL ◮ digram structure as in English: ON IE ANTSOUTINYS ARE T INCTORE ST BE S DEAMY ACHIN D ILONASIVE TUCOOWE AT TEASONARE FUSO TIZIN ANDY TOBE SEACE CTISBE

Alex A Mathematical Theory of Communication 7 / 17

slide-24
SLIDE 24

Example: approximations to English

Using 27 (26+space) alphabet. ◮ symbols independent and equiprobable: XFOML RXKHRJFFJUJ ZLPWCFWKCYJ FFJEYVKCQSGHYD QPAAMKBZAACIBZLHJQD ◮ symbols independent but with frequencies of English text: OCRO HLI RGWR NMIELWIS EU LL NBNESEBYA TH EEI ALHENHTTPA OOBTTVA NAH BRL ◮ digram structure as in English: ON IE ANTSOUTINYS ARE T INCTORE ST BE S DEAMY ACHIN D ILONASIVE TUCOOWE AT TEASONARE FUSO TIZIN ANDY TOBE SEACE CTISBE

”One opens a book at random and selects a letter at random on the page. This letter is recorded. The book is then opened to another page and one reads until this letter is encountered. The succeeding letter is then recorded.”

Alex A Mathematical Theory of Communication 7 / 17

slide-25
SLIDE 25

◮ trigram structure as in English: IN NO IST LAT WHEY CRATICT FROURE BIRS GROCID PONDENOME OF DEMONSTURES OF THE REPTAGIN IS REGOACTIONA OF CRE

Alex A Mathematical Theory of Communication 8 / 17

slide-26
SLIDE 26

◮ trigram structure as in English: IN NO IST LAT WHEY CRATICT FROURE BIRS GROCID PONDENOME OF DEMONSTURES OF THE REPTAGIN IS REGOACTIONA OF CRE ◮ first-order word approximation: REPRESENTING AND SPEEDILY IS AN GOOD APT OR COME CAN DIFFERENT NATURAL HERE HE THE A IN CAME THE TO OF TO EXPERT GRAY COME TO FURNISHES THE LINE MESSAGE HAD BE THESE

Alex A Mathematical Theory of Communication 8 / 17

slide-27
SLIDE 27

◮ trigram structure as in English: IN NO IST LAT WHEY CRATICT FROURE BIRS GROCID PONDENOME OF DEMONSTURES OF THE REPTAGIN IS REGOACTIONA OF CRE ◮ first-order word approximation: REPRESENTING AND SPEEDILY IS AN GOOD APT OR COME CAN DIFFERENT NATURAL HERE HE THE A IN CAME THE TO OF TO EXPERT GRAY COME TO FURNISHES THE LINE MESSAGE HAD BE THESE ◮ Second-order word approximation: THE HEAD AND IN FRONTAL ATTACK ON AN ENGLISH WRITER THAT THE CHARACTER OF THIS POINT IS THEREFORE ANOTHER METHOD FOR THE LETTERS THAT THE TIME OF WHO EVER TOLD THE PROBLEM FOR AN UNEXPECTED

Alex A Mathematical Theory of Communication 8 / 17

slide-28
SLIDE 28

Entropy

A set of possible events with probabilities p1, p2, . . . , pn

Alex A Mathematical Theory of Communication 9 / 17

slide-29
SLIDE 29

Entropy

A set of possible events with probabilities p1, p2, . . . , pn Need: a measure of uncertainty in the outcome

Alex A Mathematical Theory of Communication 9 / 17

slide-30
SLIDE 30

Entropy

A set of possible events with probabilities p1, p2, . . . , pn Need: a measure of uncertainty in the outcome H = −

n

  • i=1

pi log pi

Alex A Mathematical Theory of Communication 9 / 17

slide-31
SLIDE 31

Entropy

A set of possible events with probabilities p1, p2, . . . , pn Need: a measure of uncertainty in the outcome H = −

n

  • i=1

pi log pi Example: two possibilities with probabilities p and q = 1 − p.

Alex A Mathematical Theory of Communication 9 / 17

slide-32
SLIDE 32

Entropy

A set of possible events with probabilities p1, p2, . . . , pn Need: a measure of uncertainty in the outcome H = −

n

  • i=1

pi log pi Example: two possibilities with probabilities p and q = 1 − p. H = −(p log p + q log q)

Alex A Mathematical Theory of Communication 9 / 17

slide-33
SLIDE 33

Alex A Mathematical Theory of Communication 10 / 17

slide-34
SLIDE 34

Conditional entropy and entropy of a source

x, y - events H(x) + H(y) ≥ H(x, y)

Alex A Mathematical Theory of Communication 11 / 17

slide-35
SLIDE 35

Conditional entropy and entropy of a source

x, y - events H(x) + H(y) ≥ H(x, y) = H(x) + Hx(y)

Alex A Mathematical Theory of Communication 11 / 17

slide-36
SLIDE 36

Conditional entropy and entropy of a source

x, y - events H(x) + H(y) ≥ H(x, y) = H(x) + Hx(y) A source has states with entropies Hi, transition probabilities are pi(j)

Alex A Mathematical Theory of Communication 11 / 17

slide-37
SLIDE 37

Conditional entropy and entropy of a source

x, y - events H(x) + H(y) ≥ H(x, y) = H(x) + Hx(y) A source has states with entropies Hi, transition probabilities are pi(j), then H =

  • i

PiHi = −

  • i,j

Pi pi(j) log pi(j)

Alex A Mathematical Theory of Communication 11 / 17

slide-38
SLIDE 38

Conditional entropy and entropy of a source

x, y - events H(x) + H(y) ≥ H(x, y) = H(x) + Hx(y) A source has states with entropies Hi, transition probabilities are pi(j), then H =

  • i

PiHi = −

  • i,j

Pi pi(j) log pi(j) Different units!

Alex A Mathematical Theory of Communication 11 / 17

slide-39
SLIDE 39

Conditional entropy and entropy of a source

x, y - events H(x) + H(y) ≥ H(x, y) = H(x) + Hx(y) A source has states with entropies Hi, transition probabilities are pi(j), then H =

  • i

PiHi = −

  • i,j

Pi pi(j) log pi(j) Different units! Turns out, lim

N→∞

log n(q) N = H n(q) – number of the most probable sequences of length N, total probability q with any q = 0, 1.

Alex A Mathematical Theory of Communication 11 / 17

slide-40
SLIDE 40

Conditional entropy and entropy of a source

x, y - events H(x) + H(y) ≥ H(x, y) = H(x) + Hx(y) A source has states with entropies Hi, transition probabilities are pi(j), then H =

  • i

PiHi = −

  • i,j

Pi pi(j) log pi(j) Different units! Turns out, lim

N→∞

log n(q) N = H n(q) – number of the most probable sequences of length N, total probability q with any q = 0, 1. Entropy of source: bits per symbol

Alex A Mathematical Theory of Communication 11 / 17

slide-41
SLIDE 41

Noiseless case

Alex A Mathematical Theory of Communication 12 / 17

slide-42
SLIDE 42

Noiseless case

Theorem (the fundamental theorem for a noiseless channel) Let a source have entropy H bits/symbol and a channel have a capacity C bits/second. Then it is possible to encode the output of the source to transmit at the average rate C

H − ǫ symbols/second over the channel

where ǫ is arbitrarily small. It is not possible to transmit at an average rate greater than C

H .

Alex A Mathematical Theory of Communication 12 / 17

slide-43
SLIDE 43

Noiseless case

Theorem (the fundamental theorem for a noiseless channel) Let a source have entropy H bits/symbol and a channel have a capacity C bits/second. Then it is possible to encode the output of the source to transmit at the average rate C

H − ǫ symbols/second over the channel

where ǫ is arbitrarily small. It is not possible to transmit at an average rate greater than C

H .

[C] [H] = bits/s bits/sym = sym/s

Alex A Mathematical Theory of Communication 12 / 17

slide-44
SLIDE 44

Noiseless case

Theorem (the fundamental theorem for a noiseless channel) Let a source have entropy H bits/symbol and a channel have a capacity C bits/second. Then it is possible to encode the output of the source to transmit at the average rate C

H − ǫ symbols/second over the channel

where ǫ is arbitrarily small. It is not possible to transmit at an average rate greater than C

H .

[C] [H] = bits/s bits/sym = sym/s The proof involves constructing an explicit code that achieves the required rate: Shannon-Fano coding.

Alex A Mathematical Theory of Communication 12 / 17

slide-45
SLIDE 45

Noisy case

Source output: x, decoded output: y. Noise: stochastic process as well.

Alex A Mathematical Theory of Communication 13 / 17

slide-46
SLIDE 46

Equivocation

Hy(x) – equivocation.

Alex A Mathematical Theory of Communication 14 / 17

slide-47
SLIDE 47

Equivocation

Hy(x) – equivocation. Actual transmission rate: R = H(x) − Hy(x).

Alex A Mathematical Theory of Communication 14 / 17

slide-48
SLIDE 48

Equivocation

Hy(x) – equivocation. Actual transmission rate: R = H(x) − Hy(x). Capacity of a noisy channel (maximum over all sources): C = max(H(x) − Hy(x)). Transmitting: 1000 bits/second with probabilities p0 = p1 = 1

  • 2. On

average, 1 in 100 is received incorrectly.

Alex A Mathematical Theory of Communication 14 / 17

slide-49
SLIDE 49

Equivocation

Hy(x) – equivocation. Actual transmission rate: R = H(x) − Hy(x). Capacity of a noisy channel (maximum over all sources): C = max(H(x) − Hy(x)). Transmitting: 1000 bits/second with probabilities p0 = p1 = 1

  • 2. On

average, 1 in 100 is received incorrectly. Saying that rate is 990 (= 1000 × 0.99) is not reasonable: don’t know where the errors occur.

Alex A Mathematical Theory of Communication 14 / 17

slide-50
SLIDE 50

Equivocation

Hy(x) – equivocation. Actual transmission rate: R = H(x) − Hy(x). Capacity of a noisy channel (maximum over all sources): C = max(H(x) − Hy(x)). Transmitting: 1000 bits/second with probabilities p0 = p1 = 1

  • 2. On

average, 1 in 100 is received incorrectly. Saying that rate is 990 (= 1000 × 0.99) is not reasonable: don’t know where the errors occur. With the above definition of Hy(x)

Alex A Mathematical Theory of Communication 14 / 17

slide-51
SLIDE 51

Equivocation

Hy(x) – equivocation. Actual transmission rate: R = H(x) − Hy(x). Capacity of a noisy channel (maximum over all sources): C = max(H(x) − Hy(x)). Transmitting: 1000 bits/second with probabilities p0 = p1 = 1

  • 2. On

average, 1 in 100 is received incorrectly. Saying that rate is 990 (= 1000 × 0.99) is not reasonable: don’t know where the errors occur. With the above definition of Hy(x) (if y = 1 is received, probability that x = 1 was sent is 0.99, etc): Hy(x) = −(0.99 log 0.99 + 0.01 log 0.01) = 0.081 bits/symbol.

Alex A Mathematical Theory of Communication 14 / 17

slide-52
SLIDE 52

Equivocation

Hy(x) – equivocation. Actual transmission rate: R = H(x) − Hy(x). Capacity of a noisy channel (maximum over all sources): C = max(H(x) − Hy(x)). Transmitting: 1000 bits/second with probabilities p0 = p1 = 1

  • 2. On

average, 1 in 100 is received incorrectly. Saying that rate is 990 (= 1000 × 0.99) is not reasonable: don’t know where the errors occur. With the above definition of Hy(x) (if y = 1 is received, probability that x = 1 was sent is 0.99, etc): Hy(x) = −(0.99 log 0.99 + 0.01 log 0.01) = 0.081 bits/symbol. Thus the actual transmission rate is R = 1000 − 81 = 919 bits/second

Alex A Mathematical Theory of Communication 14 / 17

slide-53
SLIDE 53

Theorem (the fundamental theorem for a discrete channel with noise) Let a discrete channel have the capacity C and a discrete source the entropy per second H. If H ≤ C, there exists a coding system with an arbitrarily small frequency of errors (or an arbitrarily small equivocation Hy(x)) during transmission. If H > C it is possible to encode the source so that the equivocation is less than H − C + ǫ where ǫ is arbitrarily

  • small. There is no method of encoding which gives an equivocation less

than H − C.

Alex A Mathematical Theory of Communication 15 / 17

slide-54
SLIDE 54

Theorem (the fundamental theorem for a discrete channel with noise) Let a discrete channel have the capacity C and a discrete source the entropy per second H. If H ≤ C, there exists a coding system with an arbitrarily small frequency of errors (or an arbitrarily small equivocation Hy(x)) during transmission. If H > C it is possible to encode the source so that the equivocation is less than H − C + ǫ where ǫ is arbitrarily

  • small. There is no method of encoding which gives an equivocation less

than H − C.

Alex A Mathematical Theory of Communication 15 / 17

slide-55
SLIDE 55

Shannon-Fano coding

Image: Wikimedia Alex A Mathematical Theory of Communication 16 / 17

slide-56
SLIDE 56

Thanks!