entropy randomness and information
play

Entropy, Randomness, and Information Lecture 27 December 5, 2013 - PowerPoint PPT Presentation

CS 573: Algorithms, Fall 2013 Entropy, Randomness, and Information Lecture 27 December 5, 2013 Sariel (UIUC) CS573 1 Fall 2013 1 / 28 Part I . Entropy . Sariel (UIUC) CS573 2 Fall 2013 2 / 28 Quote If only once - only once -


  1. CS 573: Algorithms, Fall 2013 Entropy, Randomness, and Information Lecture 27 December 5, 2013 Sariel (UIUC) CS573 1 Fall 2013 1 / 28

  2. Part I . Entropy . Sariel (UIUC) CS573 2 Fall 2013 2 / 28

  3. Quote “If only once - only once - no matter where, no matter before what audience - I could better the record of the great Rastelli and juggle with thirteen balls, instead of my usual twelve, I would feel that I had truly accomplished something for my country. But I am not getting any younger, and although I am still at the peak of my powers there are moments - why deny it? - when I begin to doubt - and there is a time limit on all of us.” –Romain Gary, The talent scout. Sariel (UIUC) CS573 3 Fall 2013 3 / 28

  4. Entropy: Definition . Definition . The entropy in bits of a discrete random variable X is [ ] [ ] ∑ H ( X ) = − Pr X = x lg Pr X = x . x [ ] 1 Equivalently, H ( X ) = E lg . . Pr [ X ] Sariel (UIUC) CS573 4 Fall 2013 4 / 28

  5. Entropy intuition... . Intuition... . H ( X ) is the number of fair coin flips that one gets when getting the value of X . . Sariel (UIUC) CS573 5 Fall 2013 5 / 28

  6. Binary entropy [ ] [ ] H ( X ) = − ∑ x Pr X = x lg Pr X = x = ⇒ . Definition . The binary entropy function H ( p ) for a random binary variable that is 1 with probability p , is H ( p ) = − p lg p − (1 − p ) lg(1 − p ) . We define H (0) = H (1) = 0 . . Q: How many truly random bits are there when given the result of flipping a single coin with probability p for heads? Sariel (UIUC) CS573 6 Fall 2013 6 / 28

  7. Binary entropy [ ] [ ] H ( X ) = − ∑ x Pr X = x lg Pr X = x = ⇒ . Definition . The binary entropy function H ( p ) for a random binary variable that is 1 with probability p , is H ( p ) = − p lg p − (1 − p ) lg(1 − p ) . We define H (0) = H (1) = 0 . . Q: How many truly random bits are there when given the result of flipping a single coin with probability p for heads? Sariel (UIUC) CS573 6 Fall 2013 6 / 28

  8. Binary entropy [ ] [ ] H ( X ) = − ∑ x Pr X = x lg Pr X = x = ⇒ . Definition . The binary entropy function H ( p ) for a random binary variable that is 1 with probability p , is H ( p ) = − p lg p − (1 − p ) lg(1 − p ) . We define H (0) = H (1) = 0 . . Q: How many truly random bits are there when given the result of flipping a single coin with probability p for heads? Sariel (UIUC) CS573 6 Fall 2013 6 / 28

  9. Binary entropy: H (p) = − p lg p − (1 − p) lg(1 − p) H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 0 . 8 0 . 6 0 . 4 0 . 2 0 0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1 . H ( p ) is a concave symmetric around 1 / 2 on the interval [0 , 1] . 1 . . maximum at 1 / 2 . 2 . . H (3 / 4) ≈ 0 . 8113 and H (7 / 8) ≈ 0 . 5436 . 3 . . = ⇒ coin that has 3 / 4 probably to be heads have higher 4 amount of “randomness” in it than a coin that has probability 7 / 8 for heads. Sariel (UIUC) CS573 7 Fall 2013 7 / 28

  10. Binary entropy: H (p) = − p lg p − (1 − p) lg(1 − p) H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 0 . 8 0 . 6 0 . 4 0 . 2 0 0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1 . H ( p ) is a concave symmetric around 1 / 2 on the interval [0 , 1] . 1 . . maximum at 1 / 2 . 2 . . H (3 / 4) ≈ 0 . 8113 and H (7 / 8) ≈ 0 . 5436 . 3 . . = ⇒ coin that has 3 / 4 probably to be heads have higher 4 amount of “randomness” in it than a coin that has probability 7 / 8 for heads. Sariel (UIUC) CS573 7 Fall 2013 7 / 28

  11. Binary entropy: H (p) = − p lg p − (1 − p) lg(1 − p) H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 0 . 8 0 . 6 0 . 4 0 . 2 0 0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1 . H ( p ) is a concave symmetric around 1 / 2 on the interval [0 , 1] . 1 . . maximum at 1 / 2 . 2 . . H (3 / 4) ≈ 0 . 8113 and H (7 / 8) ≈ 0 . 5436 . 3 . . = ⇒ coin that has 3 / 4 probably to be heads have higher 4 amount of “randomness” in it than a coin that has probability 7 / 8 for heads. Sariel (UIUC) CS573 7 Fall 2013 7 / 28

  12. Binary entropy: H (p) = − p lg p − (1 − p) lg(1 − p) H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 0 . 8 0 . 6 0 . 4 0 . 2 0 0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1 . H ( p ) is a concave symmetric around 1 / 2 on the interval [0 , 1] . 1 . . maximum at 1 / 2 . 2 . . H (3 / 4) ≈ 0 . 8113 and H (7 / 8) ≈ 0 . 5436 . 3 . . = ⇒ coin that has 3 / 4 probably to be heads have higher 4 amount of “randomness” in it than a coin that has probability 7 / 8 for heads. Sariel (UIUC) CS573 7 Fall 2013 7 / 28

  13. And now for some unnecessary math . . H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 . . H ′ ( p ) = − lg p + lg(1 − p ) = lg 1 − p 2 p ( ) . . H ′′ ( p ) = p − 1 1 1 − p · = − p (1 − p ) . 3 p 2 . . ⇒ H ′′ ( p ) ≤ 0 , for all p ∈ (0 , 1) , and the H ( · ) is concave. = 4 . . H ′ (1 / 2) = 0 = ⇒ H (1 / 2) = 1 max of binary entropy. 5 . . = ⇒ balanced coin has the largest amount of randomness in it. 6 Sariel (UIUC) CS573 8 Fall 2013 8 / 28

  14. And now for some unnecessary math . . H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 . . H ′ ( p ) = − lg p + lg(1 − p ) = lg 1 − p 2 p ( ) . . H ′′ ( p ) = p − 1 1 1 − p · = − p (1 − p ) . 3 p 2 . . ⇒ H ′′ ( p ) ≤ 0 , for all p ∈ (0 , 1) , and the H ( · ) is concave. = 4 . . H ′ (1 / 2) = 0 = ⇒ H (1 / 2) = 1 max of binary entropy. 5 . . = ⇒ balanced coin has the largest amount of randomness in it. 6 Sariel (UIUC) CS573 8 Fall 2013 8 / 28

  15. And now for some unnecessary math . . H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 . . H ′ ( p ) = − lg p + lg(1 − p ) = lg 1 − p 2 p ( ) . . H ′′ ( p ) = p − 1 1 1 − p · = − p (1 − p ) . 3 p 2 . . ⇒ H ′′ ( p ) ≤ 0 , for all p ∈ (0 , 1) , and the H ( · ) is concave. = 4 . . H ′ (1 / 2) = 0 = ⇒ H (1 / 2) = 1 max of binary entropy. 5 . . = ⇒ balanced coin has the largest amount of randomness in it. 6 Sariel (UIUC) CS573 8 Fall 2013 8 / 28

  16. And now for some unnecessary math . . H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 . . H ′ ( p ) = − lg p + lg(1 − p ) = lg 1 − p 2 p ( ) . . H ′′ ( p ) = p − 1 1 1 − p · = − p (1 − p ) . 3 p 2 . . ⇒ H ′′ ( p ) ≤ 0 , for all p ∈ (0 , 1) , and the H ( · ) is concave. = 4 . . H ′ (1 / 2) = 0 = ⇒ H (1 / 2) = 1 max of binary entropy. 5 . . = ⇒ balanced coin has the largest amount of randomness in it. 6 Sariel (UIUC) CS573 8 Fall 2013 8 / 28

  17. And now for some unnecessary math . . H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 . . H ′ ( p ) = − lg p + lg(1 − p ) = lg 1 − p 2 p ( ) . . H ′′ ( p ) = p − 1 1 1 − p · = − p (1 − p ) . 3 p 2 . . ⇒ H ′′ ( p ) ≤ 0 , for all p ∈ (0 , 1) , and the H ( · ) is concave. = 4 . . H ′ (1 / 2) = 0 = ⇒ H (1 / 2) = 1 max of binary entropy. 5 . . = ⇒ balanced coin has the largest amount of randomness in it. 6 Sariel (UIUC) CS573 8 Fall 2013 8 / 28

  18. And now for some unnecessary math . . H ( p ) = − p lg p − (1 − p ) lg(1 − p ) 1 . . H ′ ( p ) = − lg p + lg(1 − p ) = lg 1 − p 2 p ( ) . . H ′′ ( p ) = p − 1 1 1 − p · = − p (1 − p ) . 3 p 2 . . ⇒ H ′′ ( p ) ≤ 0 , for all p ∈ (0 , 1) , and the H ( · ) is concave. = 4 . . H ′ (1 / 2) = 0 = ⇒ H (1 / 2) = 1 max of binary entropy. 5 . . = ⇒ balanced coin has the largest amount of randomness in it. 6 Sariel (UIUC) CS573 8 Fall 2013 8 / 28

  19. Squeezing good random bits out of bad random bits... Given the result of n coin flips: b 1 , . . . , b n from a faulty coin, with head with probability p , how many truly random bits can we extract? Sariel (UIUC) CS573 9 Fall 2013 9 / 28

  20. Squeezing good random bits out of bad random bits... . Question... . Given the result of n coin flips: b 1 , . . . , b n from a faulty coin, with head with probability p , how many truly random bits can we extract? . If believe intuition about entropy, then this number should be ≈ n H ( p ) . Sariel (UIUC) CS573 10 Fall 2013 10 / 28

  21. Back to Entropy [ ] [ ] . . entropy of X is H ( X ) = − ∑ x Pr X = x lg Pr X = x . 1 . . Entropy of uniform variable.. 2 . Example . A random variable X that has probability 1 / n to be i , for i = 1 , . . . , n , has entropy H ( X ) = − ∑ n n lg 1 1 n = lg n . . i =1 . . Entropy is oblivious to the exact values random variable can 3 have. . . = ⇒ random variables over − 1 , +1 with equal probability has 4 the same entropy (i.e., 1 ) as a fair coin. Sariel (UIUC) CS573 11 Fall 2013 11 / 28

  22. Back to Entropy [ ] [ ] . . entropy of X is H ( X ) = − ∑ x Pr X = x lg Pr X = x . 1 . . Entropy of uniform variable.. 2 . Example . A random variable X that has probability 1 / n to be i , for i = 1 , . . . , n , has entropy H ( X ) = − ∑ n n lg 1 1 n = lg n . . i =1 . . Entropy is oblivious to the exact values random variable can 3 have. . . = ⇒ random variables over − 1 , +1 with equal probability has 4 the same entropy (i.e., 1 ) as a fair coin. Sariel (UIUC) CS573 11 Fall 2013 11 / 28

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend