Illegitimi non carborundum
Ronald L. Rivest
Viterbi Professor of EECS MIT, Cambridge, MA
CRYPTO 2011 2011-08-15
1
Illegitimi non carborundum Ronald L. Rivest Viterbi Professor of - - PowerPoint PPT Presentation
Illegitimi non carborundum Ronald L. Rivest Viterbi Professor of EECS MIT, Cambridge, MA CRYPTO 2011 2011-08-15 1 Illegitimi non carborundum (Dont let the bastards grind you down!) Ronald L. Rivest Viterbi Professor of EECS MIT,
Ronald L. Rivest
Viterbi Professor of EECS MIT, Cambridge, MA
CRYPTO 2011 2011-08-15
1
Ronald L. Rivest
Viterbi Professor of EECS MIT, Cambridge, MA
CRYPTO 2011 2011-08-15
2
Ronald L. Rivest
Viterbi Professor of EECS MIT, Cambridge, MA
CRYPTO 2011 2011-08-15
3
Overview and Context The Game of “FLIPIT” Non-Adaptive Play Adaptive Play Lessons and Open Questions
4
5
6
7
8
9
10
◮ secret-sharing [S79,...] ◮ proactive crypto [HJKY95,...] ◮ signer-base intrusion-resilience [IR04,...] ◮ leakage-resilient crypto [MR04,...]
11
12
Q: “If I call the dog’s tail a leg, how many legs does it have?”
13
Q: “If I call the dog’s tail a leg, how many legs does it have?” A: “Four. It doesn’t matter what you call the tail; it is still a tail.”
14
15
16
17
18
19
20
21
◮ A password
22
◮ A password ◮ A digital signature key
23
◮ A password ◮ A digital signature key ◮ A computer system
24
◮ A password ◮ A digital signature key ◮ A computer system ◮ A mountain pass
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
◮ Note that Attacker can take over at any time.
40
◮ Note that Attacker can take over at any time. ◮ There is no “perfect defense”.
41
◮ Note that Attacker can take over at any time. ◮ There is no “perfect defense”. ◮ Only option for Defender is to re-take control
42
◮ Note that Attacker can take over at any time. ◮ There is no “perfect defense”. ◮ Only option for Defender is to re-take control
◮ The game may go on forever...
43
◮ In practice, compromise is often
44
◮ In practice, compromise is often
◮ In FL I PIT,
45
◮ In practice, compromise is often
◮ In FL I PIT,
◮ Player’s uncertainty about system state
46
◮ In practice, compromise is often
◮ In FL I PIT,
◮ Player’s uncertainty about system state
◮ A move may take control (“flip”) or have no
47
◮ In practice, compromise is often
◮ In FL I PIT,
◮ Player’s uncertainty about system state
◮ A move may take control (“flip”) or have no
◮ Uncertainty means flops are unavoidable.
48
◮ A player learns the state of the system only
49
◮ A player learns the state of the system only
◮ In basic FL I PIT, each move has feedback
50
◮ A player learns the state of the system only
◮ In basic FL I PIT, each move has feedback
◮ In variants, move reveals only current state,
51
◮ Moves aren’t for free!
52
◮ Moves aren’t for free! ◮ Player i pays ki points per move:
53
◮ Moves aren’t for free! ◮ Player i pays ki points per move:
◮ Being in control yields gain!
54
◮ Moves aren’t for free! ◮ Player i pays ki points per move:
◮ Being in control yields gain! ◮ Player earns one point for each second he is
55
◮ Let Ni(t) denote number moves by player i
56
◮ Let Ni(t) denote number moves by player i
◮ Let Gi(t) denote the number of seconds
57
◮ Score (net benefit) Bi(t) up to time t is
◮ Benefit rate is
◮ Player wishes to maximize βi = limt→∞ βi(t).
58
59
60
61
62
◮ A non-adaptive strategy plays on blindly,
63
◮ A non-adaptive strategy plays on blindly,
◮ In principle, a non-adaptive player can
64
◮ A non-adaptive strategy plays on blindly,
◮ In principle, a non-adaptive player can
◮ Some interesting non-adaptive strategies:
65
◮ A non-adaptive strategy plays on blindly,
◮ In principle, a non-adaptive player can
◮ Some interesting non-adaptive strategies:
◮ Periodic play 66
◮ A non-adaptive strategy plays on blindly,
◮ In principle, a non-adaptive player can
◮ Some interesting non-adaptive strategies:
◮ Periodic play ◮ Exponential (memoryless) play 67
◮ A non-adaptive strategy plays on blindly,
◮ In principle, a non-adaptive player can
◮ Some interesting non-adaptive strategies:
◮ Periodic play ◮ Exponential (memoryless) play ◮ Renewal strategies: iid intermove times 68
69
70
71
◮ a sentry make his regular rounds ◮ 90-day password reset
72
Theorem
◮ if α1 > 1/2k0, don’t play(!), ◮ if α1 = 1/2k0, play periodically at any rate α0,
◮ if α1 < 1/2k0, play periodically at rate
73
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
if α1 >
1 2k0 Attacker too fast for Defender
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
if α1 =
1 2k0
Defender can play with 0 benefit
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
if α1 <
1 2k0
Defender maximizes benefit with α0 =
2k0
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
Optimal Attacker play
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
Optimal Attacker play
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
Optimal Attacker play
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
Optimal Attacker play
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
Nash equilibrium at (α0, α1) = ( 1
3, 2 9)
(k0 = 1, k1 = 1.5)
2 3 1 2 1 3 1 6
2 3 1 2 1 3 1 6
Nash equilibrium at (α0, α1) = ( 1
3, 2 9)
(γ0, γ1) = ( 2
3, 1 3)
(β0, β1) = ( 1
3, 0)
86
87
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
Optimal Defender play for α1 < 1 α0 =
k0 − α1
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
Optimal Attacker play
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
Optimal Attacker play
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
Nash equilibrium at (α0, α1) = ( 6
25, 4 25)
(k0 = 1, k1 = 1.5)
1
2 3 1 3
1
2 3 1 3
Nash equilibrium at (α0, α1) = ( 6
25, 4 25)
(γ0, γ1) = ( 3
5, 2 5)
(β0, β1) = ( 9
25, 6 25)
96
97
Theorem
98
99
100
101
102
◮ Periodic strategy not very effective against
103
◮ Periodic strategy not very effective against
◮ FL I PIT with adaptive strategies can be
104
◮ Periodic strategy not very effective against
◮ FL I PIT with adaptive strategies can be
105
Theorem
106
2 3 1 3
2 3 1 3
Periodic Attacker Periodic Defender
2 3 1 3
2 3 1 3
Periodic Attacker Periodic Defender
Adaptive Attacker Exponential Defender
2 3 1 3
2 3 1 3
Periodic Attacker Periodic Defender
Adaptive Attacker Exponential Defender ∃ ? Better Defender ?
109
110
◮ Be prepared to deal with continual repeated
111
◮ Be prepared to deal with continual repeated
◮ Play fast! Aim to make opponent drop out!
112
◮ Be prepared to deal with continual repeated
◮ Play fast! Aim to make opponent drop out!
◮ Arrange game so that your moves cost much
113
114
115
116
117
118
119
120
121