a closer look at the classical fixed point analysis of
play

A closer look at the classical fixed-point analysis of WLANs Rajesh - PowerPoint PPT Presentation

A closer look at the classical fixed-point analysis of WLANs Rajesh Sundaresan Indian Institute of Science May 2017 Section 1 The classical fixed point analysis DCF: The 802.11 countdown and its Markovian caricature N nodes accessing the


  1. A closer look at the classical fixed-point analysis of WLANs Rajesh Sundaresan Indian Institute of Science May 2017

  2. Section 1 The classical fixed point analysis

  3. DCF: The 802.11 countdown and its Markovian caricature ◮ N nodes accessing the common medium in a wireless LAN. Infinite backlog of packets. Attempts to transmit HOL packet. ◮ Each node’s (backoff) state space: Z = { 0 , 1 , · · · , m − 1 } . Backoff state determines attempt probability for a node in a slot. ◮ Transitions: 0 1 i i + 1 m - 1

  4. Example design – exponential backoff ◮ Assume three states, Z = { 0 , 1 , 2 } or m = 3. ◮ Attempt probability for node in state i is c i / N . ◮ Aggressiveness of the transmission c = ( c 0 , c 1 , c 2 ). ◮ The scaling by 1 / N ensures that the overall attempt probability of a single node is O (1 / N ) so that the overall (system) attempt probability for the system is O (1). ◮ Conventional wisdom: exponential backoff: c i = c i − 1 / 2 . Double the average waiting time after every failure.

  5. Back-of-envelope analysis ◮ Observation: your collision probability depends only on the empirical measure of node states ... excepting you. ◮ ξ = current empirical measure of nodes across states. ◮ Number of nodes across states is ( N ξ 0 , N ξ 1 , . . . , N ξ m − 1 ). ◮ If you are in state 0, others states ( N ξ 0 − 1 , N ξ 1 , . . . , N ξ m − 1 ). Probability that no one else transmits is: m − 1 m − 1 1 − c 0 � N ξ 0 − 1 1 − c i � N ξ i 1 − c 0 � − 1 1 − c i � N ξ i � � � � � � · = · N N N N i =1 i =0 e −� c ,ξ � . → ◮ � c , ξ � is the attempt probability: � i ( N ξ i )( c i / N ). ◮ If N is small or if attempt probabilities don’t scale, avoid the limit.

  6. The classical fixed-point analysis 1. Conditional collision probability, when making an attempt, is the same for each node in each state. γ := 1 − e −� c ,ξ � = 1 − e − ( attempt ) This amounts to assuming that spatial distribution stabilises at ξ . 2. The system interactions decouple. 3. Focus on a node. Consider renewal instants of return to state 0. From the renewal-reward theorem: 0 1 i i + 1 m - 1 1 + γ + γ 2 + . . . attempt E [ Reward ] c 2 + . . . =: G ( γ ) = E [ RenewalTime ] = . N c 0 + γ N c 1 + γ 2 N N N 4. Solve for the fixed point: γ = 1 − e − G ( γ ) .

  7. Goodness of the approximation (from Bianchi 1998) Plot for fixed-point analysis without taking N → ∞ . W is the window size in the basic WLAN protocol.

  8. In this talk: We will see an overview of ◮ why decoupling is a good assumption; ◮ when node independent, state independent, conditional collision probability assumption holds; ◮ and going a little beyond what to do when the ’node/state independent collision probability’ does not hold.

  9. Section 2 The decoupling assumption

  10. Mean-field interaction ◮ Coupled dynamics. ◮ Embed slot boundaries on R + . Assume slots of duration 1 / N . ◮ Transition rate = prob. of change in a slot / slot duration = O (1). ◮ Transition rate for success or failure depends on the states of the other nodes, but only through the empirical measure µ N ( t ) of nodes across states. ◮ At time t , node transition rates are as follows: ◮ i � i + 1 with rate λ i , i +1 ( µ N ( t )). ◮ i � 0 with rate λ i , 0 ( µ N ( t )). ◮ In general, i � j with rate λ i , j ( µ N ( t )).

  11. The transition rates If µ N ( t ) = ξ , then ◮ Example: λ 0 , 1 ( ξ ) = ( c 0 / N )(1 − e − attempt ) = c 0 (1 − e −� c ,ξ � ) . 1 / N ◮ Write as a matrix of rates: Λ( · ) = [ λ i , j ( ξ ) ] i , j ∈Z . ◮ For ξ , the empirical measure of a configuration, the rate matrix is  c 0 (1 − e −� c ,ξ � )  − 0  . c 1 e −� c ,ξ � c 1 (1 − e −� c ,ξ � ) Λ( ξ ) = −  c 2 e −� c ,ξ � 0 − For today’s exposition, we will assume this continuous-time caricature with these instantaneous transition rates. This is different, since at most one node can transit at any time.

  12. The Markov processes, big and small ◮ ( X ( N ) ( · ) , 1 ≤ n ≤ N ), the trajectory of all the n nodes, is Markov n ◮ Study µ N ( · ) instead, also a Markov process Its state space size is the set of empirical probability measures on N particles with state space Z . ◮ Then try to draw conclusions on the original process.

  13. The smaller Markov process µ N ( · ) ◮ A Markov process with state space being the set of empirical measures of N nodes. ◮ This is a measure-valued flow across time. ◮ In the continuous-time version: the transition ξ � ξ + 1 N e j − 1 N e i occurs at rate N ξ ( i ) λ i , j ( ξ ). ◮ For large N , changes are small, O (1 / N ), at higher rates, O ( N ). Individuals are collectively just about strong enough to influence the evolution of the measure-valued flow. ◮ Fluid limit : µ N converges to a deterministic limit given by an ODE.

  14. The conditional expected drift in µ N ◮ Recall Λ( · ) = [ λ i , j ( · ) ] without diagonal entries. Then 1 h E [ µ N ( t + h ) − µ N ( t ) | µ N ( t ) = ξ ] = Λ( ξ ) ∗ ξ lim h ↓ 0 with suitably defined diagonal entries.

  15. An interpretation ◮ The rate of change in the k th component is made up of increase � ( N ξ i ) · λ i , k ( ξ ) · (+1 / N ) i : i � = k ◮ and decrease � ( N ξ k ) λ k , i ( ξ )( − 1 / N ) . i : i � = k ◮ Put these together: � � � ξ i λ i , k ( ξ ) = (Λ( ξ ) ∗ ξ ) k . ξ i λ i , k ( ξ ) − ξ k λ k , i ( ξ ) = i : i � = k i : i � = k i

  16. The conditional expected drift in µ N ◮ Recall Λ( · ) = [ λ i , j ( · ) ] without diagonal entries. Then 1 h E [ µ N ( t + h ) − µ N ( t ) | µ N ( t ) = ξ ] = Λ( ξ ) ∗ ξ lim h ↓ 0 with suitably defined diagonal entries. ◮ Anticipate that µ N ( · ) will solve (in the large N limit) Λ( µ ( t )) ∗ µ ( t ) , µ ( t ) ˙ = t ≥ 0 [McKean-Vlasov equation] µ (0) = ν ◮ Nonlinear ODE.

  17. A limit theorem Theorem p Suppose that the initial empirical measure µ N (0) → ν , where ν is deterministic. Let µ ( · ) be the solution to the McKean-Vlasov dynamics with initial condition µ (0) = ν . p Then µ N ( · ) → µ ( · ) . Technicalities: ◮ The McKean-Vlasov ODE must be well-posed. p ◮ µ N (0) → ν : Probability of being outside a ball around ν vanishes. p ◮ µ N ( · ) → µ ( · ): For any finite duration, probability of being outside a tube around µ ( · ) vanishes.

  18. Back to the individual nodes ◮ Let µ ( · ) be the solution to the McKean-Vlasov dynamics ◮ Choose a node uniformly at random, and tag it. ◮ µ N ( · ) is the distribution for the state of the tagged node at time t . ◮ As N → ∞ , the limiting distribution is then µ ( t )

  19. Joint evolution of tagged nodes ◮ Tag k nodes. ◮ If the interaction is only through µ N ( t ), and this converges to a deterministic µ ( t ), the transition rates are just λ i , j ( µ ( t )). ◮ Each of the k nodes is then executing a time-dependent Markov process with transition rate matrix Λ( µ ( t )). ◮ Asymptotically, no interaction ... decoupling. ◮ The node trajectories are (asymptotically) independent and identically distributed.

  20. Section 3 The steady state assumption

  21. The fixed-point analysis again ◮ Solve for the rest point of the dynamical system. Λ( ξ ) ∗ ξ = 0 . ◮ If the solution is unique, predict that the system will settle down at ξ ⊗ ξ ⊗ . . . ⊗ ξ . ◮ Works very well for the exponential backoff. ◮ But not in general due to limit cycles.

  22. A malware propagation example (from Benaim and Le Boudec 2008) ◮ The fixed point is unique, but unstable. ◮ All trajectories starting from outside the fixed point, and all trajectories in the finite N system, converge to the stable limit cycle.

  23. What is the issue? ◮ Large time behaviour for a finite N system: lim t →∞ µ N ( t ). If N is large, we really want: � � lim t →∞ µ N ( t ) lim . N →∞ ◮ But we are trying to predict where the system will settle from the following: � � lim N →∞ µ N ( t ) lim = lim t →∞ µ ( t ) . t →∞ ◮ We need a little bit of robustness of the ODE for this work.

  24. Does the method work? Theorem Let µ N (0) → ν in probability. Let the ODE have a (unique) globally asymptotically stable equilibrium ξ f with every path tending to ξ f . d Then µ N ( ∞ ) → ξ f . It is not enough to have a unique fixed point ξ f . But if that ξ f is globally asymptotically stable, that suffices.

  25. A sufficient condition A lot of effort has gone into identifying when we can ensure a globally asymptotic stable equilibrium. Theorem If c is such that � c , ξ � < 1 for all ξ , then the rest point ξ f of the dynamics is unique, and all trajectories converge to it. This is the case for the classical exponential backoff with c 0 < 1.

  26. The case of multiple stable equilibria for the ODE 1 0.9 0.8 0.7 0.6 Fraction in state 1 0.5 0.4 0.3 0.2 0.1 0 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Fraction in state 0 ◮ Different parameters: c = (0 . 5 , 0 . 3 , 8 . 0). ◮ There are two stable equilibria. One near (0 . 6 , 0 . 4 , 0 . 0) and another near (0 , 0 , 1).

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend