simultaneous and sequential detection of multiple
play

Simultaneous and sequential detection of multiple interacting change - PowerPoint PPT Presentation

Simultaneous and sequential detection of multiple interacting change points Long Nguyen Department of Statistics University of Michigan Joint work with Ram Rajagopal (Stanford University) 1 Introduction Statistical inference in the


  1. Simultaneous and sequential detection of multiple interacting change points Long Nguyen Department of Statistics University of Michigan Joint work with Ram Rajagopal (Stanford University) 1

  2. Introduction • Statistical inference in the context of spatially distributed data processed and analyzed by decentralized systems – sensor networks, social networks, the Web • Two interacting aspects – how to exploit the spatial dependence in data – how to deal with decentralized communication and computation 2

  3. Introduction • Statistical inference in the context of spatially distributed data processed and analyzed by decentralized systems – sensor networks, social networks, the Web • Two interacting aspects – how to exploit the spatial dependence in data – how to deal with decentralized communication and computation • Extensive literature dealing with each of these two aspects separately by different communities • Many applications call for handling both aspects in near “real-time” data processing and analysis 2-a

  4. Example – Sensor network for traffic monitoring Problem: detecting sensor failures for all sensors in the network • data: sequence of sensor measurements of traffic volume • sequential detection rule for change (failure) point, one for each sensor 3

  5. “Mean days to failure” • as many as 40% sensors fail a given day • need to detect failed sensors as early as possible • separating sensor failure from events of interest is difficult 4

  6. Talk outline • statistical formulation for detection of multiple change points in a network setting – classical sequential analysis – graphical models • sequential and “real-time” message-passing detection algorithms – decision procedures with limited data and computation • asymptotic theory of the tradeoffs between statistical efficiency vs. com- putation/communication efficiency 5

  7. Sequential detection for single change point • sensor u collects sequence of data X n ( u ) for n = 1 , 2 , . . . • λ u change point variable for sensor u • data are i.i.d. according to f 0 before the change point; and iid f 1 after 6

  8. Sequential detection for single change point • sensor u collects sequence of data X n ( u ) for n = 1 , 2 , . . . • λ u change point variable for sensor u • data are i.i.d. according to f 0 before the change point; and iid f 1 after • a sequential change point detection procedure is a stopping time τ u , i.e., { τ u ≤ n } ∼ σ ( X 1 ( u ) , . . . , X n ( u )) 6-a

  9. Sequential detection for single change point • sensor u collects sequence of data X n ( u ) for n = 1 , 2 , . . . • λ u change point variable for sensor u • data are i.i.d. according to f 0 before the change point; and iid f 1 after • a sequential change point detection procedure is a stopping time τ u , i.e., { τ u ≤ n } ∼ σ ( X 1 ( u ) , . . . , X n ( u )) • Neyman-Pearson criterion: – constraint on false alarm error PFA ( τ u ( X )) = P ( τ u < λ u ) ≤ α for some small α – minimum detection delay E [( τ u − λ u ) | τ u ≥ λ u ] . 6-b

  10. Beyond a single change point • we have multiple change points, one for each sensor • we could apply the single change point method to each sensor independently, but this is not a good idea – measurements from a single sensor are very noisy – failed sensors may still produce plausible measurement values • borrowing information from neighboring sensors may be useful – due to spatial dependence of measurements – but data sharing limited to neighboring sensors – data sharing via a message-passing mechanism 7

  11. Sample correlation with neighbors Correlation with good sensors Correlation with failed sensors 8

  12. Correlation statistics have been successfully utilized in practice, although not in a sequential and decentralized setting (Kwon and Rice, 2003) 9

  13. A formulation for multiple change points • m sensors labeled by U = { u 1 , . . . , u m } • given a graph G = ( U, E ) that specifies the the connections among u ∈ U • each sensor u fails at time λ u – λ u is endowed with (independent) prior distribution π u 10

  14. A formulation for multiple change points • m sensors labeled by U = { u 1 , . . . , u m } • given a graph G = ( U, E ) that specifies the the connections among u ∈ U • each sensor u fails at time λ u – λ u is endowed with (independent) prior distribution π u • there is private data sequence X n ( u ) for sensor u – private data sequence changes its distribution after λ u 10-a

  15. A formulation for multiple change points • m sensors labeled by U = { u 1 , . . . , u m } • given a graph G = ( U, E ) that specifies the the connections among u ∈ U • each sensor u fails at time λ u – λ u is endowed with (independent) prior distribution π u • there is private data sequence X n ( u ) for sensor u – private data sequence changes its distribution after λ u • there is shared data sequence ( Z n ( u, v )) n for each neighboring pair of sensors u and v : iid ∼ f 0 ( ·| u, v ) , for n < min( λ u , λ v ) Z n ( u, v ) iid ∼ f 1 ( ·| u, v ) , for n ≥ min( λ u , λ v ) 10-b

  16. Graphical model of change points (a) Topology of sensor network (b) Graphical model of random variables • Conditionally on the shared data sequences, change point variables are no longer independent 11

  17. Localized stopping times • Data constraint. Each sensor has access to only shared data with its neighbors • Definition. Stopping rule for u , denoted by τ u , is a localized stopping time , which depends on measurements of u and its neighbors: – for any t > 0 : � � { τ u ≤ t } ∈ σ { X n ( u ) , Z n ( u, v ) | n ≤ t, v ∈ N ( u ) } 12

  18. Performance metrics • false alarm rate PFA ( τ u ) = P ( τ u ≤ λ u ) . • expected failure detection delay D ( τ u ) = E [ τ u − λ u | τ u ≥ λ u ] . • Problem: for each sensor u , find a localized stopping time τ u τ u D ( τ u ) such that PFA ( τ u ) ≤ α. min 13

  19. Review of results for single change point detection • optimal sequential rule is a stopping rule by thresholding the posterior of λ u under some conditions: (Shiryaev, 1978) τ u ( X ) = inf { n : Λ n ≥ 1 − α } , where Λ n = P ( λ u ≤ n | X 1 ( u ) , . . . , X n ( u )) . • well-established asymptotic properties (Tartakovsky & Veeravalli, 2006): – false alarm: PFA ( τ u ( X )) ≤ α. – detection delay: | log α | � � as α → 0 . D ( τ u ( X )) = 1 + o (1) q ( X ) + d – here q ( X ) = KL ( f 1 ( X ) || f 0 ( X )) , the Kullback-Leibler information, d some constant 14

  20. Two sensor case: An initial idea λ u λ v X Z Y u v X Z Y • Idea: use both private data X 1 , . . . , X n and shared data Z 1 , . . . , Z n : τ u ( X, Z ) = inf { n : P ( λ u ≤ n | ( X 1 , Z 1 ) , . . . , ( X n , Z n )) ≥ 1 − α } . 15

  21. Two sensor case: An initial idea λ u λ v X Z Y u v X Z Y • Idea: use both private data X 1 , . . . , X n and shared data Z 1 , . . . , Z n : τ u ( X, Z ) = inf { n : P ( λ u ≤ n | ( X 1 , Z 1 ) , . . . , ( X n , Z n )) ≥ 1 − α } . • Theorem 1: The false alarm for τ u ( X, Z ) is bounded from above by α , while expected delay takes the form: „ « | log α | D ( τ u ( X, Z )) = 1 + o (1) as α → 0 . q ( X ) + d – Z not helpful in improving the delay (at least in the asymptotics!) – this suggests to use information from Y as well (to predict λ u ) 15-a

  22. Localized stopping rule with message exchange • Modified Idea: – u should use information given by shared data Z only if its neighbor v has not changed (failed) ... – but u does not know whether v has changed or not, so ... – instead of deciding this by itself, u will wait for v to tell it X Z Y u v message 16

  23. Localized stopping rule with message exchange • Modified Idea: – u should use information given by shared data Z only if its neighbor v has not changed (failed) ... – but u does not know whether v has changed or not, so ... – instead of deciding this by itself, u will wait for v to tell it X Z Y u v message Stopping rule for u ultimately hinges also information given by data sequence Y , passed to u indirectly via neighbor sensor v 16-a

  24. Localized stopping rule with information exchange • Algorithmic Protocol: – each sensor uses all data shared with neighbors that have not de- clared to change (fail) – if a sensor v stops according to its stopping rule, v broadcasts this information to all its neighbors, who promptly drop v from the list of their respective neighbors 17

  25. Localized stopping rule with information exchange • Algorithmic Protocol: – each sensor uses all data shared with neighbors that have not de- clared to change (fail) – if a sensor v stops according to its stopping rule, v broadcasts this information to all its neighbors, who promptly drop v from the list of their respective neighbors • Formally, for two sensors: – stopping rule for u , using only X : τ u ( X ) – stopping rule for u , using both X and Z : τ u ( X, Z ) – similarly, for sensor v : τ v ( Y ) and τ v ( Y, Z ) – then, the overall stopping rule for u is: 8 τ u ( X, Z ) if τ u ( X, Z ) ≤ τ v ( Y, Z ) < τ u ( X, Y, Z ) = ¯ max( τ u ( X ) , τ v ( Y, Z )) otherwise : 17-a

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend