self organizing search in the web of things
play

Self-Organizing Search in the Web of Things Kay Rmer, University of - PowerPoint PPT Presentation

Self-Organizing Search in the Web of Things Kay Rmer, University of Lbeck Networked Embedded Sensing Research Protocols Adaptation to interference Programming Models Role assignment Services Content-based Sensor Search


  1. Self-Organizing Search in the Web of Things Kay Römer, University of Lübeck

  2. Networked Embedded Sensing Research  Protocols  Adaptation to interference  Programming Models  Role assignment  Services  Content-based Sensor Search  Minimally-Invasive Management  Systems  Body sensor networks

  3. Motivation Camera Accelerometer  Mobile phones equipped GPS Ambient Light Bluetooth Microphone with sensors and connected GSM Magnetometer to the Internet WLAN Proximity  Sensors published on the Web: state of the real world available in real-time  Search the real world by  Search the real world by its current state ! its current state ! 2

  4. 3 Mensa empty quiet A Web of Real-World Places tation hectic full S upermarket quiet full S

  5. Web of Things  Web presence of things, people, and places with real-time state information  Web of real-world entities, not Web of sensors  High-level states, not raw sensor data  Searching the Web of Things  Search for real-world entities: places, people, things, …  by their current state: empty, hot, broken, …  in real-time 4

  6. Searching the Real World: Examples  Quiet picnic places at waterfront?  Route through city avoiding traffic jams?  Which rental station has bicycles available?  Where are many people who share my interests?  Which trains from A to B are not crowded?  Where to enter train to get free seat?  Supermarkets with short waiting queues? 5

  7. Problem: Content-based Sensor Search  Find sensors reading given state in real time  Potentially huge, distributed set of candidate sensors  More state updates than queries, push not a good idea!  Sensor output is highly dynamic  Indexing sensor output not a good idea!  We need only a limited number of results at a time  Heuristics to select good candidates! 6

  8. Approach: Sensor Ranking  Sensors create prediction model using past readings Indexing Time  Prediction models are published on the Web  Search engine periodically indexes prediction models  Prediction models are used to rank candidate sensors Query Time  Highest ranking sensors are read first  Goal: Minimize the number of read sensors 7

  9. System Model  Sensor maps discrete time to a finite discrete set of states: s : T  V   Sensor output time series: s( t i ) = v i Sensor Output ... V 1 V 2 V 3 V 4 ... V present ... t 1 t 2 t 3 t 4 t present ... Time 8

  10. System Model (Continued)  Prediction model maps query time and query value to a probability estimate:    P T V  : 0 , 1  P( t , v ): Probability that s( t ) = v search for v model construction Sensor ... V 1 V 2 V 3 V 4 ... V c ... ? Output ... t 1 t 2 t 3 t 4 ... t C ... t Time time window TW forecasting horizon 9

  11. Query Resolution Example: Quiet places at waterfront 1. Filter static (waterfront, occupancy) 2. Predict (quiet) 3. Rank 4. Read 5. Return .7 .1 .9 .5 .6 .2 .2

  12. Ranking Metrics  Normalized overhead for reading non-matching sensors  Ranking error e(t,v) Number of non-matching sensors above last matching sensor Rank of last matching sensor  Top-m ranking error e top (t,v,m)  Dito, but only first m sensors considered 15

  13. Ranking Metrics: Examples Suboptimal ranking: Optimal ranking: S1 S1 • e = 4/8 • e = 0/6 S2 S2 • e top = 2/5 • e top = 0/5 S3 S3 S4 S4 m m S5 S5 Last match S6 S6 Sensors that S7 S7 match the query Last match S8 S8 Sensors that do not match the S9 S9 query 16

  14. Prediction Models  Focus on people-centric sensors  Tend to show periodic behaviour  Requirements  Accurate predictions for forecasting horizons that match indexing frequencies (days - weeks)  Deal with imperfect periodic behavior 17

  15. Considered Prediction Models  Single-period prediction model (SPPM)  Assumes single dominant period of known length (e.g., 1 week)  Multi-period prediction model (MPPM)  Assumes multiple periodic processes of unknown length (e.g., 1 week, 4 weeks)  Select appropriate models at runtime 18

  16. Single-Period Prediction Model (SPPM)  Assumption: Single dominant period with length p p p p p Sensor V 1 V 1 V 2 V 4 V 4 ... ... ... ? V 3 Output t C t 1 t 3 t 4 ... t t 2 ... Time time window TW P(t,green) = 2/4 P(t,red) = 1/4 19

  17. Number of consecutive appearances of symbol α in period p at offset l Max. possible occurences Multi-Period Prediction Model (MPPM)   p l  Periodic symbol ( , , , )  α : symbol ; p : period ; l : offset ;  : support  Example: α =blue, p =2, l =1 p =2 p =2 p =2 l =1 Sensor Output   ps  blue ,2,1,2     3 21

  18. Inferring Prediction Estimates Query for value v=b at time t=6 1. Filter periodic symbols Same value:  = v  Same phase: l  t mod p  2. P(v,t) = max  v = b ?   p l b 2 0 0.7 g 3 1 0.1 b 4 2 0.9 0 1 2 3 4 5 6 23

  19. Adjustment Process  Faulty/malicious sensors, inaccurate S1 predictions may result in persistent S2 misranking S3  Individual ranking error for each sensor S4  S8 ranked to low: increase prediction value S5  S7 ranked to high: decrease prediction value S6  Idea: adjustment term for each sensor S7  Updated after each query using ranking error S8 E7 = -2/9 E8 = +6/9 S9 24

  20. Adjustment Process: Feedback Loop C3 A1 + S1 S1 S3 S1 P1 P3 + + E3 E1 C3 A2 + S2 S2 P2 P3 S3 S2 + + E2 E3 C3 A3 + S3 S3 S3 S3 P3 P3 + + E3 E3 C3 A4 + S4 S4 P4 P3 + + E4 E3 S4 S3 C3 A5 + S5 S5 S3 S5 P3 P5 + + E3 E5 25 25

  21. Evaluation  Simulation of a realistic search engine  Periodic rebuild and indexing of models (1 week)  Periodic queries for possible values  Measure average ranking error  Prediction models: Random, SPPM, MPPM  With / without adjustment 26

  22. Evaluation: Data Sets  MERL motion detector dataset  50 PIR sensors in office building  PIR output mapped to “ free” and “ occupied”  With and without a “faulty” sensor  ETH room reservation system  7 “sensors”  Room occupancy: “ free ” or “ occupied ”  With and without “synthethic” multiperiod sensor  Bicing data set (in progress)  350 bicycle rental stations in Barcelona  Number of available bicycles: “no”, “few”, “many”

  23. Average Ranking Error: MERL Average Ranking Error P P D D MPPM SPPM+AP SPPM+AP MPPM M SPPM A A N N P + A + A P M M R R S P P P P M M 30

  24. Average Ranking Error vs. Top m Average Ranking Error x5.5 x10 #Top Entries 31

  25. Summary  Ubiquitous sensors connected to Internet  Search for real-world entities by current state  Sensor Ranking, a primitive for content-based sensor search utilizing prediction models  Adjustment process to alleviate persistent inaccurate rankings  Promising results on real-world data sets  Ongoing work  Improved ranking based on correlations  Building a search engine 32

  26. Ads  Act-Control-Move: Beyond networked Sensors  Summer School, Schloss Dagstuhl, August 15-21, 2010  www.cooperating-objects.eu/school  IEEE SUTC (Sensor Networks, Ubiquitous & Trustworthy Computing)  Conference, Newport Beach, California, June 7-9, 2010  sutc2010.eecs.uci.eu  SESENA (Software Engineering for Sensor Nets)  ICSE Workshop, CapeTown, South Africa, May 3, 2010  www.sesena.info

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend