heaven a framework for a systematic comparative research
play

Heaven, a framework for a systematic comparative research approach - PowerPoint PPT Presentation

Heaven, a framework for a systematic comparative research approach of RSP Engines Stream Reasoning Workshop 2016 R. Tommasini,E. Della Valle, M. Balduini, D. DellAglio 1 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini -


  1. Heaven, a framework for a systematic comparative research approach of RSP Engines Stream Reasoning Workshop 2016 R. Tommasini,E. Della Valle, M. Balduini, D. Dell’Aglio 1 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  2. SRW 16 What is the goal of a benchmark? 2 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  3. SRW16 The goal of a domain specific benchmark 
 is to foster technological progress 
 by guaranteeing a fair assessment . 
 - Jim Gray, The Benchmark Handbook for Database and Transaction Systems, 1993 
 3 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  4. RSP Engines Vs Benchmarking Qualitative SRW16 SKB Citybench LSBench C-SPARQL EP-SPARQL YABench SRBench Morph Stream Sparkwave CSRBench CQELS INSTANS (2008) (2010) (2011) (2012) (2013) (2015) (2016) time 4 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  5. A Well-Known Hypothesis SRW16 The incremental maintenance of the materialisation 
 is faster then full re-materialisation 
 of the ontological entailment 
 when content changes are 
 small enough (e.g. about 10%). 5 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  6. Uncomfortable Truths in RSP Benchmarking 1000 SRW16 6 100 Memory (mb) 10 4 1 0.05 0.10 0.20 Naive Latency (ms) Window Cardinality Incremental (# Triples ) 6 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  7. Uncomfortable Truths in RSP Benchmarking SRW16 1000 6 100 Memory (mb) 10 4 1 1 2 5 Latency (ms) Abox Cardinality Naive (# Triples ) Incremental 7 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  8. Observations SRW16 •RSP Engines show heterogenous operational semantics that called for •Comparative Research (Benchmarks). •Unifying formal models (RSP-QL). •The community is doing a lot of effort to develop benchmarking solutions. •Each of the proposed benchmarks highlights new aspects regarding RSP performance. 8 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  9. Adoption of Comparative Research SRW16 •It is natively case driven: •It considers cases as a combination of known properties •It defines analysis guidelines through baselines •It is extensively used to analyse complex systems •It provides layered frameworks to •systematically examine cases •identify similarities/differences enabling us to catch more insights. 9 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  10. In practice SRW16 •Qualitatively, is there a solution that always outperforms the others? •If no dominant solution can be found, when does a solution work better than another one? •Quantitatively, is there a solution that distinguishes itself from the others? •Why does a solution performs better than another solution under a certain experimental condition? 10 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  11. SRW 16 Do we need comparative research? let’s take a look to what other fields do 11 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  12. •Systems standard semantics (SPARQL 1.0/1.1) 🙃 Linked •Several Benchmarks 🙃 •No benchmark principles ☹ Data •Benchmark Requirements 🙃 •Clear Metrics 🙃 Benchmarks •Baselines 🙃 •Standard by Adoption ☹ •LDBC and Hobbit 😑 12 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  13. •Systems standard semantics 
 (OWL Profiles) 🙃 •Few Benchmarks (growing) 🙃 OWL •No benchmark principles ☹ Reasoning •Benchmark Requirements 😑 •Clear Metrics 🙃 Benchmarks •Baselines 🙃 •Standard by Adoption ☹ • ORE 😑 13 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  14. •(Several) System standard semantics (SQL dialects) 🙄 •Domain Specific Benchmarks 🙃 Databases •Jim Gray’s benchmark principles 🙃 •Benchmark Requirements* 🙃 Benchmarks •Clear Metrics 🙃 •Baselines 🙃 •Standard by TPC 🙃 14 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  15. SRW 16 What about RSP? 15 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  16. • System standard semantics 
 (But unified) 🙃 • Several Benchmarks 🙃 • Seven Commandments for RSP RSP benchmarking 🙃 Benchmarks • No benchmark requirements ☹ • Baselines by adoption 😑 • No clear metrics ☹ • No standard ☹ 16 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  17. Claims SRW16 •Comparative research is not systematic for RSP: •it needs requirements/methodology •it needs fixed metrics •it needs fixed baselines 17 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  18. SRW 16 Can we enable a systematic comparative research approach for RSP Engines? 18 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  19. Heaven SRW16 A set of requirements to satisfy. An architecture for an RSP engine Test Stand. Two baseline RSP engine architectures A proof-of-concept implementation (open source) 19 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  20. [R1] RSP engine agnostic Comparability [R2] Independent from the measured key performance indicators Requirements [R3] Identify baseline RSP engines 20 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  21. Reproducibility [R4] Data independent Requirements [R5] Query independent 21 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  22. Repeatability [R6] Minimise the experimental error Requirements 22 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  23. RSP Experiment SRW16 RSP Experiment is a function Exp(____________) E T D Q , , , , K [R1] [R4] [R5] [R2] R eport 23 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  24. Test Stands Black box evaluation of complex system that controls the experimental error [R6] 24 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  25. MB MB SRW16 Interface Interface Streamer Input output Receiver RSPEngine D < , Q > T Start Stop ResultCollector K 25 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  26. Pipeline of DSMS and a reasoner; Support reasoning under the RSP Baselines ρ DF entailment regime; Data can flows from the DSMS to the reasoner via snapshots or differences. The minimal meaningful approaches to realise an They exploit absolute time, RSP engine [R3] i.e. their internal clock can be externally controlled. 26 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  27. Conclusion SRW16 RSP needs comparative analysis [Claim]. Heaven enables a systematic execution of experiments , required by comparative investigations. [Promise] 27 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  28. Promise SRW16 •Qualitatively, is there a solution that always outperforms the others? •If no dominant solution can be found, when does a solution work better than another one? •Quantitatively, is there a solution that distinguishes itself from the others? •Why does a solution performs better than another solution under a certain experimental condition? 28 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  29. Example of Intra Experiment Comparison SRW16 29 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  30. Example of Inter Experiment Analysis SRW16 30 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  31. Future Works SRW16 Systematic analysis of existing solutions A web-based environment where a users can: choose one of existing benchmarks (datasets, queries) design of experiments; consult and compare the results against the baselines or existing integrated RSP engines. 31 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

  32. Questions? Email : riccardo.tommasini@polimi.it 
 Twitter : @rictomm Github : riccardotommasini Web : streamreasoning.org 32 Stream Reasoning Workshop, Berlin, 2016 - Riccardo Tommasini - Politecnico di Milano (DEIB)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend