evaluation of query rewriting approaches for owl 2
play

Evaluation of Query Rewriting Approaches for OWL 2 Hector - PowerPoint PPT Presentation

Evaluation of Query Rewriting Approaches for OWL 2 Hector Perez-Urbina, Edgar Rodriguez-Diaz, Michael Grove, George Konstantinidis, and Evren Sirin { hector, edgar, mike, evren}@clarkparsia.com SSWS+HPCSW 2012 Clark & Parsia, LLC


  1. Evaluation of Query Rewriting Approaches for OWL 2 Hector Perez-Urbina, Edgar Rodriguez-Diaz, Michael Grove, George Konstantinidis, and Evren Sirin { hector, edgar, mike, evren}@clarkparsia.com SSWS+HPCSW 2012 Clark & Parsia, LLC

  2. ○ Semantic Research & Development firm since 2005 ○ Washington, DC ○ Cambridge, MA ○ Leading supplier of innovative semantic technologies ○ Semantic Web ○ Web Services ○ Advanced AI technologies ○ Committed to development and adoption of standards ○ Participated in several W3C working groups ○ Organized OWLED workshops to foster relationships between researchers and industry

  3. Query Rewriting TBox Conjunctive Query Query Rewriting Algorithm Rewritings

  4. Why? ○ Independent of ABox in terms of size and computation time ○ No need to update the rewritings every time the ABox changes ○ Evaluation of rewritings can be delegated ○ The same rewritings can be evaluated over different instances of data without having to be recomputed

  5. Not a Silver Bullet ○ Rewritings might get too big/complex to evaluate efficiently ( EXP even for QL) ○ Rewritings might need to be recursive or disjunctive DQs (1) for more expressive fragments of OWL ○ Expressive ontologies might generate rewritings that are more difficult to evaluate (1) Datalog Queries

  6. State of the Art

  7. Comparison

  8. Evaluation Questions ○ Can UCQs (1) be evaluated more efficiently (faster) than DQs? ○ Can Requiem be made faster by enhancing it with various optimizations (like some of the approaches in other algorithms)? (1) Union of Conjunctive Queries

  9. Blackout ○ Stardog component ○ Highly optimized version of Requiem ○ Eager query containment ○ Data oracle

  10. Empirical Evaluation ○ Empirical evaluation of Blackout (Stardog 1.0.3) and Clipper ○ Support the most expressive logics ○ Two different approaches (UCQs vs DQ) ○ Using LUBM benchmark ○ 14 queries ○ 3 different TBoxes (for QL, EL, and RL) ○ 4 different ABoxes of increasing size ○ Very simple hardware ○ 8 GB RAM ○ AMD 3.2 Ghz - 4 cores ○ First, examine performance of the rewritings computation (size/complexity, time) ○ Second, examine how fast the rewritings can be evaluated using two database systems (DLV, Stardog)

  11. Rewriting Computation Results

  12. Rewriting Evaluation Results

  13. Overall Blackout/Stardog Performance

  14. Conclusions ○ DQ-producing approaches do not necessarily produce smaller rewritings than UCQ-producing approaches ○ Some optimizations, such as the data oracle, should apply to DQs as well ○ It is not necessary to produce DQs for RL and EL in many cases ○ UCQs are amenable to straightforward parallel evaluation ○ Evaluating the rewritings dominates the overall query answering time at large scales ○ It is worth investing time on optimizing the computation of rewritings that can be evaluated more efficiently

  15. Future Work ○ Add datalog evaluation to Stardog based on Query-SubQuery algorithm ○ Parallelization of UCQ evaluation within Stardog ○ Explore the performance implications of new SPARQL 1.1 features (sub-queries, transitivity, property paths) in rewritings ○ Determine if Blackout can be used to handle queries utilizing these SPARQL 1.1 features

  16. Questions?

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend