tira configuring executing and disseminating information
play

TIRA: Configuring, Executing, and Disseminating Information - PowerPoint PPT Presentation

TIRA: Configuring, Executing, and Disseminating Information Retrieval Experiments Tim Gollub Benno Stein Steven Burrows Dennis Hoppe Webis Group www.webis.de Bauhaus-Universitt Weimar TIRA: Configuring, Executing, and Disseminating


  1. TIRA: Configuring, Executing, and Disseminating Information Retrieval Experiments Tim Gollub Benno Stein Steven Burrows Dennis Hoppe Webis Group www.webis.de Bauhaus-Universität Weimar

  2. TIRA: Configuring, Executing, and Disseminating Information Retrieval Experiments Outline · Introduction · Architecture · Case Studies · Demonstration · Summary

  3. Introduction Quotes ❑ A longitudinal study has shown consistent selection of weak baselines in ad-hoc retrieval tasks leading to “ improvements that don’t add up ”. [Armstrong et al., 2009] ❑ A polarizing article describes how biases in research approaches lead to the consideration of “ why most published research findings are false ”. [Ioannidis, 2005] ❑ The SWIRL 2002 meeting of 45 information retrieval researchers considered evaluation as a “ perennial issue in information retrieval ” and that there is a clear need for a “ community evaluation service ”. [Allan et al., 2012] ❑ “We have to explore systematically the independent parameters of experiments.” [Fuhr, Salton Award Speech, SIGIR 2012] c 3 [ ∧ ] � www.webis.de 2012

  4. Introduction Quotes ❑ A longitudinal study has shown consistent selection of weak baselines in ad-hoc retrieval tasks leading to “ improvements that don’t add up ”. [Armstrong et al., 2009] ❑ A polarizing article describes how biases in research approaches lead to the consideration of “ why most published research findings are false ”. [Ioannidis, 2005] ❑ The SWIRL 2002 meeting of 45 information retrieval researchers considered evaluation as a “ perennial issue in information retrieval ” and that there is a clear need for a “ community evaluation service ”. [Allan et al., 2012] ❑ “We have to explore systematically the independent parameters of experiments.” [Fuhr, Salton Award Speech, SIGIR 2012] c 4 [ ∧ ] � www.webis.de 2012

  5. Introduction Quotes ❑ A longitudinal study has shown consistent selection of weak baselines in ad-hoc retrieval tasks leading to “ improvements that don’t add up ”. [Armstrong et al., 2009] ❑ A polarizing article describes how biases in research approaches lead to the consideration of “ why most published research findings are false ”. [Ioannidis, 2005] ❑ The SWIRL 2002 meeting of 45 information retrieval researchers considered evaluation as a “ perennial issue in information retrieval ” and that there is a clear need for a “ community evaluation service ”. [Allan et al., 2012] ❑ “We have to explore systematically the independent parameters of experiments.” [Fuhr, Salton Award Speech, SIGIR 2012] c 5 [ ∧ ] � www.webis.de 2012

  6. Introduction Quotes ❑ A longitudinal study has shown consistent selection of weak baselines in ad-hoc retrieval tasks leading to “ improvements that don’t add up ”. [Armstrong et al., 2009] ❑ A polarizing article describes how biases in research approaches lead to the consideration of “ why most published research findings are false ”. [Ioannidis, 2005] ❑ The SWIRL 2002 meeting of 45 information retrieval researchers considered evaluation as a “ perennial issue in information retrieval ” and that there is a clear need for a “ community evaluation service ”. [Allan et al., 2012] ❑ “We have to explore systematically the independent parameters of experiments.” [Fuhr, Salton Award Speech, SIGIR 2012] c 6 [ ∧ ] � www.webis.de 2012

  7. 7 Survey of 108 Full Papers at SIGIR 2011 Introduction [ ∧ ] Users I Query Analysis I Learning To Rank P ersonalization Retrieval Models I Social Media Content Analysis W eb IR Collaborative Filtering I Users II Query Analysis II Communities Classification Retrieval Models II Image Search Ind exing W eb Queries Collaborative Filtering II Latent Semantic Analysis Multimedia IR Summarization Ver tical Search Query Suggestions Linguistic Analysis Clustering Eff ectiveness Multilingual IR Efficiency � www.webis.de 2012 c Recommender Systems T est Collections

  8. 8 Survey of 108 Full Papers at SIGIR 2011 Introduction [ ∧ ] Provision of experiment data Users I Query Analysis I Learning To Rank P ersonalization Retrieval Models I Social Media Content Analysis W eb IR Collaborative Filtering I Users II Query Analysis II Communities Classification Retrieval Models II Image Search 51% Ind exing W eb Queries Collaborative Filtering II Latent Semantic Analysis Multimedia IR Summarization Ver tical Search Query Suggestions Linguistic Analysis Clustering Eff ectiveness Multilingual IR Efficiency � www.webis.de 2012 c Recommender Systems T est Collections

  9. 9 Survey of 108 Full Papers at SIGIR 2011 Introduction [ ∧ ] Provision of experiment software Provision of experiment data Users I Query Analysis I Learning To Rank P ersonalization Retrieval Models I Social Media Content Analysis W eb IR Collaborative Filtering I Users II Query Analysis II Communities Classification Retrieval Models II Image Search 18% 51% Ind exing W eb Queries Collaborative Filtering II Latent Semantic Analysis Multimedia IR Summarization Ver tical Search Query Suggestions Linguistic Analysis Clustering Eff ectiveness Multilingual IR Efficiency � www.webis.de 2012 c Recommender Systems T est Collections

  10. 10 Survey of 108 Full Papers at SIGIR 2011 Introduction [ ∧ ] Provision of experiment service Provision of experiment software Provision of experiment data Users I Query Analysis I Learning To Rank P ersonalization Retrieval Models I Social Media Content Analysis W eb IR Collaborative Filtering I Users II Query Analysis II Communities Classification Retrieval Models II Image Search 18% 51% 0% Ind exing W eb Queries Collaborative Filtering II Latent Semantic Analysis Multimedia IR Summarization Ver tical Search Query Suggestions Linguistic Analysis Clustering Eff ectiveness Multilingual IR Efficiency � www.webis.de 2012 c Recommender Systems T est Collections

  11. Introduction Incentives for Reproducible Research ❑ Increase acknowledgment for publishing experiments, data, and software. – Encourage a paradigm shift towards open science. ❑ Decrease the overhead of publishing experiments. – The concept of TIRA is to provide “experiments as a service”. c 11 [ ∧ ] � www.webis.de 2012

  12. Architecture Design Goals 1. Local Instantiation 1 localhost:2306/programs/examples/MyProgram?p1=42&p2=Method1&p2=Method2 ❑ Enables public research on private data. 2 ❑ Enables comparisons with private software. 3 2. Unique Resource Identifiers ❑ Enables linkage of experimental results in papers with the respective experiment service. 4 tira@node1:~$ ./myprogram.sh -p1 42 -p2 "method1" ❑ Enables reproduction of results on the 5 tira@node2:~$ ./myprogram.sh -p1 42 -p2 "method2" basis of the resource identifier (digital preservation). 6 3. Multivalued Configuration ❑ Enables the specification of whole experiment series. c 12 [ ∧ ] � www.webis.de 2012

  13. Architecture Design Goals 1. Local Instantiation 1 localhost:2306/programs/examples/MyProgram?p1=42&p2=Method1&p2=Method2 ❑ Enables public research on private data. 2 ❑ Enables comparisons with private software. 3 2. Unique Resource Identifiers ❑ Enables linkage of experimental results in papers with the respective experiment service. 4 tira@node1:~$ ./myprogram.sh -p1 42 -p2 "method1" ❑ Enables reproduction of results on the 5 tira@node2:~$ ./myprogram.sh -p1 42 -p2 "method2" basis of the resource identifier (digital preservation). 6 3. Multivalued Configuration ❑ Enables the specification of whole experiment series. c 13 [ ∧ ] � www.webis.de 2012

  14. Architecture Design Goals 1. Local Instantiation 1 localhost:2306/programs/examples/MyProgram?p1=42&p2=Method1&p2=Method2 ❑ Enables public research on private data. 2 ❑ Enables comparisons with private software. 3 2. Unique Resource Identifiers ❑ Enables linkage of experimental results in papers with the respective experiment service. 4 tira@node1:~$ ./myprogram.sh -p1 42 -p2 "method1" ❑ Enables reproduction of results on the 5 tira@node2:~$ ./myprogram.sh -p1 42 -p2 "method2" basis of the resource identifier (digital preservation). 6 3. Multivalued Configuration ❑ Enables the specification of whole experiment series. c 14 [ ∧ ] � www.webis.de 2012

  15. Architecture Design Goals (continued) 4. System Independence 1 localhost:2306/programs/examples/MyProgram?p1=42&p2=Method1&p2=Method2 ❑ Enables a widespread usage of the 2 platform. ❑ Enables the deployment of any experiment software without internal modifications. 3 5. Distributed Execution ❑ Enables efficient computation of pending experiments. 4 tira@node1:~$ ./myprogram.sh -p1 42 -p2 "method1" 5 tira@node2:~$ ./myprogram.sh -p1 42 -p2 "method2" 6. Result Storage ❑ Enables retrieval and maintenance of 6 raw experiment results. . . . and Peer to Peer Collaboration ❑ Conduct shared work on the same platform. c 15 [ ∧ ] � www.webis.de 2012

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend