INEX 2012 Overview Shlomo Geva Jaap Kamps Ralf Schenkel 10 - - PowerPoint PPT Presentation

inex 2012 overview
SMART_READER_LITE
LIVE PREVIEW

INEX 2012 Overview Shlomo Geva Jaap Kamps Ralf Schenkel 10 - - PowerPoint PPT Presentation

INEX 2012 Overview Shlomo Geva Jaap Kamps Ralf Schenkel 10 years! 2002-2012 INEX 2012 Overview Shlomo Geva Jaap Kamps Ralf Schenkel Search changed a lot in 10 years! INEX teams up with CLEF in 2012 INEX11 Workshop was on Dec 13-15,


slide-1
SLIDE 1

INEX 2012 Overview

Shlomo Geva Jaap Kamps Ralf Schenkel

slide-2
SLIDE 2

INEX 2012 Overview

Shlomo Geva Jaap Kamps Ralf Schenkel

10 years! 2002-2012

slide-3
SLIDE 3

Search changed a lot in 10 years!

slide-4
SLIDE 4

INEX teams up with CLEF in 2012

slide-5
SLIDE 5

So INEX’12 ran for only nine months... Apologies to the CLEF folks for running late’ish ... which was hard ... INEX’11 Workshop was on Dec 13-15, 2011

slide-6
SLIDE 6

Huize Bergen, Vught, The Netherlands, Dec 13-15, 2010 Snippet Retrieval Relevance Feedback Tweet Contextualization Linked Data Social Book Search

Five tracks

slide-7
SLIDE 7

Social Book Search Track

slide-8
SLIDE 8

Topic title Group name Narrative Recommended books

300 topics + recommendations from the LT forum Also crowdsourcing recommendation/relevance

slide-9
SLIDE 9

Pre- & Post Cataloguing

slide-10
SLIDE 10

SBS Task Results

Detailed results discussed in the INEX sessions

slide-11
SLIDE 11

Task 2: “Prove it” task against scanned books “Structure Extraction” task @ ICDAR

slide-12
SLIDE 12

Extensive use of crowdsourcing (topics, judgments)

slide-13
SLIDE 13

Corpus: DBpedia/YAGO + Wikipedia Investigate textual and highly structured data Linked Data Track

slide-14
SLIDE 14

Three LD Tasks

  • Ad hoc retrieval -- retrieve relevant Wikipedia

entities.

  • Faceted search -- recommend a hierarchy of facet-

values obtained from the RDF data that will

  • ptimally guide the user toward relevant Wikipedia

article in a large result set.

  • Jeopardy! -- provide answers for natural language

Jeopardy! clues which are manually translated into SPARQL queries extended with keyword conditions.

slide-15
SLIDE 15

Ad hoc: Structured helped the best run(s) Faceted: evaluation is ongoing... Jeopardy!: SPARQL effective but text better

slide-16
SLIDE 16

Task description

What International Women's Day is?

Who Francesca Woodman is?

What kind of art it is?

Where this exhibition is?

What Guggenheim Museum is?

...

On #InternationalWomensDay, we are proud to present our new Francesca Woodman exhibition opening Mar 16: http://t.co/AyuRH1OF From : Guggenheim Museum

Tweet Contextualization Track

slide-17
SLIDE 17

Task description

 Given a tweet and its metadata

  • Select a number of related passages from

Wikipedia

  • Order them so that it's a comprehensive

contextualization of the tweet

→ Multi-document summarization / answer aggregation

 Evaluation:

  • Informativeness
  • Readability
slide-18
SLIDE 18

Evaluation

 1000 tweets manual or automatically collected  Evaluation of:

  • Informativeness (63 topics) by organizers
  • Readability (16 topics) by participants

 33 submitted runs + 1 organizer baseline  13 teams (Canada, Chile, France, Germany,

India, Ireland, Mexico, Russia, Spain, USA)

slide-19
SLIDE 19

Snippet Retrieval Track

slide-20
SLIDE 20

Relevance'Assessment'

 SnippetMbased%assessment%

 Assessor%reads%through%the%20%snippets%for%each%topic,%

and%judges%each%as%relevant/irrelevant.%

 DocumentMbased%assessment%

 Each%document%is%reassessed%by%the%same%assessor%using%

the%full%document%text%providing%the%‘groundMtruth’.%

 Evaluation%is%based%on%comparing%these%two%sets%of%

judgments.%

slide-21
SLIDE 21

Timeline'

 Round%1%

 Running%(very)%Late…%

 Round%2%

 Submissions%due:%%Oct%19%  Assessment:%Oct%29%–%Nov%24%  Results%released:%Dec%3%

slide-22
SLIDE 22

Relevance Feedback Track (Open Source Retrieval workshop at SIGIR)

slide-23
SLIDE 23

+ Evaluation Platform

 Track participants provided with the complete document

collection in advance (2,666,192 docs INEX Wikipedia 2009)

 Evaluation platform provides the relevance feedback modules with

topics

 Simulates a user in the loop, interacting with the search system

Evalua&on) Pla+orm)

Document) Collec,on) Assessments Relevance) Feedback) Algorithm)

slide-24
SLIDE 24

2012'(preliminary)'results'

0.1 0.2 0.3 0.4 0.5 0.6 0.7 @5 @10 @15 @20 @30 @100 @200 @500 @1000

Exact Precision - Best non-RF and best RF submissions from each participant

BASE-IND RRMRF-300D-L05 TOPSIG-2048 TOPSIG-RF4

slide-25
SLIDE 25

1

Geva et al. (Eds.)

Comparative Evaluation

  • f Focused Retrieval

LNCS 6932

Shlomo Geva Jaap Kamps Ralf Schenkel Andrew Trotman (Eds.)

123

LNCS 6932

9th International Workshop of the Inititative for the Evaluation of XML Retrieval, INEX 2010 Vugh, The Netherlands, December 2010 Revised Selected Papers

Comparative Evaluation

  • f Focused Retrieval

INEX 2010

ISSN 0302-9743

› springer.com

Lecture Notes in Computer Science

The LNCS series reports state-of-the-art results in computer science re search, development, and education, at a high level and in both printed and electronic form. Enjoying tight cooperation with the R&D community, with numerous individuals, as well as with prestigious organizations and societies, LNCS has grown into the most comprehensive computer science research forum available. The scope of LNCS, including its subseries LNAI and LNBI, spans the whole range of computer science and information technology including interdisciplinary topics in a variety of application fields. The type of material published traditionally includes – proceedings (published in time for the respective conference) – post-proceedings (consisting of thoroughly revised final full papers) – research monographs (which may be based on outstanding PhD work, research projects, technical reports, etc.) More recently,several color-cover sublines have beenadded featuring, beyond a collection of papers, various added-value components; these sublines in clude – tutorials (textbook-like monographs or collections of lectures given at advanced courses) – state-of-the-art surveys (offering complete and mediated coverage
  • f a topic)
– hot topics (introducing emergent topics to the broader community) In parallel to the printed book, each new volume is published electronically in LNCS Online. Detailed information on LNCS can be found at www.springer.com/lncs Proposals for publication should be sent to LNCS Editorial, Tiergartenstr. 17, 69121 Heidelberg, Germany E-mail: lncs@springer.com ISBN 978-3-642-23576-4 9 7 8 3 6 4 2 2 3 5 7 6 4

INEX’11 LNCS in Sep/Oct; 2012 coming as well.

slide-26
SLIDE 26

Plans for INEX 2013 are under discussion