Assessing the scholarly impact of ImageCLEF Theodora Tsikrika Alba - - PowerPoint PPT Presentation
Assessing the scholarly impact of ImageCLEF Theodora Tsikrika Alba - - PowerPoint PPT Presentation
Assessing the scholarly impact of ImageCLEF Theodora Tsikrika Alba Garca Seco de Herrera Henning Mller University of Applied Sciences Western Switzerland (HES-SO) Sierre, Switzerland CLEF 2011, Sept 21, 2011 1 Evaluation campaigns
SLIDE 1
SLIDE 2
CLEF 2011, Sept 21, 2011 2
Evaluation campaigns
Enable the reproducible and comparative evaluation through the use
- f standardised resources and common evaluation methodologies
- Benefits:
- Provide access to evaluation infrastructure
- Build reusable resources for benchmarking
- Promote exchange of ideas
- Encourage collaboration and interaction
- Resulting in:
- Development of new approaches
- Increased quality of evaluation methodologies
- Advancement of the field
SLIDE 3
CLEF 2011, Sept 21, 2011 3
Evaluation campaigns: assessing impact?
Assess the impact of the research they foster
- research that would otherwise not have been possible
- Scientific impact:
- Scholarly impact, filed patents, …
- Economic impact:
- Technology transfer, time & effort saved for researchers, …
SLIDE 4
CLEF 2011, Sept 21, 2011 4
Evaluation campaigns: impact assessment
- B. R. Rowe, D. W. Wood, A. N. Link, and D. A. Simoni.
Economic impact assessment of NIST’s Text REtrieval Conference (TREC) Program. Technical Report, Project Number 0211875, RTI International, 2010.
- C. V. Thornley, A. C. Johnson, A. F. Smeaton, and H. Lee.
The scholarly impact of TRECVid (2003–2009). JASIST, 62(4):613–627, 2011.
SLIDE 5
CLEF 2011, Sept 21, 2011 5
Objectives
- Assess the scholarly impact of ImageCLEF
- Compare different methods in performing such an assessment
SLIDE 6
CLEF 2011, Sept 21, 2011 6
ImageCLEF: cross-language image retrieval
6
SLIDE 7
Assessing the scholarly impact
- Perform bibliometric analysis
- Obtain derived publications and their citations
- Calculate metrics (e.g., h-index)
- Sources for publication and citation data
- Thomson Reuters (ISI) Web of Science
- Scopus
- Google Scholar
7
SLIDE 8
CLEF 2011, Sept 21, 2011 8
Comparison of data sources and tools
- Thomson Reuters (ISI) Web of Science
+ Coverage of more than 10,000 journals − Very limited coverage of conference proceedings + High-quality citation data + Provides metrics
- Scopus
+ Coverage of more than 18,000 titles + Includes many conference proceedings − Citation coverage only after 1996 + High-quality citation data + Provides metrics
- Google Scholar
+ Wider coverage that includes additional journals and conference proceedings, plus books, technical reports, white papers, etc. − Errors in citation data − Does not provide metrics + … but Publish or Perish (PoP) does
8
SLIDE 9
CLEF 2011, Sept 21, 2011 9
Comparison of data sources and tools
- Thomson Reuters (ISI) Web of Science
+ Coverage of more than 10,000 journals − Very limited coverage of conference proceedings + High-quality citation data + Provides metrics
- Scopus
+ Coverage of more than 18,000 titles + Includes many conference proceedings − Citation coverage only after 1996 + High-quality citation data + Provides metrics
- Google Scholar
+ Wider coverage that includes additional journals and conference proceedings, plus books, technical reports, white papers, etc. − Errors in citation data − Does not provide metrics + … but Publish or Perish (PoP) does
9
SLIDE 10
Google Scholar shortcomings
- Several entries for the same publication
- Due to mispellings
- May deflate citation count
- PoP allows manual merging of publications
- Grouping citations of different papers
- With similar titles and author lists
- E.g., conference and journal versions of a paper
- Manual data cleaning
- Inability to correctly identify publication year
10
SLIDE 11
Dataset of ImageCLEF publications
- ImageCLEF papers in CLEF working notes
- not indexed by Scopus
- ImageCLEF papers in CLEF proceedings
- published in the year following the workshop
- Papers describing ImageCLEF resources published elsewhere
- written by ImageCLEF organisers
- Papers using ImageCLEF resources pubished elsewhere
- written by groups that have participated or just registered without
submitting runs or acquired the data at a later stage
11
SLIDE 12
Dataset of ImageCLEF publications
- ImageCLEF papers in CLEF working notes
- not indexed by Scopus
- ImageCLEF papers in CLEF proceedings
- published in the year following the workshop
- Papers describing ImageCLEF resources published elsewhere
- written by ImageCLEF organisers
- Papers using ImageCLEF resources pubished elsewhere
- written by groups that have participated or just registered without
submitting runs or acquired the data at a later stage
12
SLIDE 13
ImageCLEF publications and citations
13
- ~ 70% of citations are from papers not in CLEF proceedings
- 8.62 cites per paper on average
SLIDE 14
CLEF 2011, Sept 21, 2011 14
Overview vs. Participants’ papers
CLEF proceedings All
(CLEF proceedings + ImageCLEF resources)
90% of the papers that have # citations ≥ h-index are overviews
SLIDE 15
CLEF 2011, Sept 21, 2011 15
The 2005 overview paper
Single overview for both medical and general image tasks Half life of approximately three years
SLIDE 16
General vs. Medical image annotation and retrieval
16
Publications in medical domain have slightly higher impact
SLIDE 17
CLEF 2011, Sept 21, 2011 17
General vs. Medical image annotation and retrieval
17 Scopus PoP publications citations
- 2006-2008 publications in medical domain have high impact
SLIDE 18
CLEF 2011, Sept 21, 2011 18
Citations per task
18
- Peak in second or third
year of operation
- Followed by a decline
- … unless there is a major
- verhaul of the task
- Tasks with greatest impact
– Photographic retrieval – Medical image retrieval – Medical annotation
SLIDE 19
CLEF 2011, Sept 21, 2011 19
Conclusions and Future Work
- Preliminary analysis shows important impact of ImageCLEF
- 249 publications
- 303 citations in Scopus
- 2,147 in Google Scholar
- Scopus vs. Google Scholar
- Both have advantages and limitations
- Next steps:
- Automate process (as much as possible)
- Include working notes (~ 500 papers in total)
- Include ImageCLEF derived papers (~1,000 papers in total)
- Perform analysis for the whole CLEF