IR in Context of the User: Interactive IR Evaluation Peter - - PowerPoint PPT Presentation

ir in context of the user interactive ir evaluation
SMART_READER_LITE
LIVE PREVIEW

IR in Context of the User: Interactive IR Evaluation Peter - - PowerPoint PPT Presentation

IR in Context of the User: Interactive IR Evaluation Peter Ingwersen Royal School of LIS Denmark pi@iva.dk http://www.iva.dk/pi Oslo University College, Norway Essir2011 Ingwersen 1 Agenda - 1 Introduction (20 min) Research


slide-1
SLIDE 1

IR in Context of the User: Interactive IR Evaluation

Peter Ingwersen Royal School of LIS Denmark pi@iva.dk – http://www.iva.dk/pi Oslo University College, Norway

Essir2011 1 Ingwersen

slide-2
SLIDE 2

Essir2011 2 Ingwersen

Agenda - 1

 Introduction (20 min)

 Research Frameworks vs. Models  Central components of Interactive IR (IIR)  The Integrated Cognitive Research Framework for IR

 From Simulation to ‘Ultra-light’ IIR (20 min)

 Short-term IR interaction experiments  Sample study – Diane Kelly (2005/2007)

slide-3
SLIDE 3

Essir2011 3 Ingwersen

Agenda - 2

 Experimental Research Designs with Test

persons (25 min)

 Interactive-light session-based IR studies  Request types  Test persons  Design of task-based simulated search situations  Relevance and evaluation measures in IIR  Sample study – Pia Borlund (2000; 2003b)

slide-4
SLIDE 4

Essir2011 4 Ingwersen

Agenda - 3

 Naturalistic Field Investigations of IIR (20 min)

 Integrating context variables  Live systems & (simulated) work tasks  Sample Study – Marianne Lykke (Nielsen) (2001; 2004)

 Wrapping up (5 min)

Questions are welcome during the sessions

Advanced Information Retrieval / Ricardo Baetza Yeates & Massimo Melucci (eds.). Springer, 2011, p. 91-118.

slide-5
SLIDE 5

Essir2011 5 Ingwersen

Frameworks & Models – difference?

 Frameworks describe

 Essential objects to study  The relationships of objects  The changes in the objects /

relationships that affect the functioning of the system

 Promising goals and methods

  • f research

 Frameworks contain (tacit)

shared assumptions

 ontological, conceptual, factual,

epistemological, and methodological

 The concept model

 A precise (often formal)

representation of objects and relationships (or processes) within a framework

 Modeling may also in principle

encompass human actors and

  • rganizations

 Frameworks may lead to

 Research Designs, incl.  Research Questions;

Experimental Setting; Methodology

slide-6
SLIDE 6

The Lab. Research Framework – cave with central variables (The Turn, 2005)

Docu- ments Represen- tation Database Search request Query Matching Represen- tation Query Result Evaluation Result Evaluation Relevance assessment Recall base

Pseudo RF

Essir2009 Ingwersen

7

Essir2011 6 Ingwersen

slide-7
SLIDE 7

User-centered (contextual) MODELS

 Examples (in display order)

 Wilson, 1999 (conceptual: Info. Behavior; Seek; IR)  Byström & Järvelin, 1995 (flow chart: Info. Seek)  Saracevic, 1996 (conceptual, stratified: IR)  Vakkari, 2000 (flow chart, Online Search;

Relevance)

 Wang & Soergel, 1998 (conceptual: Relevance

Assessment Process & Criteria)

Essir2011 7 Ingwersen

slide-8
SLIDE 8

2011 Peter Ingwersen 8

Information behaviour and IR

  • T. Wilson´s Onion Model, 1999 - extended:

Seeking IR

Job-related Work Tasks Interests Non-job-related Tasks and Interests Daily-life behavior

Information behaviour Interactive IR Behaviour

slide-9
SLIDE 9

2011 Peter Ingwersen 9

IS&R model, 1995: Bystöm &

Järvelin, fig. 2

Perceived Task Personal Seeking style Organization Personal Factors Situational Factors Information Need analysis Choice of Action

  • identification of

alternatives

  • ranking them
  • choosing an action

Implementation Evaluation

a) needs satisfied, task may be completed b) needs cannot be satisfied c) further information is needed

Perceived Task Personal Seeking style Organization Personal Factors Situational Factors Information Need analysis Choice of Action

  • identification of

alternatives

  • ranking them
  • choosing an action

Implementation Evaluation

a) needs satisfied, task may be completed b) needs cannot be satisfied c) further information is needed (From: The Turn, p. 69)

slide-10
SLIDE 10

2011 Peter Ingwersen 10

Saracevic´ stratified model for IIR (1996)

slide-11
SLIDE 11

2011 Peter Ingwersen 11

Wang & Soergel 1998

Type Abstract Document Values Criteria DIEs Decision processing combining deciding Knowledge of topic person

  • rganization

journal document type Decision Rules Elimination Multiple criteria Dominance Scarcity Satisfice Chain Author Title Orientation Topicality Date Series Journal Relation Authority Availability Novelty Quality Emotional Social Conditional Functional Epistemic Rejection Maybe Acceptance DIEs: Document Information Elements Values: Document Values/Worth (From: The Turn, p. 201)

slide-12
SLIDE 12

2011 Peter Ingwersen 12

IR and relevance in Seeking context – Seeking into IS&R: Vakkari 2000

slide-13
SLIDE 13

Essir2011 13 Ingwersen

Research Setting Types

 Laboratory experiments – no test persons, but

 Simulations – Log analyses (not treated in presentation)

 Laboratory study – with test persons:  ‘Ultra-light’ (short-term interaction: 1-2 retrieval runs)

– or ‘Interactive light’ (session-based multi-run interaction)

 Field experiment – experimental (artificial) situation

in natural setting with test persons

 Field study – study of natural performance or

behavior in natural setting with test persons

 Longitudinal studies

 Case study – (qualitative) study with few test persons

slide-14
SLIDE 14

Essir2011 14 Ingwersen

Variables involved in a test:

 Independent (the ‘cause’), e.g.,

Interface functionalities; Different IR models; Searcher knowledge

 Dependent (the ‘effect’), e.g.,

 Performance measures of output (recall/prec.; CumGain; usability)

 Controlled (held constant; statistically neutralized; randomized):

 Database; Information objects  Search algorithms  Simulated work task situations – Assigned TREC topics  Test persons

 Hidden variables (Moderating or Intervening), e.g.,

 Variation of test persons’ levels of experience …!!! – see the Integrated

Research Framework for IR

slide-15
SLIDE 15

Essir2009 15

Agenda - 1

 Introduction to Tutorial (20 min)

 Research Frameworks vs. Models  Central components of Interactive IR (IIR)  The Integrated Cognitive Research Framework for IR

 From Simulation to ‘Ultra-light’ IIR (20 min)

 Short-term IR interaction experiments  Sample study – Diane Kelly (2005/2007)

Essir2011 15 Ingwersen

slide-16
SLIDE 16

Essir2011 16 Ingwersen

Central Components of Interactive IR – the basic integrated framework

Information

  • bjects

IT: Engines Logics Algorithms Interface Cognitive Actor(s)

(team)

Org. Cultural Social

Context

Information

  • bjects

IT: Engines Logics Algorithms Interface Cognitive Actor(s)

(team)

Org. Cultural Social

Context

slide-17
SLIDE 17

Central Components of Interactive IR – the basis

  • f the integrated framework

Information

  • bjects

IT: Engines Logics Algorithms Interface Cognitive Actor(s)

(team)

Org. Cultural

= Cognitive transformation and influence = Interactive communication of cognitive structures = Cognitive transformation and influence over time

Social

Context

The Lab. Framework

Essir2011 17 Ingwersen

slide-18
SLIDE 18

Essir2009 18 Ingwersen

Dimensions and Range of Variables in the Integrated IIR framework: 9 dimensions from 6 components

Information

  • bjects

IT: Engines Logics Algorithms Interface Cognitive Actor(s)

(team)

Org. Cultural Social

Context

Information

  • bjects

IT: Engines Logics Algorithms Interface Cognitive Actor(s)

(team)

Org. Cultural Social

Context

Interaction

Essir2011 18 Ingwersen

slide-19
SLIDE 19

Essir2009 19 Ingwersen

Categories of Dimensions i the Cognitive Research Framework

1.

Natural work task dimension

2.

Natural search task dimension

3.

Actor characteristics dimension

4.

Perceived work task dimension

5.

Perceived search task

6.

Document dimension

7.

Algorithmic search engine dimension

8.

Algorithmic interface dimension

9.

Access and interaction dimension

Each containing multiple variables

Socio-org. task dimensions Actor dimensions Algorithmic dimensions

Essir2011 19 Ingwersen

slide-20
SLIDE 20

Essir2009 20 Ingwersen

Natural Work Tasks (WT) & Org Natural Search Tasks (ST) Actor Perceived Work Tasks Perceived Search Tasks

WT Structure ST Structure Domain Knowledge Perceived WT Structure Perceived Information Need Content WT Strategies & Practices ST Strategies & Practices IS&R Knowledge Perceived WT Strategies & Practices Perceived ST Structure/Type WT Granularity, Size & Complexity ST Granularity, Size & Complexity Experience on Work Task Perceived WT Granularity, Size & Complexity Perceived ST Strategies & Practices WT Dependencies ST Dependencies Experience on Search Task Perceived WT Dependencies Perceived ST Specificity & Complexity WT Requirements ST Requirements Stage in Work Task Execution Perceived WT Requirements Perceived ST Dependencies WT Domain & Context ST Domain & Context Perception of Socio-Org. Context Perceived WT Domain & Context Perceived ST Stability Sources of Difficulty Perceived ST Domain & Context Motivation & Emotional State

Variables with values

Essir2011 20 Ingwersen

slide-21
SLIDE 21

Essir2009 21 Ingwersen

Natural Work Tasks (WT) & Org Natural Search Tasks (ST) Actor Perceived Work Tasks Perceived Search Tasks

WT Structure ST Structure Domain Knowledge Perceived WT Structure Perceived Information Need Content WT Strategies & Practices ST Strategies & Practices IS&R Knowledge Perceived WT Strategies & Practices Perceived ST Structure/Type WT Granularity, Size & Complexity ST Granularity, Size & Complexity Experience on Work Task Perceived WT Granularity, Size & Complexity Perceived ST Strategies & Practices WT Dependencies ST Dependencies Experience on Search Task Perceived WT Dependencies Perceived ST Specificity & Complexity WT Requirements ST Requirements Stage in Work Task Execution Perceived WT Requirements Perceived ST Dependencies WT Domain & Context ST Domain & Context Perception of Socio-Org. Context Perceived WT Domain & Context Perceived ST Stability Sources of Difficulty Perceived ST Domain & Context Motivation & Emotional State

Essir2011 21 Ingwersen

slide-22
SLIDE 22

Essir2009 22 Ingwersen

Document and Source IR Engines IT Component IR Inter-faces Access and Interaction

Document Structure Exact Match Models Domain Model Attributes Interaction Duration Document Types Best Match Models System Model Features Actors or Components Document Genres Degree of Doc. Structure and Content Used User Model Features Kind of Interaction and Access Information Type in Document Use of NLP to Document Indexing System Model Adaption Strategies and Tactics Communication Function

  • Doc. Metadata

Representation User Model Building Purpose of Human Communication Temporal Aspects Use of Weights in Doc. indexing Request Model Builder Purpose of System Communication Document Sign Language Degree of Req. Structure and Content Used Retrieval Strategy Interaction Mode Layout and Style Use of NLP to Request Indexing Response Generation Least effort Factors Document Isness

  • Req. Metadata

Representation Feedback Generation

  • Document Content

Use of Weights in Requests Mapping ST History Contextual Hyperlink Structure Explanation Features Human Source (see Actor) Transformation of Messages Scheduler Essir2011 22 Ingwersen

slide-23
SLIDE 23

Essir2009 23 Ingwersen

Number of variables using the Framework

Maximum application of three independent variables simultaneously!

Can be done in pairs – and by total control of binary values of variables, e.g.

1.

Interface function X, value a/b

2.

Personal IR expertise, values none/much

3.

In domain Z, work task type: routine – but

Rich/Poorly defined There are many relevant combinations made from the Framework!

Essir2011 23 Ingwersen

slide-24
SLIDE 24

Essir2009 24

Agenda - 1

 Introduction to Tutorial (20 min)

 Research Frameworks vs. Models  Central components of Interactive IR (IIR)  The Integrated Cognitive Research Framework for IR

 From Simulation to ‘Ultra-light’ IIR (20 min)

 Short-term IR interaction experiments  Sample study – Diane Kelly (2005/2007)

Essir2011 24 Ingwersen

slide-25
SLIDE 25

IR interaction ‘Ultra-light’ – short-term IIR

Docu- ments Represen- tation Database Search request Query Matching Represen- tation Query Result Evaluation Result Evaluation Relevance assessment Recall base

Max. ONE Relevance Feedback Run Allowed

Essir2011 25 Ingwersen

slide-26
SLIDE 26

Essir2011 26 Ingwersen

Lab IR -´ultra light´ interaction

In this 1-2 run setting we have two ways of measuring performance:

1.

By assessor in pre-existing test collection

(as in TREC with unrealistically long delay between run and assessment – but equal to all).

Assessments may be applied to judging second run results (made by the test persons) – like using pseudo RF after first run

slide-27
SLIDE 27

Essir2011 27 Ingwersen

Lab IR -´ultra-light´ interaction

  • 2. By all test searchers of the first run results of the same query

session (Secondary run assessments of results used if first run done by pseudo RF).

One may pool performance scores across the set of assigned requests (topics) or simulated tasks given in experiment – because of the max. two runs:

Good: No learning effects can influence the experiment

  • or the same effects appear as when using TREC assessors

That is why this setting is ´interactive ultra-light´

Graded relevance assessments possible

Can be used OUTSIDE traditional test collections!

Bad: Quite few documents are commonly assessed for relevance per test searcher (or vary much)!

The setting is limited in realism (only 2 runs)

slide-28
SLIDE 28

Essir2011 28 Ingwersen

Interactive ‘Ultra-light’ experiment. Research question concerned with variables from actor & IT

Information

  • bjects

IT

Pseudo RF

Interface Social Cultural Task Org.

Context

Work task perception Search task perception Actor need

slide-29
SLIDE 29

Essir2011 29 Ingwersen

IIR Interactive ‘Ultra-light’ sample

 Kelly, D., Dollu, V.D. & Xin Fu (2005). The loquacious user:

A document-independent source of terms for query expansion. In: Proceedings of the 28th Annual ACM-SIGIR Conference on Research and Development in Information retrieval: 457-464. (also as IP&M article in 2007, 43(1): 30-46 ) – extended in IPM, 2007.

 RQ: Does multi-evidence of users’ information

need situation improve retrieval performance through query expansion, compared to initial request and pseudo relevance feedback?

slide-30
SLIDE 30

Essir2011 30 Ingwersen

Research setting

13 test persons supplied …

45 natural ‘topics’ to HARD TREC (title and description) and

Relevance assessments

HARD TREC collection; Lemur system (BM25)

1.

Topic title+description run by Lemur (bag-of- words) one run; serves as baseline (BL).

2.

Pseudo RF modes (top-5; top-10;…) run on top

  • f BL

3.

Each test person asked 4 questions via a form:

slide-31
SLIDE 31

Essir2011 31 Ingwersen

Research setting 2

 (Q1) state the times in the past he/she had

searched that topic;

 (Q2) describe what he/she already knows about

the topic (knowledge state);

 (Q3) state why he/she wants to know about the

topic; and

 (Q4) add any keywords that further describe the

topic.

slide-32
SLIDE 32

Essir2011 32 Ingwersen

Research setting 3

Controlled variables: BM 25; 45 topics; HARD coll.

Independent variables:

1.

Pseudo RF variations – on top of baseline (BL)

2.

Q2-Q4 combinations (term weights) – on top of BL

Dependent var.: MAP – statistical significance test

RESULTS, yield of different words (mean):

Baseline : 9.33 – Q2: 16.18 – Q3: 10.67 – Q4: 3.3

Single Q-forms outperform BL

Q2-Q3 (and Q2-Q4) combined outperform BL plus any pseudo RF

Performance increases with query length.

slide-33
SLIDE 33

Essir2011 33 Ingwersen

Summary: IIR ‘Ultra-Light’

 Strength:

 Easy to apply existing test collections, with …  Relevance assessments existing a priori (as in TREC or INEX)  New relevance assessments possible – with graded assessments

and over many assessors (test persons): weighted assessments

 Can lead to more solid interactive investigations in later studies

 Weakness:

 Are all variable values known?? (people means hidden ones!)  ‘Ultra-light’ IIR is limited in realism (1-2 iterations; context

features hardly in play)

 Limited number of documents assessed (per test person)

slide-34
SLIDE 34

Essir2009 34

Agenda - 1

 Introduction to Tutorial (20 min)

 Research Frameworks vs. Models  Central components of Interactive IR (IIR)  The Integrated Cognitive Research Framework for IR

 From Simulation to ‘Ultra-light’ IIR (20 min)

 Short-term IR interaction experiments  Sample study – Diane Kelly (2005/2007)

Ingwersen

Essir2011 34 Ingwersen

slide-35
SLIDE 35

Essir2011 35 Ingwersen

Agenda - 2

 Experimental Set-ups with Test Persons (25

min)

 Interactive-light session-based IR studies  Request types  Test persons  Design of task-based simulated search situations  Relevance and evaluation measures in IIR  Sample study – Pia Borlund (2000; 2003b)

slide-36
SLIDE 36

IR interaction ‘Light’

Docu- ments Represen- tation Database Search request Query Matching Represen- tation Query Result Evaluation Result Evaluation Relevance assessment Recall base

MANY Relevance Feedback Runs Allowed Searcher MUST DO Posteriori Relevance Assessments

Context

Essir2011 36 Ingwersen

slide-37
SLIDE 37

Essir2011 37 Ingwersen

Data Collection Means

 Observation  Thinking (talking) aloud - Introspection  Eye-tracking  Critical incidence  Questionnaires  Interviews (structured; open-ended; closed)

 Post or/and pre-interviews

 Focus groups  Diaries – Self reporting  Logging and recording of behavior (system/client logs)

 Assessments of relevance

slide-38
SLIDE 38

Essir2011 38 Ingwersen

Request Types in (‘ultra’) ‘light’ IIR

 Natural request/ real

need of test person - or

 Assigned to test person  ‘Query’ is the retrieval

mechanism’s internal translation of the REQUEST

 Topical (as TREC ‘topics)  Factual  ‘Known Item’  Other metadata  Simulated Task Situation

(Cover Story)

 ‘Sample’ as request  Simplistic request formulation

(context free)

slide-39
SLIDE 39

Essir2011 39 Ingwersen

Number of Test Persons

 Number depends on goal of research & no. of

variables:

 Behavioral field study/experiment: many persons

(>30 test persons required – and some (2-3) search jobs, to be statistically valid)

 Performance-like field experiment: many search

jobs per person (4-10) – but less (~ 15) test persons required.

 Note: Sanderson et al. paper: IIIX 2005 on no. of topics

necessary for statistical validity: > 60!! (if applying MAP on top-15)

 The best design: always > 25 persons

slide-40
SLIDE 40

Essir2011 40 Ingwersen

Test Persons …

 In order to be statistically significant, or really

indicative, each cell in the cross tabulation result matrix should contain 25-30 units (rule of thump).

 Example (performance/evaluation goal with 3

independent (binary) variables, done in pairs: 2x2x2 = 8 cells x 30 units = 240 units in total):

 You have 2 x 10 test persons (doctors & med. stud.)  They need to do 12 search jobs per person = 120

units per group over 2x2 additional variable values, for reasons of cross tabulations = 120 x 2 = 240 jobs!

 or 2 x 20 persons doing 6 jobs each.

slide-41
SLIDE 41

Essir2011 41 Ingwersen

Latin Square research design –

The Turn, p. 253-254

system X system Y 1: A, B, C 4: D, E, F 2: C, B, A 5: F, E, D 3: C, A, B 6: F, D, E 1: D, F, E 4: A, C, B 2: E, F, D 5: B, C, A 3: E, D, F 6: B, A, C system X system Y 1: A, B, C 4: D, E, F 2: C, B, A 5: F, E, D 3: C, A, B 6: F, D, E 1: D, F, E 4: A, C, B 2: E, F, D 5: B, C, A 3: E, D, F 6: B, A, C

6 test persons (1-6); 6 real / simulated work tasks/ or assigned topics (A-F)

slide-42
SLIDE 42

Essir2011 42 Ingwersen

Agenda - 2

 Experimental Set-ups with Test Persons (25

min)

 Interactive-light session-based IR studies  Request types  Test persons  Design of task-based simulated search situations  Relevance and evaluation measures in IIR  Sample study – Pia Borlund (2000; 2003a)

slide-43
SLIDE 43

Essir2011 43 Ingwersen

Simulated Work Task Situations

– or ’cover stories’ – to trigger natural information needs (& requests)

Example from study on relevance assessments on the Web (See ‘Ultra-Light’ in Bibliography: Papaeconomou, 2008): Beijing is hosting in 2008 (8th-24th August) the Olympic Games. A friend of yours, who is a big fan of the Olympic Games, wants to attend the events and asks you to join in this trip. You find this invitation interesting. You are not a big fan of the games but you always wanted to visit China, therefore you want to find information about the sightseeing in the city and the activities that the Chinese will offer during the games. Find for instance places you could visit, activities you could do in relation to the Chinese culture or in the spirit of the games.

slide-44
SLIDE 44

Essir2011 44 Ingwersen

Borlund IIR (2003b) evaluation package

Simulated situation: sim A Simulated work task situation: After your graduation you will be looking for a job in industry. You want information to help you focus your future job

  • seeking. You know it pays to know the market. You would like to find some

information about employment patterns in industry and what kind of qualifications employers will be looking for from future employees. Indicative request: Find for instance something about future employment trends in industry, i.e. areas of growth and decline.

slide-45
SLIDE 45

Essir2011 45 Ingwersen

Integrated Framework and Relevance Criteria

Docs Repr DB Request Query Match Repr Result

A: Recall, precision, efficiency, (quality of information/process) B: Usability, Graded rel., CumGain; Quality of information/process C: Quality of info & work task result; Graded R.

Work Task Seeking Task Seeking Process Work Process Task Result Seeking Result

Evaluation Criteria:

Work task context Seeking context IR context Socio-organizational& cultural context

D: Socio-cognitive relevance; Social

utility: rating; citations; inlinks;

slide-46
SLIDE 46

Essir2011 46 Ingwersen

Relevance & Evaluation Measures

The measurement of performance by use of non-binary (graded, scaled or gliding) based performance measures (generalized by Järvelin & Kekäläinen, 2002)

 Realistic assessment behaviour  Indication of users’ subjective impression of system performance

and satisfaction of information need: usability (Hornbæk, 2006;

Nielsen, 2006)  Other measurements to be used on Interaction Process:

 Display time; No. of requests/queries; Visits & Downlods  Selection patterns; Views & clicks; Social utility assessments;  No. of documents assessed; Perceived ease of process; …

slide-47
SLIDE 47

Essir2011 47 Ingwersen

Agenda - 2

 Experimental Set-ups with Test Persons (40

min)

 Interactive-light session-based IR studies  Request types  Test persons  Design of task-based simulated search situations  Relevance and evaluation measures in IIR  Sample study – Pia Borlund (2000; 2003a)

slide-48
SLIDE 48

Essir2011 48 Ingwersen

The Borlund Case (2000; 2003b)

Research questions 1) Can simulated information needs substitute real information needs?

Hypothesis is: YES!

2) What makes a ‘good’ simulated situation with reference to semantic openness and types

  • f topics of the simulated situations?
slide-49
SLIDE 49

Essir2011 49 Ingwersen

Data collection: Financial Times (TREC data) and The Herald (current) Test system: Full-text online system Probabilistic based retrieval engine Test persons: 24 university students (undergraduates and graduates) From: Computing, engineering, psychology, geography, English history, etc.

  • Info. needs:

24 real needs (1 real need per test person) 96 simulated information needs (4 simulated task situations per test person)

Location of tests: IR Laboratory at Glasgow University

Experimental Setting:

slide-50
SLIDE 50

Essir2011 50 Ingwersen

Natural Work Tasks (WT) & Org Natural Search Tasks (ST) Actor Perceived Work Tasks Perceived Search Tasks

WT Structure ST Structure Domain Knowledge Perceived WT Structure Perceived Information Need Content WT Strategies & Practices ST Strategies & Practices IS&R Knowledge Perceived WT Strategies & Practices Perceived ST Structure/Type WT Granularity, Size & Complexity ST Granularity, Size & Complexity Experience on Work Task Perceived WT Granularity, Size & Complexity Perceived ST Strategies & Practices WT Dependencies ST Dependencies Experience on Search Task Perceived WT Dependencies Perceived ST Specificity & Complexity WT Requirements ST Requirements Stage in Work Task Execution Perceived WT Requirements Perceived ST Dependencies WT Domain & Context ST Domain & Context Perception of Socio-Org. Context Perceived WT Domain & Context Perceived ST Stability Sources of Difficulty Perceived ST Domain & Context Motivation & Emotional State

Independent Variables

slide-51
SLIDE 51

Essir2011 51 Ingwersen

Document and Source IR Engines IT Component IR Inter-faces Access and Interaction

Document Structure Exact Match Models Domain Model Attributes Interaction Duration Document Types Best Match Models System Model Features Actors or Components Document Genres Degree of Doc. Structure and Content Used User Model Features Kind of Interaction and Access Information Type in Document Use of NLP to Document Indexing System Model Adaption Strategies and Tactics Communication Function

  • Doc. Metadata

Representation User Model Building Purpose of Human Communication Temporal Aspects Use of Weights in Doc. indexing Request Model Builder Purpose of System Communication Document Sign Language Degree of Req. Structure and Content Used Retrieval Strategy Interaction Mode Layout and Style Use of NLP to Request Indexing Response Generation Least effort Factors Document Isness

  • Req. Metadata

Representation Feedback Generation

  • Document Content

Use of Weights in Requests Mapping ST History Contextual Hyperlink Structure Explanation Features Human Source (see Actor) Transformation of Messages Scheduler

Controlled Variables

slide-52
SLIDE 52

Essir2011 52 Ingwersen

Agenda - 3

 Naturalistic Field Investigations of IIR (20 min)

 Integrating context variables  Live systems & (simulated/real) work tasks  Sample Study – Marianne Lykke Nielsen (2001; 2004)

 Wrapping up of Tutorial (5 min)

Questions are welcome during the tutorial sessions

slide-53
SLIDE 53

Keep things simple!

 If you can isolate one (or two) variables as

independent – then stick to that.

 Real-life studies are much more uncertain and

complex than laboratory tests

 A robust research setting is crucial  Natural search jobs (e.g. exploratory) mixed

with simulated ones (but must be realistic!)

 Test persons do relevance assessments!

Essir2011 53 Ingwersen

slide-54
SLIDE 54

Essir2011 54 Ingwersen

Agenda - 3

 Naturalistic Field Investigations of IIR (20 min)

 Integrating context variables  Live systems & (simulated) work tasks  Sample Study – Marianne Lykke Nielsen (2001; 2004)

 Wrapping up (5 min)

Questions are welcome during the tutorial sessions

slide-55
SLIDE 55

Essir2011 55 Ingwersen

Natural IR Interaction KMS Sample

 Marianne L. Nielsen (2004): Task-based evaluation

  • f associative thesaurus in real-life environment.

Proceedings of the ASIST 2004 Annual Meeting; Providence, Rhode Island, November 13 - 18, 2004. 437 - 447.

 Research setting: Danish Pharmaceutical Company  Goal: To observe if a company thesaurus (ontology)

based on human conceptual associations affects searching behavior, retrieval performance and searcher satisfaction different from a domain-based thesaurus.

slide-56
SLIDE 56

Essir2011 56 Ingwersen

Associative Thesaurus - ASSO

 Nielsen, M.L. (2001). A framework for work task based

thesaurus design. Journal of Documentation, 57 (6), 774- 797.

 Made from several association tests with 35 employees

from the company, supplying synonyms, narrow and broader concepts, based on the ”company vocabulary” (task/product-based)

 This thesaurus was larger in number of entries (379

more) and associative terms than the ”control thesaurus” – made by domain expert and based on the ”scientific vocabulary”.

slide-57
SLIDE 57

Essir2011 57 Ingwersen

Research Design - 1

 20 test persons from the basic and clinical researchers,

including marketing employees (also with scientific background)

 3 simulated search task situations (next slide) per

test person, all having same structure and based on real work tasks observed by recently logged requests to company retrieval system.

 “Blind testing” of the two thesaurus types: test

persons were told that the investigation was part of the system design process. Only the research team knew who searched which thesaurus type!

slide-58
SLIDE 58

Essir2011 58 Ingwersen

Search Job A

You are Product Manager working for Lundbeck Pharma. A physician who wants to know if the combination of Citalipram and Lithium leads to approve therapeutic effect on Bipolar Disorders, has consulted you. You need to find reports or articles investigating interaction and effect of the two drugs.

slide-59
SLIDE 59

Essir2011 59 Ingwersen

Research Design - 2

Steps in the field study (2 hours per test person): 1. Capture search skills (e-mail questionnaire) 2. Explanation session 3. Pre-search interview of searcher’s mental model concerning each search job / expectations 4. Search session with relevance assessments (logging and structured observation of each job) 5. Post-search interview of motivation & satisfaction for each search job.

slide-60
SLIDE 60

Essir2011 60 Ingwersen

Research Design - 3

 Latin square execution (slide 59) to avoid learning

effects & all search jobs are tried out on both thesauri:

 10 persons x 3 search jobs in ASSO = 30 units  10 persons x 3 search jobs in DOMAIN = 30 units

(in reality there were only 2 x 29, due to error)

 Relevance assessments: three-graded: Highly

relevant; Partially relevant; Not relevant.

 Measures: Recall/Precision; Behavior; Satisfaction

slide-61
SLIDE 61

Essir2011 61 Ingwersen

Research Design - 4

 Independent Variable:  Document Metadata Representation (two values)  Controlled Variables:

 Natural Work/Search Task Org. setting  Perceived Work Task Structure; Complexity (high)  Perceived Information Need  Database; Retrieval Engine; Interface

 Hidden Variables: Test person characteristics

slide-62
SLIDE 62

Essir2011 62 Ingwersen

Naturalistic Field Study (M.L. Nielsen) - variables

Information Objects IT

Metadata Struc.

Thesauri Task Org.

Context

Interface Social Cultural

Work task perception Search task perception Actor Char.

slide-63
SLIDE 63

Essir2011 63 Ingwersen

Document and Source IR Engines IT Component IR Inter-faces Access and Interaction

Document Structure Exact Match Models Domain Model Attributes Interaction Duration Document Types Best Match Models System Model Features Actors or Components Document Genres Degree of Doc. Structure and Content Used User Model Features Kind of Interaction and Access Information Type in Document Use of NLP to Document Indexing System Model Adaption Strategies and Tactics Communication Function

  • Doc. Metadata

Representation User Model Building Purpose of Human Communication Temporal Aspects Use of Weights in Doc. indexing Request Model Builder Purpose of System Communication Document Sign Language Degree of Req. Structure and Content Used Retrieval Strategy Interaction Mode Layout and Style Use of NLP to Request Indexing Response Generation Least effort Factors Document Isness

  • Req. Metadata

Representation Feedback Generation

  • Document Content

Use of Weights in Requests Mapping ST History Contextual Hyperlink Structure Explanation Features Human Source (see Actor) Transformation of Messages Scheduler

slide-64
SLIDE 64

Essir2011 64 Ingwersen

Natural Work Tasks (WT) & Org Natural Search Tasks (ST) Actor Perceived Work Tasks Perceived Search Tasks

WT Structure ST Structure Domain Knowledge Perceived WT Structure Perceived Information Need Content WT Strategies & Practices ST Strategies & Practices IS&R Knowledge Perceived WT Strategies & Practices Perceived ST Structure/Type WT Granularity, Size & Complexity ST Granularity, Size & Complexity Experience on Work Task Perceived WT Granularity, Size & Complexity Perceived ST Strategies & Practices WT Dependencies ST Dependencies Experience on Search Task Perceived WT Dependencies Perceived ST Specificity & Complexity WT Requirements ST Requirements Stage in Work Task Execution Perceived WT Requirements Perceived ST Dependencies WT Domain & Context ST Domain & Context Perception of Socio-Org. Context Perceived WT Domain & Context Perceived ST Stability Sources of Difficulty Perceived ST Domain & Context Motivation & Emotional State

slide-65
SLIDE 65

Essir2011 65 Ingwersen

Selected Results

 Both thesauri show same IR performance level  Both thesauri applied to Query Formulation &

Modification or as Lead-in Terms:

 Finding synonyms and /or more specific terms  Clarifying meaning (in task perspective) of terms

 ASSO applied slightly more time (used for

Narrow Terms capture)

 DOMAIN applied more in pre-search stage

slide-66
SLIDE 66

Essir2011 66 Ingwersen

Selected Results 2

 Recall / Precision:  ASSO: .14 / .32 – DOMAIN: .11 / .37  Note: test persons assessed same

documents quite differently!

 This was due to two fundamentally different

groups of test persons (hidden variable!):

 Basic researchers (exploring new drugs)  Clinical researchers (clinical drug tests)  This also concerns the satisfaction of the use of

the thesauri for IR (which was quite high)

slide-67
SLIDE 67

Essir2011 67 Ingwersen

Agenda - 3

 Naturalistic Field Investigations of IIR (20 min)

 Integrating context variables  Live systems & (simulated) work tasks  Sample Study – Marianne Lykke Nielsen (2001; 2004)

 Wrapping up of Tutorial (5 min)

Questions are welcome during the tutorial sessions

slide-68
SLIDE 68

Essir2011 68 Ingwersen

Step-by-Step into Light!

 In pure ‘laboratory experiments’ only simulations of

searcher behavior can be done;

 If one wishes to stick to existing test collections, with

existing sets of relevance assessments and ‘topics’, only IR interaction ‘ultra-light’ can be done (in order to avoid learning effects by test persons):

 Requires short-term IR interaction;  In the form of ‘laboratory studies’.  Number of test persons, search jobs and research setting follow

same line as Interactive ‘light’ IR.

slide-69
SLIDE 69

Essir2011 69 Ingwersen

Step-by-Step into Context - Light!

 IR interaction ‘light’ entails session-based IR, with test

persons’ relevance assessments and more intensive monitoring (logs; interviews; observation);

 Can be carried out as laboratory study or field experiment

 Like in ‘ultra-light’ and ‘naturalistic’ IR, number of test

persons and search jobs must assure that ‘statistically enough’ data is present in the result matrices when cross tabulating independent variables (see slides 40-41).

 IR interaction ‘light’: assigned realistic requests,

simulated task situations and searcher relevance assessments

 Naturalistic IR interaction assumes natural tasks

(mixed with simulated ones) in natural environments

slide-70
SLIDE 70

‘Ultra-light’ and ‘Light’ IIR

 Assessments can be 4-graded (Vakkari & Sormunen, 2004);  Realistic but few relevance assessments per person;  Assessments can be pooled for same search job

  • ver all test persons – weighted doc. assessments

 Common recall/precision, MAP, CumGain, P@n,

  • etc. feasible

 You require min. 30 responses per result cell  Ultra-light lab. studies are effective for tightly

controlled IIR experiments (like Kelly et al.)

Essir2011 70 Ingwersen

slide-71
SLIDE 71

Essir2011 71 Ingwersen

The Cognitive Research Framework informs about …

 Central variables to combine as independent

  • nes

 Major variables kept controlled/neutralized in a

setting

 What kind of variables that are hidden!  Dependent variables depend on the research

goals (the independent variables!) Novel possible research designs, settings and measures … there is a lot to do - really!

slide-72
SLIDE 72

Essir2011 72 Ingwersen

http://www.springeronline.com/1-4020-3850-X /

THANK YOU!