Using process indicators to Using process indicators to assess and - - PowerPoint PPT Presentation

using process indicators to using process indicators to
SMART_READER_LITE
LIVE PREVIEW

Using process indicators to Using process indicators to assess and - - PowerPoint PPT Presentation

1 Using process indicators to Using process indicators to assess and improve the quality of assess and improve the quality of LFS data collection procedures LFS data collection procedures Dag F. Gravem Statistics Norway 1 The background


slide-1
SLIDE 1

1

1

Using process indicators to Using process indicators to assess and improve the quality of assess and improve the quality of LFS data collection procedures LFS data collection procedures

Dag F. Gravem Statistics Norway

slide-2
SLIDE 2

2

The background

  • For the 4th quarter of 2011, the new SIV case management

system was introduced for the Norwegian LFS

– Built on top of Blaise – Modifications made to Blaise

  • The old CAI system: mainly list-based CATI interviewing by

locally based interviewers (CATI-L)

– Electronic lists consisting of sampled families and their members

  • The SIV system: mainly database-based CATI interviewing

by call centre interviewers (CATI-D)

– Respondents distributed to ”one by one” – With list-based CATI follow-up by locally based interviewers

slide-3
SLIDE 3

3

Implications of the SIV system

  • Respondents distributed to interviewers ”one by one”

– Treated more as sampled individuals, and less as members of a sampled family

  • The relationship between respondent and interviewer

becomes more impersonal

– Respondents communicate with several different interviewers instead of one – Call centre interviewers communicate with many more respondents per session than locally based interviewers

  • Locally based interviewers receive follow-up cases only

– This may influence their motivation

  • More process data is collected!
slide-4
SLIDE 4

4

Process data collected for each contact attempt

  • Interviewer ID
  • Entry priority

– Default – Appointment – Selected by interviewer

  • Date
  • Time
  • Result

– Interview – No answer – Non-response

  • Phone number(s) called
  • Appointment data
  • Duration of interview: does not work
  • …and lots more (too much)
slide-5
SLIDE 5

5

Some issues with the process data

  • Two different sources

– Blaise CATI-D system: Automatically generated – CATI-L follow-up: Manually recorded by interviewers

The interviewer may not always follow procedures

  • Not completely integrated

– Reflects the different routines and tasks of CATI-D and CATI-L interviewers

  • Need for more data

– More detailed nonresponse information – Time used for each contact attempt

  • On the positive side: Process data is available in a file that

also contains administrative data and questionnaire data!

slide-6
SLIDE 6

6

The SIV system: Hopes and expectations regarding improved process quality

  • Improved data collection timeliness

– At the CATI-D call centre, we may to a larger extent control which survey the interviewer is working on – Potentially leading to improved estimates

  • Reduced costs

– Resources may be allocated more efficiently – Call centre interviewers are cheaper

  • Data collection more in line with responsive design

– Easier to target underrepresented groups – Potentially increasing the representativity of the net sample

  • What do the process (and other) data tell us..?
slide-7
SLIDE 7

7

Timeliness of the data collection

  • No. of days after the end of the reference

week that interviews are completed Q1 2011 Q2 2011 Q3 2011 Q4 2011

  • No. of interviews

19 702 19 334 19 026 18 619 Median value 3 4 4 4 Mean value 7.32 8.78 9.05 10.04

  • Std. deviation

9.79 10.95 11.82 12.31

slide-8
SLIDE 8

8

Timeliness of the data collection

  • The timeliness of the data collection did not improve in the

4th quarter

  • It rather continued a negative timeliness trend throughout

2011, with

– Fewer and fewer interviews – More and more time needed to complete interviews

  • Still, a weak point in the new system was detected: the

transition from CATI-D to the CATI-L follow-up phase

slide-9
SLIDE 9

9

Timeliness, 3rd and 4th quarter compared

Number of interviews by day after the end of the reference week. Accumulated.

2000 4000 6000 8000 10000 12000 14000 16000 18000 20000 1 7 13 19 25 31 37 43 49 55 61 67 73 79 85 91 97 Day after reference week Interviews accumulated Q3 2011 Q4 2011

slide-10
SLIDE 10

10

Timeliness, 3rd and 4th quarter compared

  • The 4th quarter data collection keeps pace with the 3rd

quarter until the middle of the 2nd week of data collection.

  • The 4th quarter data collection then loses pace dramatically

through the rest of the 2nd week of data collection

  • We never managed to catch up
  • This time coincides with the transferring of non-response

from CATI-D to CATI-L

– A lag was to be expected, but not this dramatic – Why weren’t we able to catch up?

slide-11
SLIDE 11

11

Explanations

  • Routines were new to the CATI-L interviewers
  • CATI-L interviewers received cases in mid-week, and had

less time to plan and start calling

  • CATI-L interviewers only received nonresponse follow-up,

with cases that previously had been contacted by CATI-D interviewers.

– This may have had a negative impact on motivation

slide-12
SLIDE 12

12

Motivation: probability of getting an interview by contact attempt. LFS 4th quarter 2011

Contact attempt Probability of get- ting an interview First 27 % Second 26 % Third 21 % Fourth 17 % Fifth 14 % Sixth 13 %

  • CATI-L interviewers generally

received cases that had been tried, though some cases had not been tried due to CATI-D undermanning

  • We now try to limit the number of

contact attempts each case is exposed to during the CATI-D phase

slide-13
SLIDE 13

13

Data collection timeliness: some improvement in the 1st quarter of 2012

  • No. of days after the end of the reference week that

interviews are completed Q1 2011 Q2 2011 Q3 2011 Q4 2011 Q1 2012

  • No. of interviews

19 702 19 334 19 026 18 619 18 383 Median value 3 4 4 4 3 Mean value 7.32 8.78 9.05 10.04 6.95

  • Std. deviation

9.79 10.95 11.82 12.31 10.98

slide-14
SLIDE 14

14

Representativity of the net sample

  • Response rates have been declining, what about

representativity?

  • R-indicator: useful for measuring representativity over time

for one survey

– Use of different background variables – A value of 1 indicates perfect representation according to the variables used

  • LFS R-indicator variables

– Gender – Age group (11 values) – Urbanity (39 values) – Level of education (3 values)

  • Household size ought to be included
slide-15
SLIDE 15

15

Response rates and R-indicator compared

R-indicator and response rates for the LFS data collection.

0,1 0,2 0,3 0,4 0,5 0,6 0,7 0,8 0,9 1 Q4 2010 Q1 2011 Q2 2011 Q3 2011 Q4 2011 Q1 2012 R-indcator Response rate

slide-16
SLIDE 16

16

But what about LFS specific representativity?

  • Traditionally, 1st wave respondents have been slightly

underrepresented.

– This was reinforced when SIV was introduced

Quarterly response rates for the Norwegian LFS. 2009-2012.

50 55 60 65 70 75 80 85 90 Q1 2009 Q2 2009 Q3 2009 Q4 2009 Q1 2010 Q2 2010 Q3 2010 Q4 2010 Q1 2011 Q2 2011 Q3 2011 Q4 2011 Q1 2012 Quarter and year Percent Wave 1 All 8 waves

slide-17
SLIDE 17

17

LFS specific representativity, continued

  • With SIV, 1st wave respondents must be monitored extra

closely

– Recruiting to the panel is essential – In the 1st quarter of 2012, we started turning the ship…

  • We should also look at respondents’ reference weeks

– The reference weeks towards the end of each quarter suffer from lower response rates

  • Should we make an LFS specific representativity indicator

including these or other variables..?

slide-18
SLIDE 18

18

The costs of the data collection

Monthly interviewer wage payouts for LFS interviewing 2010-2012

200 250 300 350 400 450 500 J a n u a r y 2 1 A p r i l 2 1 J u l y 2 1 O c t

  • b

e r 2 1 J a n u a r y 2 1 1 A p r i l 2 1 1 J u l y 2 1 1 O c t

  • b

e r 2 1 1 J a n u a r y 2 1 2 A p r i l 2 1 2

Thousand

Year and month NOK

Payouts

  • Poly. (Payouts)
slide-19
SLIDE 19

19

Man hours

LFS Interviewer man hours. 2010-2012

1000 2000 3000 4000 5000 6000 7000 8000 Q 1 2 1 Q 2 2 1 Q 3 2 1 Q 4 2 1 Q 1 2 1 1 Q 2 2 1 1 Q 3 2 1 1 Q 4 2 1 1 Q 1 2 1 2 Quarter and year Man hours

Man hours worked

slide-20
SLIDE 20

20

The costs of the data collection

  • Not adjusted for wage increases, holidays, interview length
  • Some of the increases may be attributed to training
  • The cost figure includes data for April 2012, which was not

available at the time the paper was completed

– A more linear increase in wage costs than suggested in the paper – Still too early to determine whether SIV in itself has had an effect

  • The number of man-hours seems to have increased more

after SIV

– Call centre interviewers have lower wages than locally based interviewers

slide-21
SLIDE 21

21

Conclusions

  • Q: Did the SIV case management system lead to improved

data collection quality? A: Inconclusive

– The quality indicators presented in the paper have limitations – The new system creates and shapes data collection procedures, sometimes i unexpected ways – We work on re-shaping the system to fit the procedures we want, rather than having to adapt procedures to the system

  • Easier access to process data combined with administrative

and survey data will enable us to do analyses of how the data collection influences data quality more directly

– So far, we have had to focus on making systems and routines work – But how does e.g. the number of contact attempts influence bias, variance, mean squared error..?