Assessment & Evaluation Richard J. Shavelson Stanford - - PowerPoint PPT Presentation

assessment evaluation
SMART_READER_LITE
LIVE PREVIEW

Assessment & Evaluation Richard J. Shavelson Stanford - - PowerPoint PPT Presentation

Assessment & Evaluation Richard J. Shavelson Stanford University & Yosemite National Institutes NPS Symposium San Diego, California April 18, 2002 Sessions I & II: Evaluation Background & Group Activity Evaluation: Uses,


slide-1
SLIDE 1

Assessment & Evaluation

Richard J. Shavelson

Stanford University & Yosemite National Institutes

NPS Symposium

San Diego, California

April 18, 2002

slide-2
SLIDE 2

4/18/2002 NPS Symposium 2

Sessions I & II: Evaluation Background & Group Activity

  • Evaluation: Uses, Misuses and a Working

Definition

  • Distinctions: Assessment & Evaluation
  • Evaluation: Formative and Summative
  • Evaluation: Interpretative and Other Educational

Programs

  • Small-Group Activity: Designing an Evaluation
slide-3
SLIDE 3

4/18/2002 NPS Symposium 3

Evaluation: Uses, Misuses— The YNI Experience

  • Initial Use:

– Education Committee: Improve field-education program – Administration & Some Board Members: Demonstrate, through an independent “Stanford” evaluation, the effectiveness of YNI’s field-based education program

  • Initial Misuse: Evaluation was viewed as a marketing tool

with the imprimatur of “Stanford University”

  • Current Use: Evaluators turned initial purpose into a focus
  • n field-based education improvement, especially inquiry-

based learning

slide-4
SLIDE 4

4/18/2002 NPS Symposium 4

Evaluation: A Working Definition

Evaluation is the art (not a “model”) of bringing conceptual, political and empirical evidence to bear on some NPS Program using a variety of tools from various social sciences in order to reach a judgment as to the value of the program in meeting goals, intended and unintended.

slide-5
SLIDE 5

4/18/2002 NPS Symposium 5

Some Distinctions: Assessment & Evaluation

  • Assessment refers to a combination of social

measurements--a test, a performance assessment, an observation scale, an interview scale, or attitude scale score--that bears on an outcome of interest.

  • Evaluation is a judgment of the value of a

program under inquiry on the basis of artful and rigorous application of scientifically justifiable method

slide-6
SLIDE 6

4/18/2002 NPS Symposium 6

Example (“Summative”) Evaluation:

CPB/Annenberg Distance Education Curriculum

  • Background & Question—does it work?
  • Evaluator’s turn of question—

Exchangeability

  • Absence of evaluation model to fit context
  • Creativity in quasi-experimental design

with quantitative and qualitative data

  • Judgment of value
slide-7
SLIDE 7

4/18/2002 NPS Symposium 7

Formative & Summative Evaluation

  • Formative—

– To improve the NPS Program during its development process – Feedback to close gap between goals and present conditions

  • Summative

– To judge the overall value of the NPS Program – Best conceived as as comparing the value of the program against alternatives or standards sometimes considering program cost (e.g., CPB evaluation, class size reduction evaluation)

slide-8
SLIDE 8

4/18/2002 NPS Symposium 8

Use of Formative Evaluation— The YNI Experience

Doing Well Needs Improvement Recommendations

Staff Training: Focus on Inquiry, Community Connections, Bio-monitoring, Diversity Staff Training: YI: more inquiry activities, diversity/connections training Staff Training: All: Assess staff competencies and create training around weaknesses Pre-Trips: Realize that we have as much to learn from pre-trip as students do Pre-Trips: Improve knowledge-sharing between groups Pre-Trips: All: Create communication between Outreach/Field Science Inquiry: Staff is getting more comfortable with inquiry. Trainings on topic have occurred at all campuses. HI, OPI great progress Inquiry: YI: Need more support/training Inquiry: All: Continue to provide information HI/YI/YNI: 5 staff attend IFI training in June Stewardship: Stewardship being recognized as more than end-of-program topic Stewardship: Role/topic of service learning Stewardship: Clarify opportunities for stewardship and integrate into program Science: Inquiry activities more common, biomonitoring occurring Science: Inconsistent communication about role of science in program Science: Discuss history of science role in education and create action plan for clarifying within organization

Source: Schneider Quarterly Report 2002

slide-9
SLIDE 9

4/18/2002 NPS Symposium 9

Use of Summative Evaluation— The Bavarian Experience

Source: Bogner,

  • F. (1998). The

influence of short-term

  • utdoor ecology

education on long-term variables of environmental perspectives. Journal of Environmental Education, 29(4), 17-29.

slide-10
SLIDE 10

4/18/2002 NPS Symposium 10

Small-Group Discussion: Designing an Evaluation

  • Groups Choose a Problem:

– Design an evaluation of a National Park’s Interpretation Program – As part of a National Park’s strategic plan for linking with external field-based science education program, design an evaluation to inform selection decisions

  • Consider:

– Mission – Activities – Goals – Credible evidence

  • Groups Report Back
slide-11
SLIDE 11

4/18/2002 NPS Symposium 11

Appendix to Session I: CPB Evaluation Design

Community College Semester

Fall Class Spring Class Final Exam

T

1.29 2.82!

Treatment (T)

Site A

Spring Pretest Spring Final 3.37!!

(T) T

1.26 0.16

Site H

1.51

slide-12
SLIDE 12

4/18/2002 NPS Symposium 12

Session III: Overview

  • Evaluation Questions and Appropriate

Designs

  • Some Problems with Environmental

Education Evaluation “Models”

  • Some Problems with Interpretation of

Environmental Education Findings

slide-13
SLIDE 13

4/18/2002 NPS Symposium 13

Evaluation Questions and Appropriate Designs

  • What’s happening?

– Statistical point and relational estimates – Case and other qualitative (e.g., ethnographic) studies

  • Is there a systematic effect?

– Randomized experiments – Quasi-experiments – Correlational studies

  • How is it happening?

– Experiments – Correlational studies – Case and other qualitative (e.g., ethnographic) studies

Source: Shavelson & Town (Eds.) (2002), Scientific Research in Education. National Academies Press

slide-14
SLIDE 14

4/18/2002 NPS Symposium 14

Some Problems with Environmental Education Evaluation Models

  • Inadequate design to address question

– One-shot and pre-post studies to address summative- evaluation (“Effects”) questions – Posttest attitude questionnaires to address participants’ perceptions – Seldom formative evaluation reported

  • “A variety of research designs and data collection methods

were employed. Surveys, observation, and interviewing were used most often, but some studies discuss… focus groups, quizboard testing, or photography…. Several studies employ a pre- and posttest design… but more often… a posttest. Often the form of experimental or quasi-experimental design was unclear” (Wells & Smith, 2000, p. 1).

slide-15
SLIDE 15

4/18/2002 NPS Symposium 15

Some Problems with Environmental Education Evaluation Models (Cont’d.)

  • Over-interpretation of environmental education

findings

  • We have learned that:

– “Education programs can effectively accomplish the dual-goal of helping students achieve national Science Standards while at the same time fostering stewardship of national parks” (Reynolds, 2001, p. 1). – Based on ... case studies of schools that use environmental education [EE] as the focus for their curriculum… Sward found ‘… compared to traditional educational approaches, [EE] improves academic performance across the curriculum’ (Glenn, 2000)” (ASCD, 2001).