CCSSO 2015 Symposium: Future of Science Assessment Lei Liu, - - PowerPoint PPT Presentation

ccsso 2015 symposium future of science assessment
SMART_READER_LITE
LIVE PREVIEW

CCSSO 2015 Symposium: Future of Science Assessment Lei Liu, - - PowerPoint PPT Presentation

CCSSO 2015 Symposium: Future of Science Assessment Lei Liu, Kathleen Scalise, Madeleine Keehner and Cindy Ziker Examples and demonstrations from the new U.S. National Assessment for Educational Progress (NAEP): Virtual science lab;


slide-1
SLIDE 1

CCSSO 2015 Symposium: Future of Science Assessment

Lei Liu, Kathleen Scalise, Madeleine Keehner and Cindy Ziker Examples and demonstrations from the new U.S. National Assessment for Educational Progress (NAEP): Virtual science lab; scenario-based tasks; domain modeling; and process data in hands-on and virtual objects tasks.

slide-2
SLIDE 2

CCSSO 2015 Symposium: Future of Science Assessment

Accessible, engaging assessment for all students in the NAEP science and engineering scenario-based tasks

Kathleen Scalise Director, NAEP Science, ETS 6/22/15

slide-3
SLIDE 3

Technology-Enhanced Assessments

6/23/15

  • Innovation is a central component for the future of

educational assessment. New claims about student reasoning, behavior, and mental processes in context, along with new data sources, new scoring methods, and new performance assessment tasks are driving the next generation

  • f science, mathematics, engineering and

technology assessments.

3

slide-4
SLIDE 4

4 ¡

Measurement Technology Information Technology

Source ¡of ¡Concept: ¡Wilson, ¡M. ¡(2003). ¡The ¡technologies ¡of ¡assessment, ¡Invited ¡ PresentaDon ¡at ¡the ¡AEL ¡NaDonal ¡Research ¡Symposium, ¡Toward ¡a ¡NaDonal ¡Research ¡ Agenda ¡for ¡Improving ¡the ¡Intelligence ¡of ¡Assessment ¡Through ¡Technology. ¡Chicago. ¡

Two Types Assessment Technology Innovations

slide-5
SLIDE 5

Information Technology Innovations

6/23/15

  • NAEP Pilot 2015 employs science “scenarios” and

simulators in rich tasks.

  • NAEP also uses “hybridized” hands-on science

tasks, and blocks of discretes (single) items.

  • The tasks offer tools and animations to elicit what

students know and can do through virtual and hands-on investigations.

  • U.S. National Assessment of Educational Progress

5

slide-6
SLIDE 6

Task: Community water well in a rural

  • village. Students investigate problems,

query avatars, explore data, and provide explanations (Carr, 2013).

Carr, P. (2013). Presentations Of Selected Items/Tasks by Developers Of Those Assessments: NAEP. Presented at the Invitational Research Symposium on Science Assessment, Washington, DC.

NOTE: TEL Wells movie to be played.

Simu mula lations ns: T : TEL W Wells lls T Task k

slide-7
SLIDE 7

Task: Community water well in a rural

  • village. Students investigate problems,

query avatars, explore data, and provide explanations (Carr, 2013).

Carr, P. (2013). Presentations Of Selected Items/Tasks by Developers Of Those Assessments: NAEP. Presented at the Invitational Research Symposium on Science Assessment, Washington, DC.

NOTE: TEL Wells movie to be played.

Simu mula lations ns: T : TEL W Wells lls T Task k

slide-8
SLIDE 8

Engagement & Access Results: TEL

6/23/15

National Assessment Governing Board, May 2014

  • NCES shared information from students and

school staff after the 2014 TEL administration, including discussion of three positive themes that emerged:

  • High levels of student engagement in TEL tasks

(“now I think I might like to be an engineer”);

  • High levels of student completion of TEL

additional supplemental block;

  • Supportive reactions to TEL administration and

to task types in schools from school staff.

8

slide-9
SLIDE 9

Pump Trouble lesho hooting ng Activity

TEL Wells task is about proces

ess –

All students will (eventually) fix the pump.

We are interested in whether the proces ess is:

  • Efficien

ent: solves problem without unnecessary steps.

  • System

ematic: solves problem methodically, with a logical sequence of steps.

9 ¡

Source: ¡NCES, ¡Sept. ¡2013 ¡

slide-10
SLIDE 10

10 ¡

1 2 5 3 4

We capture proces ess d data:

  • Wha

hat is clicked (decisions/selections)

  • Or

Order der of clicks (sequences)

  • Numb

mber of clicks (frequencies)

  • Timi

ming ng of clicks (timestamps)

Provides a trail o

  • f a

actions so we can:

  • Recons

nstruct problem-solving process

  • Cha

haracterize different strategies

  • Inf

Infer underlying cognition

Source: ¡NCES, ¡Sept. ¡2013 ¡

slide-11
SLIDE 11

Characterizing “Efficient Actions”

11 ¡

What does an “efficient” pattern look like?

  • WHICH choices

you make

5 4 Source: ¡NCES, ¡Sept. ¡2013 ¡

slide-12
SLIDE 12

Characterizing “Systematic Actions”

12 ¡

What does a “systematic” pattern look like?

  • HOW you order

your choices

Source: ¡NCES, ¡Sept. ¡2013 ¡

slide-13
SLIDE 13

Games-based Assessment

13

Source: GlassLab, May 2015

slide-14
SLIDE 14

Conversation-based Assessment

6/23/15

14

Source: J. Gorin, CERA,

  • Dec. 2014
slide-15
SLIDE 15

Collaborative Tasks

Source: J. Gorin, CERA, Dec. 2014

slide-16
SLIDE 16

Multimodal Assessment: Live Performance

16

Source: J. Gorin, CERA, Dec. 2014

slide-17
SLIDE 17

Measurement Technology Innovations

6/23/15

  • Adaptivity is one example of measurement

technology innovation from NAEP.

  • In NAEP multistage adaptive tests (MST), the test

adaptation occurs based on student cumulative performance on a block of items. Multistage testing (MST) can be highly suitable because it can help better meet the needs of all students.

  • Also, NAEP doing a special study on the use of

adaptivity within the simulation tasks – “responsive” scenario-based tasks (RSBTs).

17

slide-18
SLIDE 18

Source: ¡ETS, ¡Nov. ¡2014 ¡

1 2 3 4 5 6 7

  • 5
  • 4
  • 3
  • 2
  • 1

1 2 3 4 5 6 Series1 Series2 Series3 Series4 Series5

Measurement Technology Innovation

slide-19
SLIDE 19

19

Examples of UDL tools available

  • Available Only in Discrete (Single) Items
  • 1. Elimination Tool (multiple-choice questions only)
  • 2. Highlighter Tool
  • 3. Zoom
  • 4. Word Definition (some items only)
  • Available in Discrete Items and SBTs (and

Survey Questions)

  • 5. Text to Speech (TTS)
  • 6. Hide/Show Timer
slide-20
SLIDE 20

20

Use Example: Text-to-Speech (TTS) Use on TEL Cognitive Items

  • TEL Di

Discrete It Items ms: Text-to-Speech (TTS) use ranged from about 6% to 30%.

  • TEL S

Scena nario-b

  • based T

Tasks ks ( (SBTs): In SBTs, TTS use ranged from 16% to 50% per task.

  • At the student level, 53% of students used

TTS at least once (either discrete or SBT).

Result of UDL tool use: TTS example

slide-21
SLIDE 21

Wrap-Up: Potential new directions for Science assessments

  • Ta

Tasks: Open-ended, more free-form

  • Authentically reflect real science and engineering practices
  • Eviden

ence: Includes rich process data, assistive tools

  • Pathways, sequences, timing of actions, tool choices
  • Rep

eporting: Beyond scaled scores

  • Insights into process, strategy, cognition

We have more research to do, but what we are learning can contribute to the development of more authentic, rich, and informative approaches to STEM assessment and reporting.

21 ¡

Source: ¡NCES, ¡Sept. ¡2013 ¡

slide-22
SLIDE 22

NRC ¡report ¡describes ¡that ¡a ¡“system” ¡of ¡assessment ¡is ¡needed: ¡

  • 1. Assessment ¡tasks ¡should ¡allow ¡students ¡to ¡engage ¡in ¡science ¡

pracDces ¡in ¡the ¡context ¡of ¡disciplinary ¡core ¡ideas ¡and ¡ crosscuTng ¡concepts. ¡ ¡ ¡

  • 2. MulD-­‑component ¡tasks ¡that ¡make ¡use ¡of ¡a ¡variety ¡of ¡response ¡

formats ¡will ¡be ¡best ¡suited ¡for ¡this. ¡ ¡

  • 3. Selected-­‑response ¡quesDons, ¡short ¡and ¡extended ¡constructed ¡

response ¡quesDons, ¡and ¡performance ¡tasks ¡can ¡all ¡be ¡used, ¡but ¡ should ¡be ¡carefully ¡designed ¡to ¡ensure ¡that ¡they ¡measure ¡the ¡ intended ¡construct ¡and ¡support ¡the ¡intended ¡inference. ¡ ¡

  • 4. Students ¡will ¡need ¡mulDple ¡and ¡varied ¡assessment ¡
  • pportuniDes ¡to ¡demonstrate ¡their ¡proficiencies ¡with ¡the ¡NGSS ¡

performance ¡expectaDons. ¡

22

NRC Report on Assessing NGSS

slide-23
SLIDE 23

Discussion & Questions: Future of Science Assessment

Accessible, engaging assessment for all students in the NAEP science and engineering scenario-based tasks

Contact Kathleen Scalise, 6/22/15 kscalise@ets.org,

slide-24
SLIDE 24

24 ¡

U.S. National Assessment of Educational Progress:

  • Largest nationally representative and continuing

assessment of what America's students know and can do in various subjects.

  • Provides the U.S. national and state “Report

Cards” and trend assessments, as well as many publications, products, and data tools, see http://nces.ed.gov/nationsreportcard/

What is NAEP?