Developing Praxis Tests Tennessee State Board of Education Workshop - - PowerPoint PPT Presentation

developing praxis tests
SMART_READER_LITE
LIVE PREVIEW

Developing Praxis Tests Tennessee State Board of Education Workshop - - PowerPoint PPT Presentation

Developing Praxis Tests Tennessee State Board of Education Workshop November 14, 2019 Involving Educators to Develop Praxis Tests From Design through Implementation Design Structure of Test Development Educator Advisory Committee


slide-1
SLIDE 1

Developing Praxis Tests

Tennessee State Board of Education Workshop November 14, 2019

slide-2
SLIDE 2

Involving Educators to Develop Praxis Tests

From Design through Implementation

  • Development

Advisory Committee

  • Job Analysis Survey

Determine Content Domain

  • National Advisory

Committee

  • Confirmatory Survey

Design Structure

  • f Test
  • Educator

Consultants

  • Multistate Standard-

Setting Study (MSSS) Panel

Develop and Administer Test

2

slide-3
SLIDE 3
  • Ensuring diverse perspectives by recruiting

educators …

  • across states that use Praxis
  • from varied educational settings
  • rural, suburban & urban schools
  • small, mid-size & large colleges/universities
  • Work with state agencies and associations to

build diverse committees with regards to gender and race/ethnicity

3

Involving Educators to Develop Praxis Tests

From Design through Implementation

slide-4
SLIDE 4

Praxis Development Process

Accumulation of validity evidence to support the use of Praxis tests

4

slide-5
SLIDE 5

Development Steps and Validity Chain

Select and review appropriate standards Identify relevant and important knowledge and skills

Confirmatory Survey

Confirm the relevance and importance of the test specifications Translate knowledge and skills into test specifications STEP 1: STEP 2: STEP 3: STEP 4:

DEVELOPMENT STEPS

Basing the initial knowledge/skills domain on existing standards accepted by the profession Further refining the initial domain of knowledge/skills based on input from subject matter experts (SMEs)

NAC

Independent verification of the job-relatedness

  • f the

knowledge/skills Building test specifications to reflect identified knowledge/skills

DAC, Job Analysis Survey VALIDITY CHAIN BLUE boxes represent steps that rely heavily on educators

slide-6
SLIDE 6

Development Steps and Validity Chain

Select and review appropriate standards STEP 1:

DEVELOPMENT STEPS

Basing the initial knowledge/skills domain on existing standards accepted by the profession

VALIDITY CHAIN BLUE boxes represent steps that rely heavily on educators

slide-7
SLIDE 7

Aligning to Appropriate Standards

Praxis Te Test

  • Teaching Reading:

Elementary

  • Biology: Content Knowledge
  • Special Education: Content

Knowledge & Applications Nation ional al S Stan andards

  • International Literacy

Association

  • Next Generation Science

Standards National Science Teachers Association

  • Council for Exceptional

Children

7

slide-8
SLIDE 8

Development Steps and Validity Chain

Identify relevant and important knowledge and skills STEP 2:

DEVELOPMENT STEPS

Further refining the initial domain of knowledge/skills based on input from subject matter experts (SMEs)

DAC, Job Analysis Survey VALIDITY CHAIN BLUE boxes represent steps that rely heavily on educators

slide-9
SLIDE 9

Online Job Analysis Survey

slide-10
SLIDE 10

Online Job Analysis Survey

slide-11
SLIDE 11

Development Steps and Validity Chain

Confirmatory Survey

Confirm the relevance and importance of the test specifications Translate knowledge and skills into test specifications STEP 3: STEP 4:

DEVELOPMENT STEPS NAC

Independent verification of the job-relatedness

  • f the

knowledge/skills Building test specifications to reflect identified knowledge/skills

VALIDITY CHAIN BLUE boxes represent steps that rely heavily on educators

slide-12
SLIDE 12

Test Specifications

12

Test specifications provide detailed description of the content of the test to guide

  • students preparing to the

test, and

  • preparation programs

developing curricula

slide-13
SLIDE 13

Development Steps and Validity Chain

13

Develop test items and scoring keys/rubrics Multiple reviews of each test item

Educator Consultants

Assemble and review test forms Items written to measure test specifications Verification of linkage between test items and test specifications Verification of linkage between test form and test specifications

Educator Consultants

STEP 5: STEP 6: STEP 7:

DEVELOPMENT STEPS VALIDITY CHAIN BLUE boxes represent steps that rely heavily on educators

slide-14
SLIDE 14

Evidence Gathering … … Developing Relevant Test Items

14

Develop test items and scoring keys/rubrics Items written to measure test specifications

Educator Consultants

STEP 5:

  • What must the test taker SHOW? (i.e., critical behavioral

indicators)

  • In other words, “What would someone have to know or know how to do

in order to show that knowledge or accomplish that skill?”

  • Is this necessary at the time of entry into the

profession?

slide-15
SLIDE 15

15

Test Specs to Evidence Example

Knowledge Statement: “Is familiar with the provisions of major legislation that impact the field of special education (e.g., Public Law 94-142, IDEA 2004, Section 504).” In order to conclude that the test taker “Is familiar with the provisions of major legislation …” he or she must be able to….

  • Identify the major aspects of IDEA
  • Determine when a child is eligible for a 504
  • Compare an IEP and a 504 plan
slide-16
SLIDE 16

16

Test Item Mapped to Test Specs

Sample Item: According to the least restrictive environment provision in the Individuals with Disabilities Education Act (IDEA), a student with a disability must be educated with non- disabled peers (A) when appropriate facilities are available (B) only if the student has a mild disability (C) if the student has a severe disability (D) to the greatest extent possible  Identify the major aspects of IDEA

slide-17
SLIDE 17

Development Steps and Validity Chain

17

Conduct standard- setting study Verify item- and test- level performance before reporting scores Ongoing review of each Praxis test title to assure the content domain continues to reflect the field

  • If significant changes to the content domain have occurred (e.g.,

new SPA standards), the test is redesigned (beginning at Step #1) Verification of proper performance

  • f test items prior to

scoring/reporting Using educators to recommend a performance standard to policymakers

MSSS Panel

STEP 10:

DEVELOPMENT STEPS

STEP 8: STEP 9:

VALIDITY CHAIN BLUE boxes represent steps that rely heavily on educators

slide-18
SLIDE 18

Development Steps and Validity Chain

18

Conduct standard- setting study Using educators to recommend a performance standard to policymakers

MSSS Panel DEVELOPMENT STEPS

STEP 8:

VALIDITY CHAIN BLUE boxes represent steps that rely heavily on educators

slide-19
SLIDE 19

Standard-Setting

  • The standard-setting process for a new or

revised Praxis test is the final phase in the development process

  • The credibility of the standard-setting effort is

established by properly f following a reasonable and rational s system o

  • f rules a

and proce cedures that result in a test score that differentiates levels of performance (Cizek, 1993)

19

slide-20
SLIDE 20

Standard-Setting Components

  • Standard setting involves three important

components

  • The first component is the test itself. The test is

designed to measure knowledge and skills determined to be important for competent performance as a beginning teacher.

  • The second component is the describing of the

level of knowledge and skills necessary for competent performance.

  • The last component is the process for mapping

the description onto the test.

20

slide-21
SLIDE 21

Steps in the Process

  • First step was understanding the test
  • Prior to the study, panelists were asked to review the

specifications for the test they would be evaluating.

  • At the study, following an overview of the licensure

process and standard setting, the panelists “took the test.”

  • Then the panel discussed the content of the test

and what is expected of beginning teachers.

The purpose of these activities is to familiarize the panelists with what is being measured and how it is being measured.

21

slide-22
SLIDE 22

Steps in the Process (cont’d.)

  • Next the panelists developed a profile or

description of the “just qualified candidate”

  • r JQC.
  • The JQC is the candidate who just crossed that

threshold of demonstrating the level of knowledge and skills needed to enter the profession.

  • The definition highlights the knowledge and skills

that differentiate the candidate just over the threshold from the candidate who is not quite there yet.

22

slide-23
SLIDE 23

Describing a Just Qualified Candidate

Not Yet Qualified Qualified

Still Not Qualified Just Qualified Passing Score Low Score High Score

slide-24
SLIDE 24

Steps in the Process (cont’d.)

  • Now the panelists were ready to make their

standard-setting judgments.

  • Panelists were trained in the standard setting

method, had an opportunity to practice making judgments, and then made their question-by- question judgments.

  • Modified A

Angoff method for selected-response questions– judge the likelihood that a JQC will answer a question correctly

  • Exten

ended ed A Angoff method for constructed- response questions– judge the rubric score JQC would likely earn

24

slide-25
SLIDE 25

Standard-Setting Methods (cont’d.)

  • Multiple r

rounds—Panelists made two rounds of judgments. ‒During the first round, panelists made independent judgments. ‒The judgments were summarized, both at a question and overall test level, and panelists engaged in discussions about their rationales for particular judgments. ‒After discussion, the panelists could change their original judgments.

25

slide-26
SLIDE 26

Panelists’ Evaluation

  • Critical to the validity of the standard-setting

process is that (a) panelists understand the task, and (b) implementation of the study as planned.

  • Following training and before the panelists begin

making judgments, they were asked to confirm that they understand the process and the judgment task.

  • After the study, the panelists were asked to

complete an evaluation of the study — their understanding of the steps in the process, the effectiveness of key steps, and their overall impressions of the recommended passing scores.

26

slide-27
SLIDE 27

Setting Operational a Passing Score

  • Each state reviews the information from the study

and decides what it will adopt as its passing score for the test

  • States may want to consider other information
  • Estimated conditional standard error of measurement
  • Standard error of judgment
  • Importance of minimizing false positives or false

negatives

slide-28
SLIDE 28

Development Steps and Validity Chain

28

Verify item- and test- level performance before reporting scores Verification of proper performance

  • f test items prior to

scoring/reporting

DEVELOPMENT STEPS

STEP 9:

VALIDITY CHAIN

slide-29
SLIDE 29

Item Analysis

  • How difficult is it?
  • How well does it distinguish high from low ability?
  • How do the incorrect options behave?
  • Does it have a single

le correct response?

Does each question behave as expected?

slide-30
SLIDE 30

Item Statistics

  • Diffic

fficult lty – how hard is the question for a group of test takers?

  • Di

Discrimination – how sharply does the question separate test takers who are generally strong in the subject from those who are generally weak?

  • Candidates with higher total test scores should have a

higher probability of answering a question correctly.

slide-31
SLIDE 31

Sample Item Analysis

31

Item difficulty

slide-32
SLIDE 32

Sample Item Analysis

32

Item discrimination

slide-33
SLIDE 33

Another Sample Item Analysis

33

slide-34
SLIDE 34

Differential Item Functioning

Is an item particularly hard or easy for test takers from specified demographic groups?

Focal Reference

  • Female

vs. Male

  • African American

vs. White

  • Asian American

vs. White

  • American Indian vs.

White

  • Hispanic

vs. White

slide-35
SLIDE 35

Differential Item Functioning

0.0 20.0 40.0 60.0 80.0 100.0 10 20 30 40 50

Test Score % Correct

Focal Reference 0.0 20.0 40.0 60.0 80.0 100.0 10 20 30 40 50

Test Score % Correct

Focal Reference

An item with DIF An item with no DIF

slide-36
SLIDE 36

Differential Item Functioning

  • DIF ≠ Impact
  • Impact = difference in performance of two intact groups.
  • DIF = difference in performance of two groups

conditioned on ability

  • Impact can often be explained by differences in

preparation across groups

  • DIF ≠ Item bias
  • DIF is used as one way to evaluate whether there is item

bias.

  • Content experts will review and determine if DIF found is

due to item bias.

slide-37
SLIDE 37

Converting Raw Scores to Scale Scores

  • Scaling
  • Placing a candidate’s raw score (number correct) onto

the Praxis 100 to 200 reporting scale

  • Equating
  • Putting two or more essentially parallel forms on a

common scale

37

slide-38
SLIDE 38

An Illustration of Equating Scaling

100 105 110 115 120 125 130 135 140 145 150 155 160 165 170 175 180 185 190 195 200 2 4 6 8 10 12 14 16 18 20 22 24 26 28 30 32 34 36 38 40 42 44 46 48 50

38

Scores at or below chance are scaled to 100 Scores at or above 95% are scaled to 200 Scale is established on the FIRST form.

slide-39
SLIDE 39
  • Statistical procedure to find equivalent scores on two

different forms that may be of different difficulty levels.

Scaled Score Base Form 2nd Form 3rd Form

50 … 26 25 24 … 50 … 26 25 24 … 50 … 26 25 24 … 200 … 144 142 138 … 100 Scaling Equating Equating

An Illustration of Equating

2nd Form more difficult than Base Form 3rd Form easier than 2nd and Base Forms

slide-40
SLIDE 40

Involving Educators to Develop Praxis Tests

From Design through Implementation

  • Development

Advisory Committee

  • Job Analysis Survey

Determine Content Domain

  • National Advisory

Committee

  • Confirmatory Survey

Design Structure

  • f Test
  • Educator

Consultants

  • Multistate Standard-

Setting Study (MSSS) Panel

Develop and Administer Test

40

slide-41
SLIDE 41

41

If you have any Praxis questions, please contact Kathy Pruner, Tennessee Client Relations Director kpruner@ets.org Clyde Reese, PEP Data and Validity Director creese@ets.org