7 testing testing big questions
play

7. Testing Testing: Big Questions How do teachers construct tests? - PowerPoint PPT Presentation

7. Testing Testing: Big Questions How do teachers construct tests? How are teacher-made tests like/unlike standardized tests? What information comes from test results? 7.1 Instructional Objectives 7.2 Teacher-Developed


  1. 7. Testing

  2. Testing: Big Questions 
 • How do teachers construct tests? 
 • How are teacher-made tests like/unlike standardized tests? 
 • What information comes from test results?

  3. 7.1 Instructional Objectives 7.2 Teacher-Developed Tests in the Classroom 7.3 Formative Evaluation 7.4 Classroom Grading Approaches

  4. 7.5 Criterion- Referenced Testing 7.6 Norm-Referenced Testing 7.7 Interpreting Norm- Referenced Tests Scores 7.8 Validity

  5. 7.9 Reliability 7.10 Test Bias 7.11 Using Tests Appropriately 7.12 Summary

  6. 7.1 Instructional Objectives

  7. Objectives: Checklist for learning 
 • More specific than goals 
 • What students should know or be able to do by end of lesson ➔ descriptive verbs! • Taxonomies provide hierarchies of increasing sophistication 
 • Bloom: Cognitive, affective, psychomotor

  8. Bloom’s taxonomies 
 • Cognitive most used 
 • 6 levels: remember, comprehend, apply, analyze, evaluate, create • Objective: “Students will compare and contrast yurts and tipis, in 3 key features.” 
 • Note task, level (analysis), criteria 
 ➔ “Mastery learning” system

  9. 7.2 Teacher-Developed Tests in the Classroom

  10. Classroom assessment 
 Backward planning as a “best practice” 
 1. Write objective with taxonomy-level verb and criteria for mastery 2. Create Assessment/test that fits objective 
 3. Plan learning activities that support and prepare students for mastery

  11. Classroom tests 
 • Essay: for comprehension, analysis; needs criteria 
 • Multiple choice, matching for recognition • T/F, fill blanks for recall 
 • Problem-solving for application/analysis 
 ➔ Consider pros/cons and kind of students who benefit

  12. Performance-based or authentic assessment 1 
 • Portfolio showing progress 
 • Exhibition, e.g. posters • Demonstration, e.g. slide shows, videos 
 • For individual or group assessment

  13. Authentic assessment 2 Rubric with criteria for scoring (posted for all to see) 10 points 5 points Sources Over 5 Under 5 Facts Over 10 Under 10 Format Correct Errors Graphics Over 5 Under 5

  14. 7.3 Formative Evaluation

  15. Formative assessments 1 
 • Assess/evaluate learning needs before instruction (aka “pretest”) 
 • Determine previous knowledge on topic or skill 
 • Determine readiness for skill or topic

  16. Formative assessments 2 
 • Check understanding, monitor progress during learning cycle 
 • Spot errors for re-teaching • Give feedback and suggestions 
 • Check readiness for final (summative) assessment (aka “posttest”)

  17. 7.4 Classroom Grading Approaches

  18. Assigning grades 1 
 When student gets a grade for work, what does he/she think it means? 
 • This is what I am worth • This how I compare with classmates 
 • This is what teacher thinks of me 
 • This is how well I learned

  19. Assigning grades 
 • Letter grades: A, B, C, D, F 
 • Absolute: 10 points per letter • Curve (relative): comparative scaling (force bell curve?) 
 • Descriptive (short or long) 
 • Performance rating (with rubric/criteria) 
 • Mastery checklist (# of attempts not important)

  20. 7.5 Criterion-Referenced Testing

  21. Criterion referencing 
 • Emphasis on mastery of specific skills/objectives 
 • Good for topics that can be broken into small objectives 
 • Good for topics that have hierarchy of skills (e.g. math) 
 • Must master skill A before you can understand and master skill B

  22. Criterion referencing 
 • Set-up: objective and performance criteria to prove mastery for each skill 
 (e.g. 80% correct answers) 
 • No comparisons (and no time constraints?) ➔ move to next level at own pace

  23. 7.6 Norm-Referenced Testing

  24. Norm referencing 
 • “Standardized” 
 • Comparative with other students • Achievement tests 
 (what has been learned, e.g. state/graduation test) 
 • Aptitude tests 
 (predict future success, e.g. IQ, SAT, GRE)

  25. 7.7 Interpreting Norm-Referenced Test Scores

  26. Analyzing test results (1) 
 • Raw scores ➔ derived (comparative) score • “Normed” with large samples of test-takers 
 • Norming = fitted onto normal distribution (bell curve) 
 • Bell curve: mean/average (skewed by extremes), median (middle #), and mode (most frequent) are same

  27. Analyzing test results (2) 
 Statistical descriptors 
 • Areas of distribution marked by standard deviations = deviations from average 
 • Example: IQ tests 100 = avg.; 34% either side of average 
 • Z-scores: # standard deviations +/- from average 
 • Stanines: #5 in center; 1-4 below, 6-9 above

  28. Analyzing test results (3) 
 More statistical descriptors 
 • Percentiles = % of students performing same or below 
 • Example: 80 th percentile = performs better than 80% of others 
 • Grade-level equivalents = 
 • Example: 3.4 = 3 rd grade, 4 th month

  29. 7.8 Validity

  30. How is a test valid? 
 • Validity: accuracy measure 
 • Content: match what was in curriculum • Face: appropriate format 
 • Criterion-related: items match objectives 
 • Predictive: match future performance 
 • Construct: match other tests

  31. 7.9 Reliability

  32. How is a test reliable? 
 • Reliability = consistency 
 • Test-retest • Alternate/parallel (versions) • Split-half = odds/evens • Kuder-Richardson = 1 test

  33. • Perfect = 1.0, but .80 OK 
 • 0 = no correlation • Negative value = as one factor goes up, other down

  34. 7.10 Test Bias

  35. Can a test be biased? 
 • If content or format favors one SES, race, culture, gender, or learning style 
 • Shows up in form/content of test question or answer • Partial solution: test in students’ native language 
 • Not bias: Males vary more than females in achievement scores

  36. 7.11 Using Tests Appropriately

  37. Testing: Use wisely 
 • Check validity and standard error of estimate (score +/-) 
 • Check reliability and standard error of measurement (confidence interval) caused by degree of unreliability • Consider how scores and results will be used

  38. 7.12 Summary

  39. Testing the test 
 • What are you trying to find out, and at what point in learning cycle? 
 • Does a test report skill achievement or compare students? • Does a test measure what it should, consistently and without bias to any learner?

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend