Essentials of Test Quality in Psychometrics

Essentials of Test Quality in Psychometrics
Slide Note
Embed
Share

Key factors that define a good test - reliability and validity. Learn about sources of error, types of reliability, and estimation methods in test assessment.

  • Test quality
  • Psychometrics
  • Reliability
  • Validity
  • Error sources

Uploaded on Mar 10, 2025 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

You are allowed to download the files provided on this website for personal or commercial use, subject to the condition that they are used lawfully. All files are the property of their respective owners.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.

E N D

Presentation Transcript


  1. What Makes A Test Good? Psychometrics

  2. What Makes a Test Good? A test must be: Reliable Valid Reliable = consistent Valid = measuring what it is suppose to measure

  3. Face Reliability and Validity Validity Reliability Content Test-Retest Criterion IC Alternative Construct Convergent Interrater Discriminant Factorial

  4. Reliability and Validity Reliable but not valid

  5. Reliability and Validity Neither reliable nor valid

  6. Reliability and Validity Reliable and valid

  7. Definition of Reliability reliability = refers to test's consistency A reliable test minimizes error and provides repeatable consistent results

  8. Sources of Error 1) low precision characteristics of the instrument used 2) state of the participants depressed, moody, irritable 3) state of the experimenter depressed, moody, irritable 4) state of environment noise, heat, etc.

  9. Classical Test Theory Observed score True ability Random error + = e + T

  10. Estimation of Reliability Correlations used Normally ranges from -1.0 to +1.0. Reliability coefficient is: 0 to 1.0

  11. Types of Reliability Test-retest Alternate Forms Internal Consistency Inter-rater

  12. Test-Retest Reliability Coefficient of stability is the correlation of the two sets of test scores.

  13. Alternate Forms Two versions of the same test with similar content Forms must be equal Coefficient of equivalence

  14. Internal Consistency Split half KR-20 Cronbach s alpha

  15. Inter-rater Reliability Measures scorer reliability Do the judges agree?

  16. Inter-rater reliability Scores by two independent judges and the scores are correlated Cohen s Kappa (k - Kappa Statistic)

  17. How Reliable? research r = .7 and .8 clinical settings GT r = .9

  18. Validity: Basic Concepts

  19. Reliability and Validity Validity Reliability Face Content Test-Retest Criterion IC Alternative Construct Convergent Discriminant Interrater Factorial

  20. Reliability and Validity Highly Reliable and Valid Highly Reliable but Not Valid Neither Reliable or Valid

  21. Overview Definition Face Validity Content Validity Criterion-Related Validity Construct Validity

  22. Definition of Validity The extent to which a test measures a trait that it claims to measure Involves Validation

  23. Face Validity superficial examination of domain A non-statistical approach established by non- experts Does the test "look valid?

  24. Content-related Validity Does the test assess the domain of the construct that is supposed to measure?

  25. Content-related Validity involves examination of the literature domain must be systematically analyzed consult with experts in the field

  26. Content Validity: Limitations biases Consider biases of the panelists level of expertise of the panelists Are they really experts?

  27. Criterion-related validity relationship between test and criterion Criterion (Y) (behavior) GPA SAT = Scholastic Achievement Test GPA = Grade Point Average Predictor (X) (test) - SAT

  28. Two Kinds of Criterion Validity Concurrent Predictive

  29. Concurrent Criterion-related Validity test and criterion are measured at approximately the same time

  30. Construct-related validity How does a test correlated with similar tests? Test Gold Standard Your Test

  31. Construct-related validity: Types Convergent validity Discriminant validity

  32. Convergent validation A test should correlate highly with another test that it is theoretically related to r = positive and strong Gold Standard WAIS WAIS = Wechsler Adult Intelligence Scale Your Test IQ Measure

  33. Discriminant validation A test ought not to correlate with a theoretically unrelated test Extraversion r = close to zero Your Test IQ Measure

  34. Multi-trait Multi-method Matrix Campbell & Fiske (1959) Correlation of: 2 or more traits by 2 or more methods Methods: e.g., self-report v. peer observations Traits: e.g., Extraversion and Warmth

  35. Multitrait Multimethod Es Ep As Ap Es MM (reliability) MH HM HH Ep MM HH HM As MM MH Ap MM

  36. Monotrait Monomethod (MM) reliability Same trait, same method High r, little error Low r, lots of error

  37. Monotrait Heteromethod (MH) Same trait, different method Validity coefficients expected to be high High r, evidence of construct validity Low r, indicates opposite

  38. Heterotrait Monomethod (HM) Different trait, same method Coefficients expected to be low to moderate high r, poor construct validity low r, good construct validity

  39. Heterotrait Heteromethod (HH) Different trait, different method Coefficients expected to be the lowest high r, poor construct validity low r, good construct validity

  40. Multitrait Multimethod Es Ep As Ap Es MM (reliability) MH HM HH Ep MM HH HM As MM MH Ap MM

Related


More Related Content