Measurement: Instrument Validity & Reliability

MEASUREMENT: INSTRUMENT VALIDITY & RELIABILITY

PUBH 475


Learning Objectives

  • Differentiate between validity and reliability
  • Explain 4 ways to establish instrument validity
  • Explain 4 ways to establish instrument reliability

PSYCHOMETRICS: VALIDITY & RELIABILITY


Validity and Reliability

Validity

  • Definition:
    • Refers to:
    • Research design (internal validity)
    • Ability to generalize based on sampling (external validity)
    • Accuracy of an instrument in measuring what it is supposed to measure

Reliability

  • Definition:
    • Refers to:
    • Consistency or stability of an instrument in measuring whatever it measures.

I. Validity

What is it?

  • Definition:
    • Degree to which an instrument measures what it is supposed to measure
  • Focus:
    • Accuracy
  • Methodologies:
    • Process involves qualitative and quantitative methodologies

REVIEW

  • Concept: Attach theory
  • Construct: Variable
  • Operationalization

1. Construct Validity

What is it?

  • Definition:
    • Extent to which an instrument or test measures the construct it is supposed to measure
  • Relevance:
    • Most relevant of the validities, testing the theoretical framework
  • Testing:
    • Tested rigorously on convergence and discrimination

2. Criterion Validity

What is it?

  • Definition:
    • Extent to which a new instrument or test correlates with an established test (standard or “criterion”) examining a similar theoretical construct
  • Testing:
    • Tested rigorously on its concurrent and predictive abilities

3. Content Validity

What is it?

  • Definition:
    • Extent to which an instrument or test adequately samples or captures relevant material of the construct
  • Testing:
    • Tested rigorously to see if items are redundant or unnecessary
    • Utilizes quantitative and qualitative methods to establish
  • Example:
    • Developing a “stress” measurement instrument

4. Face Validity

What is it?

  • Definition:
    • Extent to which an instrument or test will be easy for study participants to understand and complete correctly
  • Assessment:
    • Determined by experts and users
  • Example:
    • Depression measurement instrument

Ways of Establishing Instrument Validity

  1. Construct validity
    • Is the construct accurately measured by the measurement instrument?
  2. Criterion-related validity
    • Do the measurement instrument scores correlate with scores on some concrete standard or criterion in the real world?
  3. Content validity
    • Does the measurement instrument assess the full range of relevant phenomena?
  4. Face validity
    • Does the measurement instrument appear to cover the phenomenon and can be easily understood?

II. Reliability

What is it?

  • Definition:
    • Extent to which an instrument or test will produce the same or nearly the same result each time it is used
  • Focus:
    • Stability or consistency over time
  • Interpretation:
    • Less variation = more reliable

Ways of Establishing Instrument Reliability (1)

  1. Parallel Forms Reliability
    • Extent to which different forms of the same instrument or test produce the same result
    • Example:
      • Different versions of standardized tests
  2. Internal Consistency Reliability
    • Extent of intercorrelations between/among items of the same construct
    • Example:
      • Perceived Stress Scale (PSS): 10 items should “hang” together, each measuring a different part of stress
    • Statistical Measurement:
      • Measured (e.g., Cronbach’s alpha, $\alpha$) where a higher number implies greater internal consistency

Ways of Establishing Instrument Reliability (2)

  1. Test-Retest Reliability
    • Extent of stability of the same measurement over time
    • Example:
      • PSS administered twice, one month apart
  2. Rater Reliability
    • Extent of consistent measurement or rating within the rater (intrarater) or between raters (interrater)
    • Example:
      • Multiple trainers administering physical activity intervention
    • Statistical Measurement:
      • Measured (e.g., Kappa coefficient, $\kappa$) where a higher number indicates greater rater reliability

Putting it Together…

  • Reliable and Not Valid:
    • Low Validity, Low Reliability: Not Reliable, Not Valid
    • Both Reliable and Valid

Coming Up…

  • Next topic: Questionnaire development (asynchronous)
  • Due date: Midterm #2