Lecture 9 - Psychometric Methods

Validity Overview

  • Validity: Refers to the degree to which a test measures what it claims to measure, which is essential for ensuring that the test results are meaningful and applicable in real-world scenarios. Validity goes beyond mere accuracy, as it encompasses whether the test is effective in measuring both the intended constructs and the real-life situations they represent.

  • There are three traditional forms of validity that researchers typically consider:

  • Content Validity: Involves the systematic examination of the test content to ensure that it adequately covers the construct it intends to measure. It requires clear definitions and observable behaviors related to the test, ensuring that the test items are representative of the entire domain of the construct.

  • Criterion-related Validity: Tests intended to predict outcomes and focuses on how well test scores correlate with external criteria that are relevant to the construct being measured. This form of validity can be further divided into concurrent and predictive validity, depending on the timing of the measurement.

  • Construct Validity: Assesses the extent to which a test measures a theoretical construct, encompassing both the adequacy of the test in measuring the construct and the theoretical understanding underlying it. This form of validity is crucial for establishing the legitimacy of psychological theories through empirical testing.

Construct Validity
  • Definition: The extent to which a test measures a theoretical construct. Construct validity is developed in the 1950s to bridge the gap between psychological measurement and theoretical frameworks, addressing the lack of evidence linking tests to psychological theories. Evidence of construct validity lies in the acceptance and understanding of the underlying constructs by test developers and researchers.

Concrete vs. Abstract Constructs
  • Concrete Constructs: Observable and measurable traits or behaviors; behaviors that help infer deeper, abstract constructs. Examples include behaviors such as physical endurance or specific cognitive tasks.

  • Abstract Constructs: Hypothetical traits that are not directly observable and can only be inferred through a range of behaviors. Examples include intelligence or emotional resilience. Starting with a well-defined construct is crucial for gathering robust evidence of construct validity, as it offers a clear focus for interpreting test outcomes.

Steps for Establishing Construct Validity
  1. Construct Explication: Define the construct clearly, ensuring that the definition comprehensively encapsulates its scope and implications.

  2. Identify Observable Behaviors: Determine relevant behaviors that relate to the construct and can provide empirical evidence for its existence.

  3. Identify Related Constructs: Gain insight into other constructs that might inform the primary construct, creating a broader theoretical framework.

  4. Identify Behaviors Related to Other Constructs: Ensure a comprehensive understanding by exploring relationships with behaviors associated with other constructs, enriching the validity analysis.

Nomological Network
  • Nomological Network: A theoretical model that visually displays how different constructs and their associated behaviors are interrelated. This framework aids in forming coherent testable hypotheses, enabling empirical testing to validate theoretical assumptions.

Gathering Evidence for Construct Validity
  1. Theoretical Evidence:

  • Literature Review: Conducting a thorough exploration of theoretical and empirical literature that supports the test and hypotheses helps establish a solid nomological network of relationships.

  1. Psychometric Evidence:

  • Reliability: Ensuring test scores consistently measure the same constructs across different measurements, maintaining internal consistency and stability.

  • Convergent Evidence: Demonstrating a strong correlation with other established tests that measure similar constructs, thus reinforcing the validity of the test under consideration.

  • Discriminant Evidence: Establishing a weak correlation with unrelated constructs, demonstrating that the test is specifically measuring the intended variables without interference from other factors.

  • Experimental Interventions: Evaluating if the construct changes in expected ways over time and if different groups score differently, adding to the evidence of the test's validity.

Reliability
  • Reliability refers to a test's ability to consistently measure the same constructs over time and across various conditions. Reliability coefficients typically range from 0 to 1.0, with higher scores indicating stronger consistency and dependability of the test scores, which is fundamental for ensuring valid interpretations.

Convergent Validity
  • Convergent validity is recognized when scores on the test correlate strongly with scores on other tests that measure the same or similar constructs. A generally accepted threshold for convergence is Pearson's r > 0.30, indicating a significant relationship that supports the test's validity.

Discriminant Validity
  • Discriminant validity ensures that scores on a test do not correlate with unrelated constructs. Validity coefficients should showcase clear, non-significant correlations with constructs that are conceptually different, reinforcing the test's focused validity.

Multitrait-Multimethod Design
  • Trait = Construct: This design is used to validate measures of constructs across different methods and traits, providing robust evidence for the reliability and validity of the test through a comprehensive assessment approach.

Factor Analysis
  • Factor analysis is a statistical procedure that helps explain correlations among test items or measures, facilitating the identification of underlying factors.

  • Exploratory Factor Analysis: Conducted without formal hypotheses to uncover potential underlying factors in the data, generating insights into the dimensionality of the constructs.

  • Confirmatory Factor Analysis: Tests specific theories by confirming known constructs, utilizing model fit to validate the theoretical assumptions.

  • General Procedure: Involves inputting raw scores into specialized software for generating a correlation matrix, which is then used to identify factor solutions that adequately fit the data.