reliability

0.0(0)
studied byStudied by 0 people
call kaiCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/17

encourage image

There's no tags or description

Looks like no tags are added yet.

Last updated 8:23 PM on 1/21/26
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No analytics yet

Send a link to your students to track their progress

18 Terms

1
New cards

Reliability

-test is repeatable and will give the same results

-refers to the consistency and dependability of the data collection process

-within itself, overtime, over circumstances

2
New cards

Internal Consistency/Reliability (Within Itself)

-how consistently each item behaves relative to the overall tests

-how well a test measures what we want to measure

-harmony among different items within a measurement tool aiming to assess the same concept

-if I get a high score on a whole test, do I also get a high score on each item

-Cronbach alpha, split-half, item total correlations

3
New cards

Test-Retest (Overtime)

-the degree to which a test will produce the same result over repeated administration

-consistency of test response overtime: assuming there is no intervention/change to the thing being measured between tests

-shorter time period likely to have higher consistency

-the degree to which a test will produce the same results over repeated administration

-I did well last time, will I do well the next time?

-rank order correlations, intra-class correlation

4
New cards

Inter-Rater Reliability (Over Circumstances/Over Raters)

-the degree to which multiple raters will give the same results for a given target

-extend to which different raters produce consistent results when rating the same thing

-do mothers and fathers give their child the same score?

-Cohen kappa, intra-class correlation

5
New cards

Cronbach’s Alpha

-internal consistency/reliability

-how closely related a set of items are as a group

-value 0.7 is considered acceptable, with larger values indicating greater reliability

6
New cards

Split-Half

-internal consistency/reliability

-splits a test in 2 halves and measures the correlation between them

-higher correlation indicates greater internal consistency

-1 is considered perfect correlation

7
New cards

Item Total Correlations (Item Discrimination)

-internal consistency/reliability

-differentiates between a question score and the overall test scores

-0-0,19: indicate that the question is not discriminating well

-0.2-0.39: indicate good discrimination

-0.4+ indicates very good discrimination

8
New cards

Factors Impacting Internal Consistency

-sensitive to sample size: the bigger the sample, the more consistent it appears

-sensitive to number of items: more items the more reliable

-sensitive to difficulty: adding lots of “easy” or “difficult'“ times will decrease the reliability, but adding lots of “average” items will increase it

-gives the reliability of a test, not the reliability of the scores

9
New cards

Rank Order Correlations

-test-retest reliability

-do respondents stay in the same order overtime?

-less than 0.2: negligible

-0.2-0.29: weak

-0.3-0.39: moderate

-0.4-0.69: strong

-more than 0.70: very strong

10
New cards

Intra-Class Correlation (ICC)

-test-retest reliability

-estimates the proportion of variability in scores as a function time

-less than 0.5: poor

-0.5-0.75: moderate

-0.75-0.90: good

-more than 0.90: excellent

11
New cards

Factors Impacting Test-Retest Reliability

-time interval: shorter intervals tend to nil higher reliability but may introduce memory bias, while longer intervals risk chances in the measured construct

-test content: tests with ambiguous or poorly defined items are less likely to yield consist results

-participant factors: variations in participants conditions between administration

-environmental factors: difference in testing environments

12
New cards

Cohen’s Kappa

-inter-rater reliability

-measures how 2 raters agree with each other beyond what is expected by chance

-used to assess the reliability of comparative agreement between raters

13
New cards

Intra-Class Correlation (ICC)

-inter-rater reliability

-estimates the proportion of variability in test scores that are not a function of rater (1- rater effect)

14
New cards

Factors Impacting Inter-Rater Reliability

-subjectivity: raters may interpret criteria differently, especially for qualitative assessments

-chance agreement: methods like percent agreement may overestimate reliability by ignoring chance agreement

-rater bias: personal biases or fatigue can influence ratings, reducing reliability

-sample dependence: reliability estimates can vary based on the sample or context, limiting generalizability

15
New cards

Latent Variables

-provide detail on the mature of the relationship between items and constructs

-allow for scoring to be adjusted to increase the reliability of a measure

-allow for estimation go score reliability rather than scale reliability

-exploratory factor analysis, confirmatory factor analysis, item response theory

16
New cards

Exploratory Factor Analysis (EFA)

-data driven approach where number and structure of latent variables is unspecified

-often mistaken for principal components analysis

17
New cards

Confirmation Factor Analysis (CFA)

-theory driven approach

-number of latent variables and structure are specified by the researcher

18
New cards

Item Response Theory (IRT)

-theory driven approach similar to CFA (differences are largely technical)

-Rasch Model is commonly used

structural assumptions are not always met