Test retest reliability is the stability of scores of an instrument when
Stability of scores of an instrument when administered more than once to same participants under similiar coniditons, can be 2 minutes apart, 2 weeks, 2 years
11
New cards
Does the instrument product the same results with repeated tested, is an example of
Stability - Test retest reliability
12
New cards
Trying to assess fairly stable attributes
Test retest reliability
13
New cards
Higher value of Cronbach alpha reflect
Higher internal consistency (0.7 and above) thus they are more accurate in terms of measuring the same characteristic
14
New cards
A cronbach's alpha of 0.5 would mean
Lower internal consistenct
15
New cards
A low reliability score indicates that items on the instrument are
Not well correlated and therefore item is unreliable which jeopardizes study findings
16
New cards
A researcher reports a cronbach's alpha of 0.95 for efficacy measurment scale, What form of reliability?
Some attributes change over time E.g knowledge, stress, Participants can respond differently since they have experienced testing once
18
New cards
Parallel or alternate form reliability
This is where you use differently worded items when you’re retesting for test retest reliability
19
New cards
Internal consistency/homogeneity is asking if items on a survey
Do the items on a survey belong together and measure the same concept/characteristics? Items in a scale reflect measure the same concept.
20
New cards
Equivalence: Interrater reliability is asking does the tool
Would the tool produce the same results when equivalent procedures are used?
21
New cards
Internal consistency example
Depression scale - measures characteristics of depression
22
New cards
Test rest reliability advantage
Some traits we don't want to see consistency over time
23
New cards
This is the most commonly used test of internal consitency/homogenity
Cronbach's alpha: each item in scale is comapred with oteher items in scale and a total score is used to analyze data
24
New cards
Cronbach's alpha measures the ability to ensure that
It measures the ability to ensure that everything in your tool and all the items are measuring the same thing
25
New cards
Interrater or Interobserver Reliability is also called
Equivalence
26
New cards
consistency or agreement among 2 or more observers who use the same measurement tool
Equivalence (interrater reliability)
27
New cards
consistency or agreement among 1 observers who use the same measurement tool
Intra rater reliability
28
New cards
Equivalence has to do with the realibility of the ____ and not ___
It has to do with the reliability of the observer and not the instrument
29
New cards
Study on facial expressions of infants undergoing painful procedures in hospital, how can researchers estabilish interrater reliability
measuring the facial changes the same way according to the structured instrument, must be done consistently
30
New cards
Equivalence is used with strucutred instruments mostly used to observe
Used to observe behaviours
31
New cards
Reliability coefficient are used to describe
Reliability values in a numerical value, ranges form 0-1
32
New cards
Reliability coefficient ranges from
0-1
33
New cards
Often designtaed as r
reliability coefficient
34
New cards
The closer the reliability coefficient to 1 indicates
higher the reliability of measure
35
New cards
1 interrater reliability means
Perfect reliability among 2 observers
36
New cards
0 interrater reliability means
no reliability
37
New cards
A reliability coefficient of _____ or higher is considered acceptable
0.7
38
New cards
Internal consistency is a type of
Reliability
39
New cards
Refers to whether a measurment instrument measures what it intends to measure. (accuracy)
Validity
40
New cards
3 major kinds of validity (3c)
Content with subtyle face validity, criterion related, construct validity,
41
New cards
Used in development of measurments tool
Content validity: Determines wheter measurment tool and items in measure represent content reseaerch intends to measure
42
New cards
How is content validity carried out
Items in the tool are given to experts on the content and experts are asked to judge whether items reflect content or not
43
New cards
Content validity index (CVI) is used to measure
Central validity index is used to measure the relevance of each item based on expert's judgment (1-4)
44
New cards
CVI of 1 would indicate
Not relevant at all to content
45
New cards
2 types of criterion related validity
Predictive validity and concurent validity
46
New cards
Criterion related validity measures relationship between
Scores from the instrument with external criteria (gold standard), we want to see them related to gold standard measures and if they measure same attribute
47
New cards
Construct validity
Extent to which a test measures a theoretical construct of trait
48
New cards
Face validity
Subtype of content validity - does instrument appear to be measuring appropriate concept
49
New cards
How is face validity carried out
Colleagues are used to review content of instrument as to whether it reflects the concept the researcher intends to measure - based on judgment, not as accurate
50
New cards
With face validity, we report a stastical numver T/F
F, With face validity, you tend not to report any statistical number, not rigorous
51
New cards
Gold standard measures are
Ideal way to measure something, usually not achieveable due to fesiability and time restriants
52
New cards
Should we use face validity as an alternative to other validity
It’s not considered an alternative to other validity
53
New cards
Predictive validity
Correlate measure of concept and future meausre of same concept, helping predict future performance
54
New cards
Does pap smear predict cancer rates, does high school grades predict post secondary GPA
Predictive validity
55
New cards
Concurrent validity
Correlation of 2 measures of same construct administed at same time. High correlation indicates high agreement.
56
New cards
Pulse oximeter compared with arterial blood gas (gold standard)
Concurrent validity
57
New cards
What would be the gold stadnard of postpartium depression tool
Diagnosis by psychatrist
58
New cards
A measurement can be reliable but not valid
Researcher measures anxiety
59
New cards
Systematic error or constant error
Attributed to relatively stable characteristics of the study population that bias behaviour, who is entering my study and biases researcher may have
60
New cards
Social desirability
Bias where participants respond to what researchers want
61
New cards
Construct validity is the extent to which a test measures
Theory or theoretical construct or trait by testing different hypothesis relatinoships
62
New cards
Construct validity: hypothesis testing approach
Researcher develops hypothesis about behaviors of individuals with varying scores of a measure, then gather data to test hypotheses
63
New cards
Construct validity: Convergent
2 or more tools measuring same concept. E.g 2 depression tools will both measure depression
64
New cards
Construct validity: Divergent
Tool on one concept differentiates form another concept e.g tool on depression is different from tool on measuring stress
65
New cards
Construct validity: Contrast group/known group
2 known groups expected to score either high or low on measurement tool and they both fully use the tool, research then sees if we can distinguish E.g highly athletic vs not athletic uses tool regarding self perception of the athleticism
66
New cards
Construct validity: Factor analysis
Assessing stastically whether itesmc cluster around main dimension or main theoretical contrcut we want to assess
67
New cards
A measure can be reliable but not valid E.g
E.g research measures anxiety by measuring temperatures, however temperature would not be a valid indicator of anxiety
68
New cards
A valid instrument is always reliable E.g
A instrument that is erractic, inconsistent an dinaccurate cannot validly measure the attribute
69
New cards
We can't have validity without
Reliability
70
New cards
Reliable but not valid example
Always getting the weight, but it is off by 2 pounds each time
71
New cards
Valid but not reliable
Getting the correct weight one time, but not all the time. NOTE: If something is not reliabile that it is not truly valid
72
New cards
Not reliable not valid example
Never getting the right weight, getting different weight each time
73
New cards
Reliable valid
Getting the correct weigh each time and consistently
74
New cards
Measurment errors
difference between what exists in reality and what is being measured by an instrument
75
New cards
Measurment error: Random or chance
Unpredictable and difficult to control errors, cannot be corrected
76
New cards
Examples of random or chance error (4)
Transient human condition (hunger, fatigue), variance in measurement procedure (room temp, misplacement of BP cuff), error in data processing, scores varying in no particular direction
77
New cards
Measurment error: Systematic error or constant error (3)
Erorrs due to bias or improperly calibrated instrument, or something else being measured in addition to concept
78
New cards
Systematic error or constant error examples (3)
Socioeconomic status, weight scale not being calibrated, measurement error is not random thus measuring something else in addition to concept
79
New cards
Questions to ask when critquing data quality (3)
Appropriate methods, is reliability adequate (above 0.7), is validity adequate
80
New cards
LEVELS OF MEASURMENT
81
New cards
To identify levels of measurment in a research study
82
New cards
Measurment
Numbers assigned to variables or events based on a set of rules
83
New cards
Norminal measurement
Dichotomus (2 true values) or categorical (more than 2 values). Mutually exclusive (true or false) Researcher assigns number to each category but numvers are meaningless
84
New cards
mutually exclusive means if you have 1 fav color,
you can't have a second
85
New cards
Ordinal measurment
Categorical, variables given a number that can be placed in meanigful numical order
86
New cards
Example of ordinal measurment
Stages of cancer diagnosis, stage II is higher than 0 but we cannot say they are twice as severe as stage 0.
87
New cards
Ordinal measurement: do we know the distance between variables
No
88
New cards
Interval measurment
Variables are continuous and rank ordered. Equal intervals between numbers. Zero is arbitarily assigned but not absolute
89
New cards
level of measurement commonly used for psychological and psychosocial variables and tests
Interval because there is no true 0, just cuz someone scores 0 on IQ test doesn't mean they have no IQ
90
New cards
Example of interval
Temperature on centigrade scale. Difference btween 20 and 30 is and 30 to 40 is equivalent. Zero does not represent absence of temperature
91
New cards
What differentiates ratio and interval
Zero on interval data is arbitarily
92
New cards
Stastical potential for using interval level
We can caculate mode, average or mean
93
New cards
What is the highest lvel of measurment
Ration
94
New cards
Ratio measurment
Variables are continour and rank ordered and have absolute meanigful zero. Distance btween values are numerically equal
95
New cards
The levels of measurments helps determine
Type of stastitical analysis to use when analyzing data
96
New cards
What graph or chart to represent nominal and ordinal
Bar graph/chart
97
New cards
What graph to represent interval or ration
Histogram
98
New cards
Mneomic to remember levels of measurment
NOIR
99
New cards
Examples orf ratio level
many physical measures: Weight, height, hemoglobin levels, pulse rate, BP, age in years, time
100
New cards
The higher the level of measurment
The more flexibility in choosing stasticial procedures (interval and ratio) therefore maximum information can be obtained from data