1/49
Looks like no tags are added yet.
Name | Mastery | Learn | Test | Matching | Spaced |
---|
No study sessions yet.
Dependent/Paired T-test
If it's the same sample, where they are both in the experimental group and the controlled group, then it is better to use the ____.
Error
If you have measured something that you were not planning to measure, and you are not supposed to measure, then that is an ___.
Measurement Error
Refers to all factors associated with the process of measuring some variable, other than the variable being measured
Error
Everything that is associated with measuring a variable that you weren't actually intending to measure is considered an
Types of Error
Random Error
Systematic Error
Random Error
Caused by unpredictable fluctuations and inconsistencies of other variables in the measurement process
Random Error
Error is unpredictable. You didn't mean to make it.
Example: Someone caused a disturbance while you were taking the exam.
Systematic Error
Typically constant or proportionate to what is presumed to be true value of the variable being measured.
Systematic Error
Systematic
It does not affect the score consistency.
Usually, when discovered, it can be corrected.
Example: Imagine you were given a 12-inch ruler to measure objects that are supposed to be the same length as the ruler. However, you later find out that the ruler you bought is actually 13 inches. Now that you know your ruler is 13 inches, for everything you measure that was supposed to be the same size as the ruler, you just need to subtract 1 inch to make it 12 inches.
4o mini
Dynamic assessment
Pretest-intervention-posttest design is an example of
Remote assessment
The use of tools of psychological evaluation to gather data and draw conclusions about a subject who is not in physical proximity to the person or people conducting the evaluation
Nomothetic approach; psychological testing
Barnum effect is a known limitation of
Test Construction
Gina was hired by a company for a test development project. She asked 7 experts to validate the items. She is in what phase of test development?
Reliability
This refers to the consistency of test scores in a measurement.
Reliability
Consistency
Example: You conduct Test 1 and get a set of scores. You then conduct another test and get another set of scores. The scores should be consistent.
Reliability
It is not the consistency of the scores themselves, but rather the consistency of the scores across tests.
Example: Height: You want to know if the height measuring scale is reliable. So today, you measure your height and find that you are 5'3". To know if the height measuring scale is reliable, you should measure your height again tomorrow, and if the same height is recorded, the measurement should be consistent.
Katatagan
At UP, based on their studies, reliability is referred to as ____—so if your test is reliable, then your test is matatag (stable/consistent).
Test-retest reliability
It requires the administration of the same test to the same people twice
Test Conceptualization > Test Construction > Test Try Out > Item Analysis > Test Revision
What is the correct sequence of Test Development
Alternate forms reliability
Different tests, two administration, same sample
Split-Half reliability
Same test, one administration
Inter-Rater reliability
Same test, different raters, at the same time.
Test-Retest Reliability
Estimate of reliability by correlating pairs of scores from the same people on two different administrations of the same test.
Appropriate for: tests that measure a construct that is relatively stable over time and tests that employ outcome measures (e.g., reaction time, grip test)
Test-Retest Reliability
Appropriate for: tests that measure a construct that is relatively stable over time and tests that employ outcome measures (e.g., reaction time, grip test)
Test-Retest Reliability
Estimate of reliability by correlating pairs of scores from the same people on two different administrations of the same test
Alternate forms reliability
You are conducting the test twice, and you administer it to the same sample or group of people. However, you are using two different forms of the test (Set A or Set B).
Split-Half reliability
You are using the same test only once, you are just dividing it.
Test-Retest Reliability
We are estimating the reliability of our test by correlating the scores from the first administration and the second administration of the test.
You have the same test, you administer it to the same people twice, and then the data set or scores from the two tests you conducted will be correlated with each other to determine if the scores are consistent. If they are, then the test is considered reliable.
Test-Retest Reliability
is best used if we are testing the reliability of a measure where the outcome measures are employed, such as reaction time or grip strength tests.
Test-Retest Reliability
It is best for outcome-based tests.
Example: Grip test - This test measures how strong your grip is, and since it's outcome-based, if you want to know if your grip test is reliable or consistent, it's better to use ____.
First test: You take the grip test and get a score.
Second test: You take the grip test again and get another score. The scores you get from the two grip test administrations are then compared with each other to see if the scores are consistent.
KR21
Donny wants to establish the reliability of his newly constructed test. It is a multiple-choice type of test with only one correct answer, further he made sure that all of the items have a p = 0.50. What should Donny use?
item difficulty index
p value refers to the
item difficulty index
In the context of test development, p refers to the
KR20
KR21
Split-Half Reliability
Cronbach's Alpha
Measure of Internal Consistency
KR20
Or Kuder-Richardson 20
Calculates the reliability of a test in which the items are dichotomous
KR20
Or Kuder-Richardson 20
KR20
Calculates the reliability of a test in which the items are dichotomous
KR20
The items of the test are dichotomous.
The difficulty level of the items is not equal—meaning some items are easier while others are harder.
KR20
In practice, in real life, ___ is commonly used because it's difficult to create a test with the same level of difficulty for all items.
In tests with right and wrong answers, the difficulty levels are often unequal.
Measure of Internal Consistency
Or inter-item consistency
Measures of internal consistency
How consistent are the items with each other?
Example: Suppose there are 10 items in a test. You look at the items or answers to see how consistent they are with each other.
Measure of Internal Consistency
These include: KR20, KR21, Split-Half Reliability, Cronbach's Alpha.
KR21
Calculates the reliability of a test in which the items are dichotomous with equal difficulty
KR21
The items of the test are dichotomous.
The difficulty level of the items is equal—meaning all items have the same level of difficulty.
KR20 and KR21
____ are used when the items are dichotomous (one correct answer, the other is wrong).
Example: True or False / Yes or No.
dichotomous
Multiple-choice questions are still considered _____ if there is only one correct answer (because one option is correct, and everything else is wrong, making it dichotomous).
Psychology act of 2009
RA10029 refers to
RA10029
Psychology act of 2009 is
Coefficient Alpha
Or cronbach's alpha
Is the mean of all possible split-half correlations, corrected by the Spearman Brown formula
Useful for tests with non-dichotomous items (e.g., Likert Scale)
Received the most acceptance and is widely used today
Ranges from 0 to 1
Coefficient Alpha
Is the mean of all possible split-half correlations, corrected by the Spearman Brown formula
Coefficient Alpha
Useful for tests with non-dichotomous items (e.g., Likert Scale)