Phase III Standardization

0.0(0)
studied byStudied by 0 people
0.0(0)
full-widthCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/28

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No study sessions yet.

29 Terms

1
New cards

What are the 7 steps in phase 3?

  1. Sampling plan

  2. Selecting appropriate respondents

  3. Specifying administration and scoring methods

  4. Piloting and revising tests

  5. Analyzing the Results

  6. Revising the test

  7. Validation and cross-validation

2
New cards

Step 1: Sampling Plan

  • Define target population (age, special needs, etc.) and comparison group.

  • Standardization sample: represents the population the test is intended for; determines norms.

  • Ideally use random sampling; sample must be representative

  • Best method: use population proportionate stratified random sampling (age, gender, SES, culture, education).

Determine appropriate sample size.

3
New cards

Step 2: Selecting Appropriate Respondents

  • Population: all target audience member 

  • Sample: Survey given to subset of population  

  • Probability sampling: Uses statistics to ensure sample is representative

  • Types: simple random, stratified random, cluster, systematic random.

  • Non-probability sampling: Sampling where not all have equal chance of being selected 

4
New cards

Types of Probability Sampling

  • Simple random sampling

  • Stratified random sampling

  • Cluster sampling

  • Systematic random sampling

5
New cards

Simple Random Sampling

Every member of population has an equal chance of selection to be in sample

6
New cards

Stratified Random Sampling

Population divided into subgroups (e.g., age, gender, SES).

7
New cards

Cluster sampling

Used when it’s not possible to list all individuals in a population; often used for surveys with large target populations.

  • The population is divided into clusters, then clusters are randomly selected.

8
New cards

Systematic Random Sampling

Select every nth person (e.g., every 5th).

9
New cards

Nonprobability Sampling

  • Convenience Sampling

    • Select any available participants; not all have equal chance.

10
New cards

Sample size

Refers to the number of people needed to represent the target population accurately.

11
New cards

Homogeneity of the population:


How similar the people in your population are to one another

  • The more dissimilar members are = more variation in sample

12
New cards


Sampling error

A statistic that shows how much error is due to lack of representation of the target population

13
New cards

Step 3: Specifying Administration & Scoring

  • Decide how the test is administered (oral, written, computer, group/individual).

  • Decide scoring method: hand, software, or publisher.

14
New cards

Raw scoring methods: Cumulative/Summative

  • Assumes that the more a test taker responds in a particular way, the more they have of the attribute being measured

    •  (e.g., more correct answers, or higher Likert scale ratings).

15
New cards

Raw scoring methods: Ipsative Model

Test takers are given 2 or more options to choose from 

  • Uses forced-choice items 

  • Typically yields nominal data because it puts test takers in categories (e.g., # of T/F, Y/N, Agree/Disagree).

    • Shows the test taker where they are relative to themselves 

    • Example: place an “X” next to the word in each pair that best describes your personality

16
New cards

Raw scoring methods: Categorical

Puts test taker in particular group/class 

  • Scores are not compared to other test takers, but compared within the test taker (which scores are high or low).

  • Yields nominal data, placing test takers in categories (e.g., # of T/F, Y/N, Agree/Disagree).

17
New cards

Step 4: Piloting and Revising Tests

Cant assume the test will perform as expected 

  • Pilot test investigates the test’s reliability and validity 

  • Administer test to sample from target audience 

  • Analyze data and revise test to fix any problems uncovered, many aspects to consider

18
New cards

Setting up the Pilot Test

Test situation should match actual circumstances in which test will be used (in sample characteristic setting) 

  • Must use apa code of ethics

19
New cards

Conducting the Pilot Test

Evaluates test performance 

  • Depth and breadth depends on the size and complexity of target audience and construct being measured

20
New cards

Step 5: Analyzing the Results 

Can gather both quant and qualitative info 

  • Use quantitative info for such things as item characteristics, internal consistency, test-retest, inter-rater convergent and discriminate validity and in some instances predictive validity 

  • Qualitative data may be used to help make decisions

21
New cards

Conducting Quantitative Item Analysis

How developers evaluate the performance of each test items

22
New cards

Conducting Quantitative Item difficulty (p-value)

The % of test takers who respond correctly - optimal ≈ 0.5.

23
New cards

Conducting Quantitative Discrimination index (D)

Compares how well an item separates high scorers from low scorers 

  • Formula:
    D = U (% in upper group correct) − L (% in lower group correct)


  • Desirable value: 30 and above.

24
New cards

Conducting Quantitative Item-total correlation

A measure of strength and direction of relation between the way test takers respond to one item and they way they respond to all items as a whole

  • 0.3 and above is desirable 

25
New cards

Interitem correlation matrix

Displays the correlation of each item with every other item.

  • Checks internal consistency; Phi coefficients for dichotomous items.

  • Phi coefficients: The results of correlating two dichotomous variables

26
New cards

Item-response theory (IRT)

  • Estimates a test taker’s ability regardless of how hard or easy the items are.

  • Estimates item difficulty and discrimination regardless of the ability of the people taking the test.

27
New cards

Item bias

Occurs when an item is easier for one group than another.

  • Test items should be equally difficult for all groups.

To eliminate bias use Item Characteristic Curves (ICCs) to evaluate.

28
New cards

STEP 6: Revising the Test

  • Finalize items based on content validity, difficulty, discrimination, inter-item correlation, and bias.

  • When new items need to be revised or add items → re-pilot must occur to ensure that changes produce desired results

29
New cards

Step 7: Validation and Cross-Validation

  • Validation: gathering evidence that the test reliably and accurately measures the intended construct.

  • Content validity: First checked during test development to ensure the test measures the right constructs (construct validity).

  • Criterion prediction: determined later through data collection.

  • Cross-validation: After final revisions show reliable and valid scores, the test is administered to a different sample to confirm results.