PSYC3020 - Exam 3

0.0(0)
studied byStudied by 0 people
call kaiCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/95

encourage image

There's no tags or description

Looks like no tags are added yet.

Last updated 1:54 PM on 11/6/25
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No analytics yet

Send a link to your students to track their progress

96 Terms

1
New cards

explain some of the major laws that govern the use of tests and measurement in the United States

2
New cards

compare the motivations and consequences of various lawsuits related to testing

3
New cards

contrast the motivations and consequences of various lawsuits related to testing

4
New cards

list several important ethical rules of conduct for measurement folks

5
New cards

Individuals with Disabilities Education Act (IDEA, 1997)

  1. all children are entitled to a free and appropriate public education

  2. testing used primarily to help place children in correct programs and measure progress

  3. each child given an Individual Education Program (IEP)

  4. children with disabilities educated in the least restrictive environment

  5. both students and parents are involved in decision-making

  6. mechanisms needed to ensure the above five principles

6
New cards

Truth in Testing Law (1979)

  • passed after investigation of Education Testing Services by New York Public Interest Research Group

  • require testing companies:

    • disclose all validity studies of a test

    • full disclosure on meaning of scores and how they’re calculated

    • provide copies of test questions, correct answers, and student’s answers if the student requests it

7
New cards

No Child Left Behind (NCLB, 2002)

  • assure that all children meet or exceed their state’s level of academic achievement

  • issues:

    • standardized achievement tests used to measure “school performance”

    • unrealistic standards

    • everyone must be tested at grade-level

    • expensive with limited funding

    • only applied to public schools

8
New cards

Every Student Succeeds Act (ESSA, 2015)

  • revised NCLB

    • relaxes requirements about testing every student in a school

    • extra funding and interventions for high schools with more needs

    • states given more control over how to help students and schools in needs

  • issues:

    • still using standardized achievement testing to evaluate school performance

    • still only applied to public schools

    • still limited in funding

9
New cards

Family Educational Rights and Privacy Act (FERPA)

  1. parents/students can inspect student’s education records, but schools are not required to provide copies

  2. parents/students can request records be corrected

  3. schools need written permission from parents/students to release any information

  • exceptions:

    • school officials

    • destination schools after tranfser

    • specified officials for audit/evaluation

    • etc.

10
New cards

What was Hobson v. Hansen (1967) about?

  • standardized tests used to place students in different learning tracks

  • African American children were disproportionately placed into based tracks while white children moved to other tracks

11
New cards

What was the consequence in Hobson v. Hansen (1967)?

  • grouping would be permissible if based on innate ability

  • problem: tests used were influenced by cultural experiences

12
New cards

What was Diana v. State Board of Education about?

  • intelligence tests used to place students in EMR tracks

  • problematic for bilingual children

  • tests were standardized for only white children

13
New cards

What was the consequence in Diana v. State Board of Education?

  • further research revealed bilingual children receive higher IQs if tested in their primary language

14
New cards

What was Larry P. v. Wilson Riles (1979) about?

  • 1/6th of African American elementary-school children tracked to EMR classes based on IQ scores

  • side 1 argued:

    • retesting done by African American psychologists yielded higher IQ scores, EMR placement was detrimental long-term

  • side 2 argued:

    • IQ scores were valid and unbiased

    • retesting was not standardized

15
New cards

What was the consequence in Larry P. v. Wilson Riles (1979)?

  • practice of IQ tests for EMR tracking ended

  • mixed feelings about outcome

16
New cards

What was Parents in Action on Special Education v. Hannon (1980) about?

  • racial bias only found for a subset of items on teh WISC, WISC-R, and Stanford-Binet IQ tests

17
New cards

What was the consequence in Parents in Action on Special Education v. Hannon (1980)?

  • racial bias findings didn’t justify removal of tests

  • conflicted with Larry P. v. Wilson Riles

18
New cards

What was Griggs v. Duke Power Company about?

  • raised concerns about segregation in the workplace

  • company claimed education was needed for advancement and created a test

    • nobody passed this test 

    • concerns about validity of the test

19
New cards

What was the consequence in Griggs v. Duke Power company?

  • employment test results must be valid and reliable

20
New cards

What was Watson v. Fort Worth Bank and Trust about?

  • misrepresentation of African American personnel

  • passed over for promotion multiple times

21
New cards

What was the consequence in Watson v. Fort Worth Bank and Trust?

  • lower courts argued that statistical bias only applied to psychological tests

  • Supreme Court disagreed

22
New cards

key ethical principles

  1. no physical, emotional, or psychological harm

  2. consent is important

  3. reasonable and appropriate incentives

  4. responses are made anonymous

  5. confidentiality must be ensured

  6. careful reporting of information

  7. use of appropriate assessment techniques

  8. test scores must be sufficiently valid and reliable

  9. tests should have a purpose

23
New cards

define test fairness described by psychometricians

24
New cards

define test bias as described by psychometricians

25
New cards

compare test fairness and bias

26
New cards

contrast test fairness and bias

27
New cards

describe how threats to test fairness weaken validity arguments

28
New cards

describe the ways test developers ensure test fairness

29
New cards

principles for making assessments using universal design

30
New cards

describe various ways of detecting bias and their limitations

31
New cards

test fairness

  • validity issue combining morality, philosophy, and legality

32
New cards

What are views on test fairness?

  • equitable treatment during testing

  • accessibility to the measured constructs

  • validity of individual test score interpretations for intended uses

  • lack of measurement bias

33
New cards

What is equitable treatment composed of?

  • standardization and consistency of administration

  • qualified test administrators

  • flexibility

34
New cards

What is accessibility composed of?

  • respondents can accurately record their responses

  • congruence of construct intended to be measured

  • congruence of constructs needed to respond to the measure

35
New cards

What is validity of individual test scores composed of?

  • heterogeneity within groups

    • group level accommodations or modifications are not always appropriate

36
New cards

What are the kinds of threats to fairness?

  • content

  • context

  • response process

  • lack of opportunity to learn

37
New cards

What is content? How does it threaten fairness?

  • problems with words or vocabulary inside an item

    • terms may be more likely known by another group

    • offensive language

    • representativeness within a question

38
New cards

What is context? How does it threaten fairness?

  • problems surrounding a test or measurement 

    • stereotype threat

    • unclear instruction

    • advanced or unfamiliar language

    • differential treatment

39
New cards

What is response process? How does it threaten fairness?

  • problems with the processes used to take in an item, process, and respond

  • occur between test-taker and item

    • faking good

    • misinterpreted communication

    • lack of accessibility

40
New cards

What is lack of opportunity to learn? How does it threaten fairness?

41
New cards

universal design

  • assessment development approach that maximizes accessibility of the test for all of the intended takers

  • begins by defining constructs precisely with clear differentiation from construct-irrelevant invariance

42
New cards

What are best practices for content and wording?

  • test takers share the same experience

  • appropriate complexity of sentences and vocabulary

  • shorter sentences

43
New cards

What are best practices for formatting?

  • text formatting

  • typefaces

  • white space

  • contrast

  • illustrations

44
New cards

accommodation

  • changes made to a test to improve accessibility

  • doesn’t affect the measured construct

45
New cards

modifications

  • changes made to a test to improve accessibiility

  • does affect the measured construct

46
New cards

When are accommodations necessary?

  • not appropriate if the affected ability is directly relevant to the construct being measured

  • not appropriate for an assessment if the purpose of the test is to assess the presence and degree of the disability

  • not necessary for all students with disabilities

47
New cards

test bias

  • system difference in scores between groups due to some unrelated factor

  • empirical observation

48
New cards

What are total-scores test bias?

  • difference-difference bias

  • cleary model

49
New cards

What is item-specific test bias?

  • content evaluation

  • differential item functioning (DIF)

50
New cards

difference-difference bias

  • bias evidenced by differences in scores among groups

51
New cards

cleary model

  • test scores are unbiased if equivalent scores from different groups equally predict some criterion

  • linear regression with interaction effect

52
New cards

problems with cleary model

  • assumes all relevant predictors or covariates are included

  • assumed unbiased criterion

  • interaction/model tests are often underpowered

53
New cards

content examination

  • review items for obvious cultural, racial, or gender related bias 

54
New cards

differential item functioning (DIF)

  • occurs when an item behaves differently among groups

  • respondents from different groups equated across scores

55
New cards

Mantel-Haenszel Test

  • simplest approach for detecting DIF

  • procedure:

    • group respondents into score groups

    • create a contingency table, for each score range group, of incorrect/correct responses and comparison group membership

    • calculate expected counts and variances of counts within each score range group

    • use all information to calculate chi-square statistic

  • limitation: how are ranges picked

  • advantage: can be used with smaller samples

56
New cards

parent model

  • allow item parameters to differ between groups

57
New cards

nested models

  • constrain a single item’s parameters to be equal across groups

58
New cards

bias and fairness

  • evidence of test bias does not mean a test is unfair

  • DIF may be detected, but might not cause impactful differences in scores

59
New cards

understand how to create table of specifications for item development

60
New cards

understand how to use a table of specifications for item development

61
New cards

describe different item formats

62
New cards

what kind of tests are different item formats suited for

63
New cards

describe and write good achievement test items

64
New cards

be familiar with Bloom’s taxonomy

65
New cards

alternative ways of defining the cognitive demands of items

66
New cards

describe and write good survey items

67
New cards

How is a test made?

  • choose and define constructs

  • determine the best method to use

  • develop possible questions or items

68
New cards

What are types of tests or measurements?

  • achievement

  • aptitude

  • ability or intelligence

  • personality

  • neuropsychology

  • career interests

69
New cards

achievement test

  • assess an individual’s level of knowledge in a particular domain

70
New cards

aptitude test

  • measure an individual’s potential to succeed in an activity requiring a particular skill or set of skills and can predict future outcomes

71
New cards

ability or intelligence test

  • assess one’s level of skill or competence in a wide variety of areas

72
New cards

personality test

  • assess an individual’s unique and table set of characteristics, traits, or attitudes

73
New cards

neuropsychological test 

  • asses the functioning of the brain as it relates to everyday behaviors, including emotions and thinking

74
New cards

vocational or career test

assess and individual’s interests and help classify those interests as they relate to particular jobs and careers

75
New cards

content/table of specifications

  • define the construct or content domain you are measuring in excruciating detail

76
New cards

What must be defined in clinical/psychological assessments?

  • define the construct and describe the associated observable behaviors

77
New cards

What must be defined in organizational tests?

  • define the knowledge and skills needed to do a job successfully

78
New cards

What must be defined in educational assessments?

  • describe the curriculum to be assessed

79
New cards

What are selected-response formats?

  • Likert format

  • category format

  • multiple-choice

80
New cards

What are constructed-response formats?

  • essay questions

  • interview questions

  • performance assessment

81
New cards

Likert format

  • people presented with a statement and asked to use a rating scale to respond according to the anchor

82
New cards

anchor

  • labels for different positions on the Likert scale

83
New cards

category format

  • rating scale between a and b

84
New cards

How might category format lead to reliability and validity issues?

  • too many categories

85
New cards

guidelines for survey items

  1. every item is important and requires a response

    1. item should apply to all respondents unless filter questions are used to exclude a participant

  2. avoid double-barreled items

  3. item should be technically accurate

  4. item should be a complete question or sentence with a simple structure

  5. use as few words as possible in each item stem and options

  6. use simple, familiar words

  7. use specific, concrete words to specify concepts clearly

  8. avoid negatively worded or connotative inconsistent items and options

  9. avoid leading or loaded items that suggest an appropriate response

86
New cards

guidelines for ordinal scales

  1. balance the item stem

  2. choose an appropriate rating scale length

  3. avoid the middle or neutral category

  4. provide balanced scales where categories are relatively equal distance apart conceptually

  5. verbally label all response categories

  6. align response options in one column (single item) or horizontally on one row (multiple items)

  7. response categories should be exhaustive, including all plausible responses

  8. response categories should be mutually exclusive

  9. response categories should approximate the actual distribution of the characteristic in the population

87
New cards

guidelines for nominal scales

  1. avoid the “other” option

  2. use forced-choice items instead of check-all-that-apply items

88
New cards

multiple choice

  • stem

  • options

    • correct: correct answer or key

    • incorrect: distractors, foils, or misleads

89
New cards

Bloom’s taxonomy

  • used to gauge the cognitive demand of test items

90
New cards

What are parts of the cognitive dimension of Bloom’s taxonomy?

  • remember

  • understand

  • apply

  • analyze

  • evaluate

  • create

91
New cards

cognitive demands

  • poor reliability in labeling questions using Bloom’s taxonomy

  • revised taxonomy:

    • recall

    • comprehend

    • use (or apply)

92
New cards

guidelines for achievement tests

  • content concerns

  • style concerns

  • writing the stem

  • writing the options

93
New cards

content concerns

  1. base each item on one type of content and cognitive demand

  2. use new material to elicit higher-level thinking

  3. keep the content of items independent of one another

  4. avoid opinions unless qualified

94
New cards

style concerns

  1. edit and proof items

  2. keep linguistic complexity appropriate to the group being tested

  3. minimize the amount of reading in each item

95
New cards

writing the stem

  1. state the central idea clearly and concisely in the stem and not in the options

  2. word the stem positively, avoid negative phrasing

96
New cards

writing the options

  1. use only options that are plausible and discriminating

  2. make sure that only one of these options is the right answer

  3. place options in logical or numerical order

  4. keep options independent

  5. avoid using options none-of-the-above, all-of-the-above, or I don’t know

  6. word options positively, avoiding negative words such as NOT

  7. avoid giving clues to the right answer

  8. make all distractors plausible

  9. avoid the use of humor