Psychological Assessment and Testing – Comprehensive Review

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/326

flashcard set

Earn XP

Description and Tags

A set of 500 English question-and-answer flashcards covering psychological testing, assessment principles, psychometrics, statistics, test construction, major tests, ethics, and related concepts for exam review.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

327 Terms

1
New cards

What does the term “psychological testing” refer to?

The process of measuring psychology-related variables by means of devices or procedures designed to obtain a sample of behaviour.

2
New cards

In psychological testing, what is typically the end product?

A numerical test score or series of scores used for evaluation.

3
New cards

Who is considered the key decision-maker in psychological assessment?

The assessor conducting the evaluation.

4
New cards

What is an ecological momentary assessment?

An “in-the-moment” evaluation of problems and related variables at the time and place they occur.

5
New cards

Define collaborative assessment.

An assessment approach in which the assessor and assessee work as partners from initial contact through feedback.

6
New cards

What is meant by therapeutic assessment?

An approach that encourages self-discovery and new understanding while assessment is conducted.

7
New cards

Describe a dynamic assessment.

An interactive approach that follows the model: evaluation → intervention → re-evaluation.

8
New cards

What is a psychological test?

A device or procedure designed to measure variables related to psychology.

9
New cards

List two basic elements every test item has.

A stimulus (the item itself) and a response to be scored.

10
New cards

What is a cut-score?

A reference point, derived by judgment, used to divide data into two or more classifications.

11
New cards

What does psychometric soundness refer to?

The technical quality and scientific adequacy of a test.

12
New cards

Who is a psychometrist?

A professional who uses, scores, and interprets psychological data.

13
New cards

Differentiate achievement and aptitude tests.

Achievement measures prior learning; aptitude estimates potential for future learning.

14
New cards

What does an intelligence test aim to measure?

General potential to solve problems, adapt, think abstractly, and learn from experience.

15
New cards

Define non-standardised interview.

An unstructured interview pursuing relevant ideas in depth without preset questions.

16
New cards

What is a semi-standardised interview?

An interview with optional probing on a limited set of prepared questions.

17
New cards

Explain the SORC model in behavioural observation.

Stimulus, Organismic variables, Response, and Consequence framework for analysing behaviour.

18
New cards

What makes a personality test ‘structured’?

It uses self-report statements and fixed response alternatives.

19
New cards

Give an example of a projective personality test.

Rorschach Inkblot Test or Thematic Apperception Test.

20
New cards

What is a speed test designed to evaluate?

How many items a test-taker can answer correctly within a time limit.

21
New cards

Contrast speed and power tests.

Speed emphasises time; power emphasises difficulty of items answered correctly.

22
New cards

Define a norm-referenced test.

A test that compares an individual’s score to a normative sample.

23
New cards

What is a criterion-referenced test?

A test that evaluates performance against a fixed standard or skill set.

24
New cards

What are the four main steps in the psychological assessment process?

Determine referral question, acquire knowledge, collect data, interpret data.

25
New cards

What is an actuarial assessment?

Evaluation relying on statistical rules rather than clinical judgment.

26
New cards

Describe mechanical prediction.

Use of computer algorithms plus statistics to generate assessment findings.

27
New cards

What is extra-test behaviour?

Observations about how an examinee behaves during testing that aid interpretation.

28
New cards

Define psychological trait.

A relatively enduring way one individual differs from another.

29
New cards

What is a psychological state?

A temporary pattern of thinking, feeling, or behaving in a specific situation.

30
New cards

What is cumulative scoring?

The assumption that endorsing more keyed responses indicates more of the trait.

31
New cards

Why is reliability important?

It indicates consistency and dependability of test scores across occasions or forms.

32
New cards

What is classical test theory’s basic equation?

Observed score = True score + Error.

33
New cards

Name three potential sources of measurement error.

Assessors, measuring instruments, and random events such as luck.

34
New cards

Differentiate random and systematic error.

Random error fluctuates unpredictably; systematic error is consistent and proportional.

35
New cards

What is test–retest reliability appropriate for?

Measuring stability of traits expected to remain constant over time.

36
New cards

Define carry-over effect.

When experience with the first test administration influences the second administration.

37
New cards

What statistic is usually used for test–retest correlation?

Pearson r or Spearman rho, depending on data level.

38
New cards

What is split-half reliability?

Correlation between two halves of one test administered once.

39
New cards

What formula adjusts split-half coefficients to full-length?

Spearman-Brown formula.

40
New cards

Explain Cronbach’s alpha.

An internal-consistency index for items scored in more than two categories.

41
New cards

What does KR-20 measure?

Internal consistency for dichotomous items with unequal difficulty.

42
New cards

When would KR-21 be preferred?

For dichotomous items all having equal difficulty.

43
New cards

What is parallel-forms reliability?

Correlation between scores on two versions with equal means and error variances.

44
New cards

Define inter-scorer reliability.

Agreement consistency between two or more scorers.

45
New cards

What is Fleiss’ kappa used for?

Inter-rater agreement for three or more raters on categorical data.

46
New cards

What does restriction of range do to reliability?

It lowers correlation coefficients and reliability estimates.

47
New cards

Explain standard error of measurement (SEM).

The standard deviation of error scores, reflecting score precision.

48
New cards

What is a confidence interval in testing?

A range around an observed score likely to contain the true score.

49
New cards

Define validity in testing.

The degree to which evidence supports intended interpretations of test scores.

50
New cards

What is content validity?

How well test items represent the construct’s domain.

51
New cards

Describe face validity.

Whether a test appears to measure what it claims to, to test-takers.

52
New cards

What problem does construct underrepresentation present?

The test fails to capture essential aspects of the construct.

53
New cards

What is construct-irrelevant variance?

Score variance due to factors unrelated to the construct of interest.

54
New cards

Name the formula often used for expert item ratings in content validation.

Lawshe’s Content Validity Ratio (CVR).

55
New cards

What is criterion validity?

How well test scores predict or relate to an external criterion.

56
New cards

Differentiate predictive and concurrent validity.

Predictive uses future criteria; concurrent uses simultaneous criteria.

57
New cards

Define sensitivity regarding a test.

The percentage of true positives correctly identified.

58
New cards

What does specificity measure?

The percentage of true negatives correctly identified.

59
New cards

What is incremental validity?

The additional predictive value gained by a new test beyond existing predictors.

60
New cards

Explain convergent evidence for construct validity.

High correlations with other measures of the same construct.

61
New cards

Explain discriminant evidence.

Low correlations with measures of different constructs.

62
New cards

What is a multitrait-multimethod matrix used for?

Simultaneously evaluating convergent and discriminant validity across traits and methods.

63
New cards

Who developed factor analysis?

Charles Spearman.

64
New cards

What is cross-validation?

Re-testing a test’s validity on a new independent sample.

65
New cards

Define validity shrinkage.

The tendency for validity coefficients to decrease after cross-validation.

66
New cards

What is a rating error called when scores cluster at the middle?

Central tendency error.

67
New cards

What is the halo effect?

Rater’s inability to discriminate among separate traits, giving uniformly high ratings.

68
New cards

Contrast leniency and severity errors.

Leniency inflates ratings; severity deflates them.

69
New cards

Define utility analysis.

Cost-benefit evaluation of a testing program’s practical value.

70
New cards

What is a Taylor-Russell table used for?

Estimating improvement in selection decisions when a test is added.

71
New cards

What is a selection ratio?

The proportion of applicants hired to those available.

72
New cards

Explain the Angoff method.

Experts estimate probability of minimally competent candidates answering each item correctly to set cut scores.

73
New cards

What does IRT stand for?

Item Response Theory.

74
New cards

In IRT, what does item discrimination indicate?

How well an item differentiates between high and low trait levels.

75
New cards

Define item difficulty in IRT.

The trait level where an examinee has a 50% chance of endorsing the keyed response.

76
New cards

What is computerized adaptive testing (CAT)?

Interactive testing where item selection is based on previous responses.

77
New cards

What are floor effects?

Scores clustering at the lower limit of a test.

78
New cards

What are ceiling effects?

Scores clustering at the upper limit of a test.

79
New cards

Name two primary scaling levels used in psychological measurement.

Nominal and ordinal, plus interval and ratio.

80
New cards

Give an example of a comparative scale.

Paired-comparison ranking of preferences.

81
New cards

What is a Likert scale?

A format where respondents rate agreement on ordered categories.

82
New cards

Define ipsative scale.

Respondents choose between options, producing intra-individual profiles.

83
New cards

What is pilot testing?

Preliminary study to refine items before full test development.

84
New cards

How many subjects per item are recommended for tryout?

At least 5, preferably up to 10.

85
New cards

What is item difficulty index (p)?

Proportion of test-takers who answered the item correctly.

86
New cards

Describe item discrimination index (D).

Difference in item success between high and low scoring groups.

87
New cards

What is the optimal average item difficulty for maximal discrimination?

Approximately 0.50.

88
New cards

What is the formula for optimal p on a four-option multiple choice item?

(1 + chance level) / 2 = (1 + 0.25)/2 = 0.625.

89
New cards

Define item reliability index.

Product of item-score SD and item-total correlation, reflecting internal consistency.

90
New cards

What is differential item functioning (DIF)?

An item showing different probabilities of endorsement across groups at equal trait levels.

91
New cards

What is scoring drift?

Gradual departure of scorers from an anchor protocol over time.

92
New cards

What is an anchor protocol?

A model scoring guide to resolve scoring discrepancies.

93
New cards

List the four measures of central tendency.

Mean, median, mode, and sometimes midrange.

94
New cards

When is median preferred over mean?

When the distribution is skewed or contains extreme scores.

95
New cards

Define variance.

The average squared deviation of scores from the mean.

96
New cards

What does standard deviation represent?

The square root of variance; average distance of scores from the mean.

97
New cards

What is a percentile rank?

Percentage of scores in a distribution that fall below a specific score.

98
New cards

Explain positive skew.

Distribution tail extends toward higher scores; mean > median > mode.

99
New cards

What is kurtosis?

The peakedness or flatness of a distribution’s centre.

100
New cards

What values describe a normal curve?

Mean = median = mode, symmetrical, mesokurtic.