Psychological Testing and Assessment – Chapter 1
Psychological Testing vs. Psychological Assessment
Psychological Testing
Process of measuring psychology–related variables through standardized devices/procedures.
Goal: obtain a score (usually numerical) that serves as a sample of behaviour.
Psychological Assessment
Broader, problem-solving process that integrates multiple data sources (tests, interviews, observation, case files, special apparatus).
Seeks to answer referral questions, form evaluations, diagnose, or guide intervention.
Distinctions (Table 1-1-style)
Objective
Testing: gauge an attribute quantitatively.
Assessment: answer a complex question/decision.
Process
Testing: often group‐based; little focus on how answers produced.
Assessment: individualized; emphasises process (e.g., error patterns, strategies).
Evaluator’s role & skill
Tester ≈ technician; interchangeable.
Assessor = key expert; integrates, interprets, decides tools.
Outcome
Testing ⇒ score(s).
Assessment ⇒ integrated report & recommendations.
Historical Foundations
Early 20ᵗʰ-century France
1905 Alfred Binet & Théodore Simon: first practical mental test to identify children needing special support.
French reforms pursued liberté, égalité, fraternité; universal education → need for fair placement tools.
William Stern creates concept; warns against oppressive use.
Military expansion of testing
WW I & II: rapid screening of recruits for intelligence & emotional fitness (group tests).
Post-war: proliferation into personality, neuropsychology, workplace, etc.
Varieties of Assessment
Therapeutic Psychological Assessment
Collaborative, feedback throughout; aims for immediate self-insight.
Dynamic Assessment
Interactive “test → intervention → retest” cycle; targets learning potential.
Retrospective Assessment
Reconstructs past functioning using historical data (even on deceased).
Remote Assessment
Data gathered when assessor & assessee are geographically separated.
Ecological Momentary Assessment (EMA)
“In-the-moment” capture of behaviours & cognitions in natural contexts.
Smartphone example (Ben-Zeev):
Microphone samples speech; GPS & Bluetooth track location & movement; accelerometer for activity.
Data relates social engagement & mobility to depression level changes.
Raises privacy/ethics; research shows patients accept if benefits explained.
The Process of Assessment
Referral & Clarification → Define question(s).
Preparation & Tool Selection
Guided by assessor’s experience, literature, institutional policy.
Formal Administration
May be collaborative, therapeutic, or dynamic.
Interpretation & Report
Written integrated document; feedback sessions.
Tools of Psychological Assessment
Tests
Defined as measuring devices; differ by content, format, admin, scoring, psychometric quality.
Example distinctions:
Content: intelligence vs. personality; theory influences item type.
Format: paper, computer (local/online), time limits, adaptive.
Scores: raw, cut scores, pass vs. fail → emotional impact (Olympic study: bronze happier than silver).
Psychometric soundness = reliability, validity, utility.
Interview
Structured exchange; evaluates verbal & non-verbal cues; modes: face-to-face, phone, text, online.
Variants: panel interviews, motivational interviewing.
Portfolio
Collection of work products (art, writing, teaching aids) for evaluation.
Case-History Data
Archival records, social-media posts, letters, media, employment files.
Basis for case study narratives; used in threat assessment & neuropsych.
Behavioral Observation
Direct or electronic monitoring; naturalistic vs. lab; e.g., playground interaction or Test of Grocery Shopping Skills.
Role-Play Tests
Simulated scenarios to assess skills (leadership, substance-abuse coping).
Computers & Technology
CAT (Computer Adaptive Testing): item selection & realtime scoring.
CAPA (Computer-Assisted Psychological Assessment): administration, scoring, interpretive, consultative, integrative reports.
Q-Interactive: dual-iPad test platform; immediate scoring.
Pros: standardization, speed, global reach, eco-friendly.
Cons: security, identity verification, tech glitches, accessibility.
Alternate media: video scenarios, VR environments, biofeedback, plethysmograph.
Settings & Applications
Educational: achievement, aptitude, diagnostic, informal evaluations; grading & placement.
Clinical: diagnosis, treatment planning, malingering detection, competency decisions.
Counseling: career interests, adjustment, retirement planning.
Geriatric: quality of life, dementia vs. pseudodementia, cognitive screens.
Business & Military: personnel selection, promotion, leadership potential, product design; MAP feature—USAF psychologist evaluates Military Training Instructor candidates via interview, 360 ratings, cognitive screen.
Governmental Credentialing: licensure, certification, board exams.
Research: all sub-fields; measurement core to empirical studies.
Parties in the Enterprise
Test Developers & Publishers: design, validate, publish; guided by AERA-APA-NCME Standards (2014).
Test Users: psychologists, HR, clinicians, educators; qualifications debated; shortages of measurement experts.
Testtakers (Assessees): variable anxiety, motivation, coaching, physical state; includes deceased via psychological autopsy.
Society at Large: laws, court rulings, public opinion shape testing; e.g., protests against high-stakes school tests.
Other Stakeholders: organizations sponsoring development, scorers, marketers, academic reviewers.
Accommodation & Alternate Assessment
Legal/ethical duty to adapt tests for disabilities while preserving validity.
Accommodation types
Change stimulus format (large print, Braille, sign language).
Change response mode (speech → writing).
Modify environment (quiet room, extended time).
Interpersonal supports (presence of service animal/helper).
Variables to weigh
Assessee capabilities.
Purpose of assessment.
Impact on score meaning.
Assessor competence.
Everyday Psychometrics vignette: 1 in 7 Americans need accommodations; importance of documenting modifications.
Ethical & Practical Obligations in Testing
Before: choose appropriate test; secure materials; prepare environment; establish rapport.
During: follow standardized protocol; note unusual incidents.
After: score accurately; protect data; interpret per guidelines; provide understandable feedback.
Information Sources for Tests
Test Catalogues: quick descriptions; marketing bias.
Test Manuals: development history, norms, administration, psychometrics.
Professional Books: deep dives, clinical tips, theory.
Reference Volumes: Buros Mental Measurements Yearbook, Tests in Print—independent reviews.
Journals: Psychological Assessment, Educational and Psychological Measurement, etc.
Online Databases: ERIC, APA PsycTESTS®, PsycINFO; ETS test database.
Unpublished Measures: Directory of Unpublished Experimental Mental Measures, ETS Test Collection.
Pros & Cons table: catalogues (pro: accessible; con: sales-oriented), manuals (detailed; may lack criticism), reference books (expert insight; potential bias), Buros (objective; not as detailed), journals (current; variable quality), databases (comprehensive; watch for bogus sites).
Key Numerical / Statistical Mentions
Americans live with a disability affecting daily living.
Smartphone EMA study: 10-week monitoring; speech & GPS metrics significantly predicted depression change (Ben-Zeev et al., 2015).
Population ageing: > Americans aged ; > aged (≈52-fold increase since 1900).
Binet–Simon intended to flag children whose achievement would lag ever further—not measure total intelligence.
Test output example: score → fail vs. → pass; emotional cut-off effects demonstrated in Olympic medalist study.
Ethical / Philosophical Themes
Stern’s dictum: Testing must work “for” people, never degrade them.
Privacy & consent paramount in smartphone/remote assessments.
Clear definitions prevent professional turf wars & misuse.
Balance between societal need to categorise & individual uniqueness.