Methods
Overview of Developmental Methods
1. Learning Goals
Understanding Different Theoretical Positions: Explore different "isms" relating to human knowledge including Nativism, Constructivism, Core Knowledge, Empiricism, Information Processing Theory, as well as the contributions of key theorists like Piaget and Vygotsky.
Nature vs. Nurture Dichotomy: Recognize the significance and limitations of this exploration in human development.
Experimental Design Principles: Understand how different experimental designs function in research.
Logic of Infant Research Methods: Describe methods such as looking-time paradigms.
Evaluation of Research Methods: Assess pros and cons of different experimental approaches.
2. The Scientific Method
Definition: A structured process of inquiry used to test ideas.
Key Steps:
Ask a Question: Initial curiosity or inquiry.
Formulate a Hypothesis: A testable prediction.
Test the Hypothesis: Conduct experiments to gather data.
Draw a Conclusion: Analyze results to confirm or refute the hypothesis.
3. Experimental and Correlational Studies
3.1 Experimental Studies
Definition: Researchers manipulate one or more independent variables (IVs) to observe effects on dependent variables (DVs).
Causal Claims: Allows statements like "X causes Y" due to controlled conditions.
Components:
Independent Variable (IV): The manipulated variable to determine its effect.
Dependent Variable (DV): The measured outcome to assess changes resulting from the IV.
3.2 Correlational Studies
Definition: Observational approach where variables are measured without manipulation.
Limitations: Can show relationships or associations but cannot determine causation.
Variables in Correlational Studies:
Independent Variable (IV): Considered the predictor but not actively manipulated.
Dependent Variable (DV): The measured outcome thought to be influenced by the IV.
4. Research Questions and Variables
Various examples exhibit how to structure research questions with corresponding IVs and DVs:
Cooperative Play and Friendship Quality: IV - Amount of cooperative play; DV - Friendship quality ratings.
Parental Praise Style and Child Persistence: IV - Praise style (effort vs ability-focused); DV - Time spent on puzzles.
Sharing Behavior: IV - Relationship to recipient (friend vs stranger); DV - Number of stickers shared.
Risk-Taking in Adolescents: IV - Peer presence (alone vs with others); DV - Risky choices in driving simulations.
5. Correlation vs. Causation
Concept: Just because two variables are correlated does not entail a causal relationship.
Example: Exploring if praise influences persistence can be approached observationally by recording existing praise versus experimentally altering praise style across contexts.
Discussion: Reflections on whether the observed increase in persistence is genuinely due to praise or if they are related through a third variable.
6. Dependent Variables Considerations
6.1 Operationalization and Conceptual Criticisms
Persistence: Measured by time on a task but may confound with interest, enjoyment, or compliance.
Fairness: Decisions made in a token division task may inaccurately reflect true fairness principles; influenced by cultural biases and personal traits.
Justice: Some interpretations link justice too tightly with punishment; this limits the measure's validity.
6.2 Reliability and Validity
Reliability: Consistency of measures over time (test-retest, inter-rater reliability).
Validity: Ensures measures are accurately gauging intended concepts.
Internal Validity: Are you testing what you intend to?
External Validity: Generalizability of findings beyond the laboratory setting.
7. Study Design Types
7.1 Longitudinal Studies
Definition: Track the same individuals over time for developmental insights.
Pros: Provides rich, informative data.
Cons: Time-consuming and expensive.
7.2 Cross-Sectional Studies
Definition: Snapshots of different age groups at one time.
Pros: Faster data collection.
Cons: Less depth in developmental understanding.
8. Experimental Integrity Checks
Manipulation Checks: Additional measures to ensure participants correctly experienced the manipulation intended by the researchers.
Control Conditions vs Treatment Conditions: Establish baselines for comparison against experimental manipulations.
9. Infant Research Methods
9.1 Habituation Studies
Definition: Measures infant attention and learning by showing stimuli repeatedly until looking time decreases.
Implications: A rise in looking time to new stimuli (dishabituation) signifies recognition of novelty.
9.2 Violation of Expectation Paradigms
Definition: Evaluates infant attention to expected versus unexpected events based on cognitive expectations about the world.
Use: Indicators of understanding can be drawn from longer looking times at surprising events.
10. Critique of Observational and Experimental Methods
10.1 Ambiguity of Looking Time
Interpretations: Possible meanings for prolonged looking times include surprise, curiosity, confusion, or preference, marking a challenge in inference made from such data.
Reliability Issues: Minor procedural or stimulus changes can lead to fluctuations in infant looking times.
11. Sampling and Generalizability Issues
11.1 Sampling Sources
Places to recruit child participants include labs, community schools, and parks, each with inherent challenges.
11.2 Sample Size Considerations
Power: The ability of a study to detect a genuine effect, akin to brightness illuminating what exists in darkness.
12. Confounds and Validity Threats
Confounds: Additional variables that may bias results when not accounted for.
Randomization: To ensure groups are initially similar and reduce bias.
Counterbalancing: To deal with potential order effects from participants experiencing multiple conditions.
13. Reading Scientific Literature
13.1 Anatomy of a Scientific Paper
Structure Typicalities: Includes Introduction, Method, Results, Discussion sections.
Peer Review and Types of Articles: Understanding the distinction between review articles and original empirical research.
13.2 Responsibilities and Critical Thinking
Conduct critical evaluations of research based on methodology and findings validity, alongside recognizing the possibility of bias even in published research papers.