Qualitative Research
Purpose: Understand and interpret social interactions.
Data Sources: Open-ended responses, field notes, interviews, observations.
Sample Size: Typically small; results less generalizable.
Analysis: Identifying patterns and themes.
Subjectivity: More subjective, providing in-depth understanding.
Quantitative Research
Purpose: Test hypotheses, assess cause-and-effect relationships, and make predictions.
Data Sources: Numerical data that undergoes statistical analysis.
Sample Size: Larger than qualitative; results often more generalizable.
Objectivity: More objective, focused on measurement and quantification.
Understanding vs. Testing: Qualitative focuses on understanding while quantitative focuses on testing.
Sample Sizes: Qualitative usually has small, unrepresentative samples; quantitative typically has large samples.
Data Analysis Methods: Qualitative involves thematic analysis; quantitative relies on statistical methods.
Internal Validity: Refers to the extent to which a researcher can conclusively say that the independent variable caused the observed effects in the dependent variable.
External Validity: Concerns how well the results can be generalized to the broader population.
Importance of decolonizing learning and research in contexts like South Africa.
Qualitative approaches can challenge Eurocentric methods and incorporate local knowledge.
Researcher should engage in critical reflexivity on their assumptions and the power dynamics involved in research.
Both qualitative and quantitative methods can provide insightful findings when used together.
Quantitative Designs: Experimental, quasi-experimental, non-experimental, correlational, descriptive, survey
Qualitative Designs: Observation, case study, phenomenology
Definition: Involves manipulation of one or more variables while keeping others constant to observe effects.
Control Groups: Participants assigned to experimental or control groups to enable comparison.
Random Assignment: Ensures each participant has an equal chance of being assigned to any group to avoid bias.
Independent Variable (IV): The variable that is manipulated.
Dependent Variable (DV): The outcome variable that is measured.
Selection Bias: Differences between groups at study start can skew results.
Mortality (Attrition): Loss of participants over time affects results.
Instrumentation Threat: Changes in measurement tools between pre-tests and post-tests may distort data.
History Threat: External events may affect results during the study.
Maturation Threat: Internal changes in participants over time can influence outcomes.
Testing Threat: The effects of repeated measurements can alter participants' performance.
Multiple Treatment Effect: Results may vary due to combined treatments rather than a single one.
Small Sample Size: Conclusions drawn from small samples can lead to skewed results.
Advantages: Control over variables allows for identifying cause-and-effect relationships.
Disadvantages: Limited generalizability to real-world situations; expensive and time-consuming.
Shares similarities with true experimental design but lacks full control and random assignment.
Example: Study comparing learner annoyance from aircraft noise without random allocation of groups.
Provides summaries of thoughts, feelings, and behaviors regarding phenomena.
Focus on frequency and association of variables.
Involves gathering data from large populations to assess perceptions, attitudes, and behaviors.
Method: Can be conducted through questionnaires or interviews.
Describes relationship strengths between variables but cannot infer causation.
Example: Study showing positive correlation between height and weight.
Observational Design: Directly observes individuals in natural settings without manipulation.
Case Study Design: In-depth observation of individuals or small groups.
Phenomenology Design: Focuses on the subjective experiences and social realities of individuals.