1/136
Looks like no tags are added yet.
Name | Mastery | Learn | Test | Matching | Spaced |
|---|
No study sessions yet.
Alternative hypothesis H1
Assumes there is an effect or relationship and that any observed effect is not due to chance. Therefore the data is statistically significant.
Randomisation
Assigning particapants into different experimental groups by chance to prevent bias
True value
Actual correct value of the quantity being measured
Measured value
Results you get from taking a measurement with an instrument
Generalisation
Refers to the extent that the sample data can be accurately applied to the entire population.
Study design
Is a plan that outlines the methodology for conducting a research, which details how the data is collected measured and analysed
Probability distribution [new]
A statistical function [rule] that describes all the possible outcomes of a random variable and measures the likelihood of observing each outcome.
Central Limit theorem
states that for large samples (n > 30), the sampling distribution of the mean will be approximately normally distributed, and the mean of this distribution equals the population mean.
Box Plot
visually displays the distribution of a dataset by showing its five-number summary:
the minimum, lower quartile (Q1), median(Q2), upper quartile (Q3), and maximum.
It uses a box to represent the middle 50% of the data (the interquartile range) and "whiskers" to show the rest of the data's range
Z scores/ Standard score
A statistical measure that indicates how many standard deviations a data point is from the mean of a dataset
Probability Distribution
A probability distribution is a smooth curve that represents the likelihood of different outcomes. It is an idealised version of real-world data, and the area under the curve shows the probability of values occurring.
if histogram follows the probability distribution, it means the pattern of your sample data is similar to the expected population pattern.
It Serve as an idealised model of the population and Lets you see how well your sample matches the expected pattern (probability distribution curve)
Measure of dispersion/spread
Describes how spread out a set of data is from its central value, indicating the data's variability. This includes the range, interquartile range and standard deviation
Interquartile Range (IQR)
A measure of statistical dispersion, showing the spread of the middle 50% of your data/scores
Bootstrap
A statistical method that involves resampling a dataset with replacement to create many simulated samples. A method to transform data.
Transforming data
Is the process performed to make the data more suitable for statistical analysis by helping it meet the assumptions of tests
Skewness
Refers to the asymmetry of a data distribution, with main types:
positive skew (right-skewed) - tail on the right
negative skew (left-skewed) - tail on the left
Data Trimming
The process of removing a specific percentage of the most extreme values from the top and bottom of a dataset to reduce the influence of outliers.
However outliers can be meaningful and could force our data into a false distribution.
Frequency Distribution
A table or graph that shows how often values or events occur within a dataset. it is often visualised using a histogram.
Kurtosis
How much a distribution deviates from normal by looking at the spread.
refers to how tall or short the curve is on the graph
Bimodal
A distribution that has two distinct peaks or modes, indicating two high-frequency values or clusters within the data
Median
The middle value of a data set and mostly used with ordinal data, skewed data or non parametric data.
Measure of Central Tendency
Numerical values that describe the centre of a dataset, with the three most common being the mean, median, and mode. They are single (value), statistical models of the data
Reliability
Refers to the consistency and repeatability of a measurement or assessment under the same conditions
Test- retest reliability
The ability to measure and produce consistent results when the same entities are tested at two different points in time
Content Validity
Is a measure of how well a test or instrument covers all relevant aspects of the concept it is designed to measure.
The instrument/methodology actually measures the effect the researcher is interested in.
example: does simon says really test short term memory well.
Validity
The degree to which a measurement/tool or test accurately measures what it is intended to measure.
Accuracy of a method.
Measurement Error
The difference between the true value and the measured value.
This can occur due to human error or limitation in the measurement tool
Correlation Analysis
A statistical method used to evaluate the relationship between two variables, determining the strength and direction of their linear association.
Coefficient Corelation
A statistical measure that quantifies the strength and direction of a relationship between two variables, ranging from -1 to +1
Predicted Value
An estimate of the dependent variable based on the values of independent variables in a regression model
Expected value
Represents the average value you would expect if an experiment were repeated many times and is a weighted average of all possible values.
Research Question
a clear, focused, and concise question that serves as the foundation for a research project, guiding the study's direction and defining its scope
Level of data
Refers to the four scales of measurement—nominal, ordinal, interval, and ratio—that determine how precisely data is recorded
Independent Variable
What the researcher manipulates and controls to see if it has an effect on the dependent variable (DV)
Within Subject cons (repeated measure)
Longer experiments
Pro Within Subjects (Repeated measures)
Participants characteristics are not a problem (they take part in all conditions)
Requires fewer participants
Pros of Between subjects (Independent Sample)
performance not influenced by boredom, fatigue, practice effect
Shorter experiments
Cons of Between subjects (Independent sample)
Hard to match participants
More participants required
Qualitative
Non-numerical data, focusing on meaning, experience, and in-depth understanding rather than measurement
Post hoc test
a statistical analysis performed after a primary test, such as an ANOVA, to determine which specific group means are statistically different from one another.
Telling us where the difference lies.
Participant variables
Differences in participants' backgrounds that could affect the outcome, such as age, intelligence, or prior experience.
Demand characteristics
Cues from the environment or experiment that might lead participants to act in a way they believe is expected.
Experimenter effects
Unintentional actions or behaviors by the researcher that could influence participant responses.
Left Skew
Negative skew, describes a data distribution with a long tail extending to the left,
Negative scores are skewing the data. (Below the mean)
indicating most values are clustered at the right side.
Pearson correlation
measures the strength and direction of a linear relationship between two continuous variables, resulting in a coefficient that ranges from -1 to +1. A value
It is a parametric statistical test
Right Skew
Describes a distribution where most data points are clustered at the low end, with a long tail extending to the right due to a few high-value outliers.
Positive scores (greater than the mean) which are skewing the data)
since scores are higher at the start, its higher then the mean and then drops which makes the asymmetrical, skewed data.
Bias
a systematic error that causes a result to be inaccurate or skewed, meaning it does not accurately represent the population it's meant to study
Situational variables
Environmental factors that are not part of the experiment design, like the temperature, time of day, or noise levels.
Order effects
An order effect is a change in the results of a psychology experiment that occurs because of the specific sequence in which participants are exposed to different conditions or treatments
Representive sample
a smaller group of a larger population that accurately reflects the key characteristics of that group, such as demographics and behaviors
representative sample
A smaller group selected from a larger population that accurately reflects the characteristics of the whole population
Practice Effect
An improvement in performance on a task that results from repeated exposure or practice, rather than from a specific intervention
Extraneous Variable
Is anything other than the independent variable that could potentially affect the results of an experiment. If not controlled, these variables can lead to inaccurate conclusions about the relationship between the independent and dependent variables.
Qualitative
Non numerical information that describes the qualities characteristics or attributes.
Confounding variables
An extraneous (external) factor that influences both the independent and dependent variables in a study, leading to a misleading association between them.
We want to control this third variable.
Null hypothesis significant testing
Null hypothesis significance testing (NHST) is a statistical method used to determine if a relationship or effect observed in a sample is likely to be real in the population or simply due to chance
Research study
Systematic and detailed investigation into a specific problem using scientific methods
Casual statement
A statement that describes the cause and effect relationship between variables or events
Inferential statistics
Focuses on using a representative sample data to make generalisations about the population and to test the hypothesis, relationship and effect
Descriptive statistics
Used to summarise and describe the characteristics of a data set. This includes measure of central tendency and measure of dispersion.
Only tells us about the sample not the population.
Ecological Validity
The extent to which the findings of a research study can be generalized to real-life settings.
Evidence that ensures the results of the study/experiments can be applied and allow inferences to real world conditions.
Sphericity
It is the assumption that the variances of the differences between all possible pairs of within-subject conditions are equal
Predictor variables
As another name for the Independent Variable (IV). It is what the experimenter manipulates or controls.
Goodness of fit
A statistical assessment that determines how well a sample of data fits a specific distribution or model.
It asses the inconsistency between the observed and expected value from the model, helping to assess how well the model represents the data.
Sample
Is a subset of a larger population that is studied to make inferences about the entire group.
Standard Deviation (SD)
Measures how spread out the data points are from the average (mean) of the data set. It indicates how much individual data points typically deviate from the mean value.
Low Standard Deviation: Data points are clustered tightly around the mean, showing consistency.
High Standard Deviation: Data points are more spread out, showing greater variability.
R²
The proportion of variance in the dependent variable that is explained by the independent variable(s) in the model. Used for linear regression (one IV and one DV).
𝜔2
is an effect size measure used primarily in (ANOVA) that quantifies the magnitude of the relationship between an independent variable and a dependent variable.
Adjusted R-squared (Adj. R²)
The proportion of variance in the dependent variable that is explained by the independent variable(s) in the model, taking into consideration of the sample size and its predictors. Used for multiple regression (multiple IVs).
Effect size
A statistical measure that quantifies the strength and the direction between two variables. The magnitude of difference between groups.
Magnitude
Refers to the strength and direction of the relationship or effect between variables. General descriptive term. Linked to standardized beta (\beta) because \beta expresses the strength of the effect in standard deviation units.
Small: r or \beta \approx 0.1
Medium: r or \beta \approx 0.3
Large: r or \beta \approx 0.5 or higher.
Degrees of freedom (df)
The number of independent values in a calculation that are free to vary after constraints are applied.
Hypothesis
A testable statement predicting whether there is a relationship or effect between variables and whether it is likely due to chance.
Test statistic
A numerical value calculated from data which is used to determine whether a result is statistically significant and to help decide whether to reject a null hypothesis.
P-value
The probability of obtaining the observed result, or a more extreme result, assuming the null hypothesis is true. If P > 0.05, we typically fail to reject the null hypothesis.
Bar charts
A graph that uses rectangular bars to show the frequency, counts or proportion of different categorical data.
Type of data - categorical [nominal/ordinal].
The bars are not touching each other, and each bar represents its own category. It is a descriptive statistic to visually compare categories.
One-tailed hypothesis
Predicts a specific direction of the effect of the IV on the DV.
Two-tailed hypothesis
Predicts an effect exists but does not specify the direction.
Type I error
Occurs when you incorrectly reject the null hypothesis, concluding there is an effect when in reality there is none.
Also known as a false positive (seeing an effect that isn't there).
Type II error
Occurs when you fail to reject the null hypothesis when it is actually false, concluding there is no effect when there is one.
Also known as a false negative (not seeing an effect when there is one).
Regression coefficient
The slope in a regression model that tells the expected change in the dependent variable for a one-unit increase in the independent variable.
Unstandardized B
Represents the expected change in the DV for a one-unit increase in the IV using the original units of the DV. Used to interpret actual impact.
Standardized beta (\beta)
The expected change in the DV for a one standard deviation increase in the IV, which can be used to signify the strength and direction between the variables.
Allows comparison of the direction and strength of effects across different variables.
Statistics
Numerical summary measured from a sample data (e.g., sample mean or sample mode). It represents a property of the sample.
Between-Subjects Design
Different participants are assigned to different conditions, allowing comparison of the effects of the IV across groups.
This is also known as an independent groups design.
Bias
Occurs when systematic errors or researcher influence affect how data are collected, analysed, or interpreted, making the results less representative of the true population.
This can come from the researcher, participants or the sampling process.
Central Measures of Tendency
Descriptive statistics that depict the overall 'central' trend of a set of data. There are three key measures: mean, median and mode.
CI (Confidence interval)
A range from a set of values. If we were to repeat the sampling method multiple times, we would be confident that about 95% of the confidence intervals generated would contain the true population parameter
Cohen's d
A measure of effect size that assesses the strength of the difference between two means in terms of standard deviation.
Small effect: d \approx 0.2
Medium effect: d \approx 0.5
Large effect: d \approx 0.8
Spearman rho
A Spearman's correlation was conducted to evaluate the relationship between students' French written and oral exam grades. There was a significant positive relationship,
rs(38)=.65,p<.001r sub s open paren 38 close paren equals .65 comma p is less than .001
𝑟𝑠(38)=.65,𝑝<.001
.
Non parametric
Pearson r correlation
"A Pearson correlation was calculated to assess the relationship between self-efficacy and breastfeeding exclusive." There was a significant positive correlation between self-efficacy and breastfeeding exclusive, \(r(df)=.45\), \(p<.001\).
Control group
The group in an experiment that is not manipulated and is seen as a baseline to compare the effects of the IV on the dependent group.
Correlation coefficient
Measures the strength and direction of a linear relationship between two variables; does not imply causation.
Descriptive Statistics
Statistical methods used to summarise, organise, and describe the main features of a dataset or sample. Includes measures like mean, median, mode, standard deviation, and visual tools like graphs and tables to help us understand the data.
DV
The variable that is measured or observed to see if an effect has occurred due to the IV. Also known as the outcome variable.
Parameter
A numerical value that describes a characteristic of the whole population (e.g., population mean).
Standard error
Measures how much a sample mean is likely to vary from the population mean.
It acts as an indicator of the sample mean's accuracy as an estimate of the population mean.
Nominal data
Categorical data that categorizes variables into distinct groups or labels without any natural order or quantitative value.
ex, gender, age, hair colour (brown, white), nationality (English, Chinese)
Histogram
A graphical representation of the frequency distribution of continuous data, showing how often values occur within certain ranges (bins). Bars touch each other, indicating continuous/scale data [interval/ratio].
Ordinal data
Categorical data that has a natural, ordered ranking but the intervals between ranks are unknown.
Ex, Likert charts, educational levels, 1-5 scales