1/106
Looks like no tags are added yet.
Name | Mastery | Learn | Test | Matching | Spaced |
---|
No study sessions yet.
Categorical Variables
Variables that take on values as category names or group labels.
Frequency Tables
Tables showing the number of cases in each category.
Relative Frequency Tables
Tables showing the proportion or percentage of cases in each category.
Quantitative Variable
Variable that takes on numerical values for a measured quantity.
Discrete Quantitative Variable
Variable with a finite or countable number of values.
Continuous Quantitative Variable
Variable with uncountable or infinite values.
Center
Value that separates the data roughly in half.
Spread
The scope of values from smallest to largest.
Clusters
Natural subgroups in which values fall.
Gaps
Holes where no values fall in the data.
Unimodal Distribution
Distribution with one peak.
Bimodal Distribution
Distribution with two peaks.
Skewed Distribution
Distribution spreading thinly towards higher or lower values.
Bell-shaped Distribution
Symmetric distribution with a center mound and sloping tails.
Descriptive Statistics
Presentation of data including average values, variability measures, and distribution shape.
Inferential Statistics
Drawing inferences from limited data.
Median
Middle number in a set of numbers.
Mean
Average found by summing items and dividing by the number of items.
Variability
Key concept in statistics describing the spread of data.
Range
Difference between the largest and smallest values.
Interquartile Range (IQR)
Range of the middle 50% of data.
Variance
Average of squared differences from the mean.
Standard Deviation
Square root of the variance, indicating typical distance from the mean.
Simple Ranking
Arranging elements to determine a value's position.
Percentile Ranking
Indicates the percentage of values at or below a specific value.
Z-Score
Number of standard deviations a value is above or below the mean.
Parallel Boxplots
Graphical representation showing the comparison of multiple datasets, indicating median, quartiles, and outliers.
Normal Distribution
A bell-shaped symmetrical distribution with the mean equal to the median, following the empirical rule of 68-95-99.7 for standard deviations.
Correlation Coefficient
A numerical measure indicating the strength and direction of a linear relationship between two variables, ranging from -1 to 1.
Coefficient of Determination (r^2)
The proportion of variance in the response variable explained by the variation in the explanatory variable in a linear regression model.
Residuals
Differences between observed and predicted values in a regression model, with a sum that is always zero.
Outliers
Data points that significantly deviate from the overall pattern in a dataset, affecting the regression analysis and interpretation.
Influential Scores
Scores whose removal would sharply change the regression line, especially points with extreme x-values.
High Leverage
Points with x-values far from the mean of x-values, having the potential to strongly influence the regression line.
Regression Outlier
A point with a large residual compared to others, affecting the regression line but not necessarily influential.
Correlation Coefficient (r)
Indicates the strength and direction of a linear relationship between two variables.
Simple Random Sample (SRS)
A sampling method where every possible sample of the desired size has an equal chance of being selected.
Stratified Sampling
Dividing the population into homogeneous groups and picking random samples from each stratum.
Cluster Sampling
Involves dividing the population into heterogeneous groups and selecting entire clusters randomly.
Systematic Sampling
Involves selecting every kth individual from a list after choosing a random starting point.
Sampling Variability
Refers to the natural presence of sampling error in a sample, which can be described using probability and tends to decrease with larger sample sizes.
Observational Studies
Involves observing and measuring without influencing the subjects, aiming to show associations between variables.
Experiments
Involve imposing treatments on subjects, measuring responses, and aiming to establish cause-and-effect relationships.
Experimental Units
Objects on which an experiment is performed, while subjects refer to individuals if the units are people.
Placebo Effect
The phenomenon where individuals respond to any perceived treatment, even if it is inactive.
Blinding
Occurs when subjects are unaware of the treatment they are receiving.
Double-blinding
When both subjects and evaluators are unaware of the treatment allocation.
Matched Pairs Design
Compares two treatments based on responses of paired subjects, often using the same individual for both treatments.
Replication
Involves having more than one experimental unit in each treatment group to enhance the reliability of results.
Law of Large Numbers
States that as the number of trials in an experiment increases, the relative frequency of an event tends to approach its true probability.
Guess Strategy
A strategy in a standard literacy test where the test taker selects answers randomly when the correct answer is unknown.
Score 60-79
A range of scores in a standard literacy test considered passing but not superior, falling between 60 and 79.
Does not score 60-79
The probability of a test taker not achieving a score between 60 and 79 in a standard literacy test.
Strategy "Answer (c)" and Scores 80-100
The joint probability of a test taker choosing answer (c) and scoring between 80 and 100 in a standard literacy test.
Strategy "Longest Answer" or Scores 0-59
The probability of a test taker choosing the longest answer or scoring between 0 and 59 in a standard literacy test.
Strategy "Guess" given Score 0-59
The probability of a test taker using the guess strategy given that their score falls between 0 and 59 in a standard literacy test.
Scored 80-100 given Strategy "Longest Answer"
The probability of a test taker scoring between 80 and 100 given that they chose the strategy of selecting the longest answer in a standard literacy test.
Guess Strategy and Scoring 0-59 Independence
The assessment of whether the strategy of guessing and scoring between 0 and 59 are independent events in a standard literacy test.
Strategy "Longest Answer" and Scoring 80-100 Mutual Exclusivity
The evaluation of whether the strategy of choosing the longest answer and scoring between 80 and 100 are mutually exclusive events in a standard literacy test.
Cumulative Probability Distribution
A function, table, or graph linking outcomes with the probability of less than or equal to that outcome occurring.
Normal Distribution
Provides a model for how sample statistics vary under random sampling, often calculated using z-scores.
Central Limit Theorem
States that for sufficiently large sample sizes, the sampling distribution of the mean will be approximately normal.
Biased and Unbiased Estimators
Bias means the sampling distribution is not centered on the population parameter; unbiased estimators are centered on the population parameter.
Sampling Distribution for Sample Proportions
Focuses on the proportion of successes in a sample, approximating a normal distribution for large sample sizes.
Sampling Distribution for Differences in Sample Proportions
Deals with differences obtained by subtracting sample proportions of one population from another.
Sampling Distribution for Sample Means
The variance of sample means is the population variance divided by the sample size squared.
Sampling Distribution
The distribution of sample means or proportions taken from a population, with a mean equal to the population mean and a standard deviation equal to the population standard deviation divided by the square root of the sample size.
Confidence Interval
A range of values that is likely to contain the true population parameter with a certain level of confidence, typically expressed as (point estimate ± margin of error).
Standard Error
A measure of how much the sample statistic typically varies from the population parameter, calculated as the standard deviation of the sampling distribution.
Normality Assumption
The assumption that the sampling distribution of sample means or proportions is approximately normal if certain conditions are met, like the sample size being large enough.
Type I Error
Mistakenly rejecting a true null hypothesis in hypothesis testing, with a probability denoted as α (alpha).
Type II Error
Mistakenly failing to reject a false null hypothesis in hypothesis testing, with a probability denoted as β (beta).
Power of a Test
The probability of correctly rejecting a false null hypothesis, influenced by the sample size and significance level chosen for the test.
P-value
A measure that helps determine the strength of the evidence against the null hypothesis in hypothesis testing.
Type I error
Occurs when the null hypothesis is wrongly rejected when it is actually true.
Type II error
Happens when the null hypothesis is not rejected when it is false.
Confidence Interval
A range of values that is likely to contain the true population parameter.
Two-sample z-interval
A method used to estimate the difference between two population proportions.
Null hypothesis
A statement that there is no significant difference or relationship between the variables being studied.
Alternative hypothesis
A statement that there is a significant difference or relationship between the variables being studied.
t-distribution
A probability distribution that is used when the population standard deviation is unknown.
Standard error
An estimate of the standard deviation of a sampling distribution.
Significance Test
A statistical method used to determine whether there is enough evidence to reject the null hypothesis.
Type-I Error
Mistakenly rejecting a true null hypothesis, leading to the consequence of discouraging customers from purchasing a product that might actually deliver as advertised.
Confidence Interval
An estimate of a population parameter that asks for a range of values within which the true parameter is likely to fall.
Type-II Error
Mistakenly failing to reject a false null hypothesis, potentially resulting in missed opportunities for necessary actions or improvements.
Significance Level
The threshold used to determine whether there is enough evidence to reject the null hypothesis in a hypothesis test.
Power
The probability of correctly rejecting a false null hypothesis, indicating the effectiveness of a test in detecting a true effect.
Hypothesis Test
A statistical method to assess the validity of a claim about a population parameter by comparing sample data to the null hypothesis.
Paired Data
Involves analyzing the differences between two related measurements, often using a one-sample analysis on the paired differences.
Simulation
A method to estimate the likelihood of observing a certain outcome by random chance alone, often used to determine P-values in hypothesis testing.
Two-Sample T-Test
A statistical test to compare the means of two independent samples, assessing whether there is a significant difference between the population means.
Confidence Interval for the Difference of Two Means
Estimating the range within which the true difference between two population means is likely to lie, based on sample data.
Chi-Square Test for Goodness-of-Fit
A statistical test used to determine if there is a significant difference between observed and expected frequencies in different categories.
Chi-Square Statistic (χ²)
The sum of weighted differences between observed and expected frequencies in a chi-square test.
P-value
The probability of obtaining a chi-square value as extreme as the one observed, assuming the null hypothesis is true.
Degrees of Freedom (df)
The number of categories minus one in a chi-square distribution.
Chi-Square Test for Independence
A statistical test to determine if there is a significant association between two categorical variables.
Chi-Square Test for Homogeneity
A test used to compare samples from two or more populations to see if they have the same distribution.
Sampling Distribution for the Slope
The distribution of sample slopes in linear regression models.