1/84
Looks like no tags are added yet.
Name | Mastery | Learn | Test | Matching | Spaced |
|---|
No study sessions yet.
Bernoulli Distribution
The probability distribution of a random variable with two possible outcomes, each with a constant probability of occurrence.
Binomial Distribution
The distribution that models n independent replications of a Bernoulli experiment, each with a probability p of success.
Chi-square critical value
A threshold for statistical significance for certain hypothesis tests and defines confidence intervals for certain parametersA
Chi-square goodness of fit test
A statistical hypothesis test used to determine whether a variable is likely to come from a specified distribution or not.
Chi-square statistic
The sum of squares of the differences between observed frequency, fo, and expected frequency, fe, divided by the expected frequency in each cell.
Combination
Unique outcomes of predictable event.
Complement
The set of all outcomes in the sample space that is not included in the event.
Conditional Probability
The probability of occurrence of one event A, given that another event B is known to be true or has already occurred.
Continuous Random Variable
A random variable that has outcomes over one or more continuous intervals of real numbers.
Cumulative Distribution Function
A specification of the probability that the random variable X assumes a value less than or equal to a specified value x.
Discrete Random Variable
A random variable for which the number of possible outcomes can be counted.
Discrete Uniform Distribution
A variation of the uniform distribution for which the random variable is restricted to integer values between a and b (also integers).
Empirical Probability Distribution
An approximation of the probability distribution of the associated random variable.
Event
A collection of one or more outcomes from a sample space.
Expected Value
The notion of the mean or average of a random variable; the weighted average of all possible outcomes, where the weights are the probabilities.
Experiment
A process that results in an outcome.
Exponential Distribution
A continuous distribution that models the time between randomly occurring events.
Goodness of Fit
A procedures that attempts to draw a conclusion about the nature of a distribution.
Independent Events
Events that do not affect the occurrence of each other.
Intersection
A composition with all outcomes belonging to both events.
Joint Probability
The probability of the intersection of two events.
Joint Probability Table
A table that summarizes joint probabilities.
Marginal Probability
The probability of an event irrespective of the outcome of the other joint event.
Multiplication Law of Probability
The probability of two events A and B is the product of the probability of A given B, and the probability of B (or) the product of the probability of B given A, and the probability of A.
Mutually Exclusive
Events with no outcomes in common.
Normal Distribution
A continuous distribution described by the familiar bell-shaped curve and is perhaps the most important distribution used in statistics.
Outcome
A result that can be observed.
Permutation
A mathematical technique that determines the number of possible arrangements in a set when the order of the arrangements matters.
Poisson Distribution
A discrete distribution used to model the number of occurrences in some unit of measure.
Probability
The likelihood that an outcome occurs.
Probability Density Function
The distribution that characterizes outcomes of a continuous random variable.
Probability Distribution
The characterization of the possible values that a random variable may assume along with the probability of assuming these values.
Probability Mass Function
The probability distribution of the discrete outcomes for a discrete random variable X.
Random Variable
A numerical description of the outcome of an experiment.
Sample Space
The collection of all possible outcomes of an experiment.
Standard Normal Distribution
A normal distribution with mean 0 and standard deviation 1.
Tree Diagram
A visualization of a multistep experiment. It can help with counting the outcomes.
Triangular Distribution
Defined by three parameters: the minimum, a; maximum, b; and most likely, c.
Uniform Distribution
A function that characterizes a continuous random variable for which all outcomes between some minimum and maximum value are equal likely.
Union
A composition of all outcomes that belongs to either of two events.
Autocorrelation
Correlation among successive observation over time and identified by residual plots having clusters of residuals with the same sign. Autocorrelation can be evaluated more formally using a statistical test based on the measure, Durbin—Watson statistic.
Coefficient of Determination (R²)
The tool gives the proportion of variation in the dependent variable that is explained by the independent variable of the regression model and has the value between 0 and 1.
Coefficient of multiple determination
Similar to simple linear regression, the tool explains the percentage of variation in the dependent variable. The coefficient of multiple determination in the context of multiple regression indicates the strength of association between the dependent and independent variables.
Curvilinear Regression Model
The model is used in forecasting when the independent variable is time.
Dummy Variables
A numerical variable used in regression analysis to represent subgroups of the sample in the study.
Exponential Function
Exponential functions have the property that y rises or falls at constantly increasing rates. y = abx
Homoscedasticity
The assumption means that the variation about the regression line is constant for all values of the independent variable. The data is evaluated by examining the residual plot and looking for large differences in the variances at different values of the independent variable.
Interaction
Occurs when the effect of one variable (i.e., the slope) is dependent on another variable.
Least-squares regression
The mathematical basis for the best-fitting regression line.
Linear function
Linear functions show steady increase or decrease over the range of x and used in predictive models. y = a +bx
Logarithmic Function
Used when the rate of change in a variable increases or decreases quickly and then levels out, such as with diminishing returns to scale. y = ln(x)
Multicollinearity
A condition occurring when two or more independent variable in the same regression model contain high levels of the same information and, consequently, are strongly correlated with one another and can predict each other better than the dependent variable.
Multiple Correlation Coefficient
Multiple R and R Square (or R²) in the context of multiple regression indicate the strength of association between the dependent and independent variables.
Multiple linear regression
A linear regression model with more than one independent variable. Simple linear regression is just a special case of multiple linear regression.
Overfitting
If too many terms are added to the model, then the model may not adequately predict other values from the population. Overfitting can be mitigated by using good logic, intuition, physical or behavioral theory, and parsimony.
Parsimony
A model with the fewest number of explanatory variables that will provide an adequate interpretation of the dependent variable.
Partial Regression coefficient
Represent the expected change in the dependent variable when the associated independent variable is increased by one unit while the values of all other independent variable are held constant.
Polynomial function
y = ax² + bx + c (second order—quadratic function), y = ax³ + bx² + dx + e (third order—cubic function), and so on. A second order polynomial is parabolic in nature and has only one hill or valley; a third order polynomial has one or two hills or valleys. Revenue models that incorporate price elasticity are often polynomial functions.
Power Function
Define phenomena that increase at a specific rate. Learning curves that express improving times in performing a task are often modeled with power function having a > 0 and b < 0. y = axbA
R² (R-squared)
A measure of the “fit” of the line to the data; the value of R² will be between 0 and 1. The larger the value of R², the better the fit.
Regression analysis
A tool for building mathematical and statistical models that characterize relationships between a dependent variable and one or more independent, or explanatory, variables, all of which are numerical.
Residuals
Observed errors which are the differences between the actual values and the estimated values of the dependent variable using the regression equation.
Significance of Regression
A simple hypothesis test checks whether the regression coefficient is zero.
Simple Linear Regression
A tool used to find a linear relationship between one independent variable, X, and one dependent variable, Y.
Standard Error of the Estimate, SYX
The variability of the observed Y-values from the predicted values.
Standard Residuals
Residuals divided by their standard deviation. Standard residuals describe how far each residual is from its mean in units of standard deviations.
Cyclical Effect
Characteristic of a time series that describes ups and downs over a much longer time frame, such as several years.
Delphi Method
A forecasting approach that uses a panel of experts, whose identities are typically kept confidential from one another, to respond to a sequence of questionnaires to converge to an opinion of a future forecast.
Double Exponential Smoothing
A forecasting approach similar to simple exponential smoothing used for time series with a linear trend and no significant seasonal components.
Econometric Model
Explanatory/casual models that seek to identify factors that explain statistically the patterns observed in the variable being forecast.
Historical Analogy
A forecasting approach in which a forecast is obtained through a comparative analysis with a previous situation.
Holt-Winters Models
Forecasting models similar to exponential smoothing models in that smoothing constants are used to smooth out variations in the level and seasonal patterns over time.
Index
A single measure that weighs multiple indicators, thus providing a measure of overall expectation.
Indicator
Measures that are believed to influence the behavior of a variable we wish to forecast.
Mean Absolute Deviation (MAD)
The absolute difference between the actual value and the forecast, averaged over a range of forecasted values.
Mean Absolute Percentage Error (MAPE)
The average of absolute errors divided by actual observation values.
Mean Square Error (MSE)
The average of the square of the differences between the actual value and the forecast.
Root Mean Square Error (RMSE)
The square root of mean square error (MSE).
Seasonal Effect
Characteristic of a time series that repeats at fixed intervals of time, typically a year, month, week, or day.
Simple Exponential Smoothing
An approach for short-range forecasting that is a weighed average of the most recent forecast and actual value.
Simple Moving Average
A smoothing method based on the idea of averaging random fluctuations in the time series to identify the underlying direction in which the time series is changing.
Smoothing Constant
A value between 0 and 1 used to weight exponential smoothing forecasts.
Stationary Time Series
A time series that does not have trend, seasonal, or cyclical effects but is relatively constant and exhibits only random behavior.
Time Series
A stream of historical data.
Trend
A gradual upward or downward movement of a time series over time.