MARK 5343 Exam 2

0.0(0)
Studied by 0 people
call kaiCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/82

encourage image

There's no tags or description

Looks like no tags are added yet.

Last updated 4:10 AM on 4/30/26
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No analytics yet

Send a link to your students to track their progress

83 Terms

1
New cards

Logistic Regression

Makes probabilistic predictions of whether you are a 1 or a 0

2
New cards

Logistic regression – what is different here regarding the DV? the IVs?

DV: Binary, dummy coded, nonmetric Y that is predicted

IV: Predictors that create a variate called Z

3
New cards

What are odds ratios?

<p></p>
4
New cards

What are log odds ratios?

knowt flashcard image
5
New cards

What is the variate in logistic regression considered to be

Z


ln(P1/P0)


a logit

6
New cards

What does it mean when Z is positive

Increased P(1) (probability of event 1)

7
New cards

What does it mean when Z is negative

Increased P(0) (probability of event 0)

8
New cards

We turn the variate into P(1) and P(0)… how?

ez over 1 + ez = P(1)


1 - P(1) = P(0)


Determines strongest alignment to DV

<p>e<sup>z</sup> over 1 + e<sup>z</sup> = P(1)</p><div data-type="horizontalRule"><hr></div><p>1 - P(1) = P(0)</p><div data-type="horizontalRule"><hr></div><p>Determines strongest alignment to DV</p>
9
New cards

Null model

Specifies in advance the P(1) and P(0) based on the observed DV proportions. Then, we introduce X material and hopefully improve over the null.


ln(P1/P0) = Z + 0(X)


Model only has the intercept and no predictors

10
New cards

Parameter estimation to get the best bs is Iterative

11
New cards

Model produced probability of being “what you are”

12
New cards

Likelihood function (L)

The joint probability based on the model for the entire sample (Multiplication of actual probabilities)

Maximizes L to find the best b’s

L = 1 if probabilities are perfect

13
New cards

Log likelihood function (LL)

Sum of the log of each probability


Model fit diagnostic that has meatier numbers than L and should be minimized


LL = Ln(.8) + Ln(.45) + Ln(.9)


Can use to compute pseudo R 2

14
New cards

-2LL

Makes LL useable in Chi-squared test of significance


Produces x2 statistic that follows a chi-squared distribution. The closer this is to 0, the better


We want the proposed model’s to be lower than the null’s

15
New cards

-2LL difference (Likelihood Ratio Test)

Difference from null model to test whether the proposed model’s estimation is significantly better than the null model


-2(LLnull - LLproposed)


Can also use to compare two models

16
New cards

Logistic output pieces that parallel multiple regression output pieces, including if stepwise

Variate: Z vs. Y-hat

Model Fit: L, LL, pseudo R2 vs. R2

b’s P-Value: Wald vs. t-statistic

Comparable X’s Impact: EXP(b) vs. Standardized coefficients (beta)

Equation: B vs. Unstandardized B

Variance Explained: pseudo R2 vs. R2

Other Model Fit: F test vs. Chi-square test of -2LL difference

17
New cards

Pseudo R-square (use Nagelkerke)

Model fit test that determines how much variability a model can explain of its outcome (explanatory power)


Want this number to be close to 1


Ex. 64% of the variance in Y is explained by the predictors

18
New cards

Significance of individual variables in logistic regression

Wald test statistic

19
New cards

Sign and size of bs and the impact on P(1)

A 1 unit increase in the predictor variable results in b increase/decrease in the Y/log-odds


Can assess directionality of relationship

20
New cards

Exp(b)

Can compare IVs relative impacts on P(1)


1 = no change in the odds

exp(b) - 1 = % change in odds

21
New cards

What model fit test is distinct for a categorical DV

Classification Matrix

22
New cards

Understanding prediction accuracy in the classification matrix

correct prediction freq + correct prediction freq = % accuracy at prediction


look at the diagonal

<p>correct prediction freq + correct prediction freq = % accuracy at prediction</p><div data-type="horizontalRule"><hr></div><p>look at the diagonal</p>
23
New cards

What are the two benchmarks for classification accuracy?

Cpro and Cmax

24
New cards

Cmax

The larger percentage between percent positive and percent negative in a classification matrix

25
New cards

Cpro

percent positive2 + (percent negative)2

26
New cards

How to compute Cpro and Cmax criteria for judging classification (1.25 x Cpro, 1.25 x Cmax)

Cmax:

If

Cpro:

27
New cards

Cpro - square actual proportions and add together, the chance standard to beat by 25%

28
New cards

Cmax – assign everyone to largest category, percent accuracy doing that? Beat by 25%

29
New cards

Logistic Regression validation with a hold-out/split sample

30
New cards

Two group discriminant Analysis – nature of DV and IVs

31
New cards

Discriminant analysis big picture idea

Find best bs that make groups as separate / discriminated as possible

32
New cards

How many discriminant functions in relation to number of groups?

33
New cards

Read outputs from running in SPSS, direct (simultaneous) or stepwise

34
New cards

Know differences in coefficient types: unstandardized, standardized, loading/structure

35
New cards

Which two are most used for variable “importance”?

36
New cards

Discriminant analysis statistical test of overall model fit

37
New cards

Classification accuracy (another test of model fit/strength, higher percentages better)

38
New cards

How do I judge quality of classification accuracy with Cpro and Cmax? (same as logistic)

39
New cards

Discriminant analysis validation with a hold-out/split sample

40
New cards

Validation via the “one-at-a-time left out” method (labeled cross-validated in SPSS)

41
New cards

Three + groups discriminant analysis – nature of DV and IVs

42
New cards

How many discriminant functions in relation to number of groups 3+?

43
New cards

Tests of significance for each function

44
New cards

Types of coefficients for each function

45
New cards

Outputs for direct or stepwise estimation, SPSS, 2D plots

46
New cards

What is the potency index?

47
New cards

Use of potency index for variable “importance” – one number for each IV even if multiple functions

48
New cards

Classification matrix and accuracy interpretation (know what it looks like for 3+ groups)

49
New cards

Map with Centroids (green, blue, red, people from slides, pink x boxes are centroids)

50
New cards

What is conjoint, what does it allow you to do?

51
New cards

Ratings-based full profile method (previous textbook chapter)

52
New cards

Defining attributes and levels

53
New cards

Use of fractional factorial designs to build profiles [the design… the X side predicting ratings]

54
New cards

Concepts of total utility (rating) and part-worth utilities (bs)

55
New cards

Addition of other profiles for “validation” - hold out profiles

56
New cards

Addition of other profiles for simulation (see what would happen in a market with certain offerings)

57
New cards

Respondent task = ratings task (vs. choice task)

58
New cards

Effect coding (vs. dummy coding) the design matrix

59
New cards

Multivariate statistical technique used to estimate part-worths (regression)

60
New cards

The “nice” property of effects coded estimates (sum to 0)

61
New cards

Determine fit at individual level, statistics from the chapter (Multiple R, Tau nonparametric)

62
New cards

Deletion of cases with bad fit (estimation or hold out) and/or illogical part-worth patterns

63
New cards

Restating part worth utilities (book process – lowest level of each attribute 0)

64
New cards

Rescaling part worth utilities (book process – each attribute brought 100 to the table)

65
New cards

Computing attribute importance (High minus low for each attribute, over sum of all hi minus low, easy if low within each attribute was made 0, now just high for each attribute over the sum of all highs)

66
New cards

Ability to “Segment” people based on individual-level information

67
New cards

Use of simulation profiles to simulate market share (book uses 2 existing products, 1 new)

68
New cards

Get total utility for each product configuration in the simulation

69
New cards

Use values to compute discrete predicted choice (maximum utility rule)

70
New cards

Or use values to compute probabilistic predicted choice (BTL, or Logit Choice rule)

71
New cards

Core understanding of leap from ratings to “Choice-Based” (with HB estimation):

72
New cards

Choice sets

73
New cards

Choice alternatives

74
New cards

Probabilistic prediction within each choice set

75
New cards

HB estimation - borrowing from upper model

76
New cards

Individual part-worth utilities

77
New cards

Averaging over every nth draw from last set of draws (e.g., every 10th draw from last 1000 gives 100 values to average)

78
New cards

Uses of individual-level utility estimates as before

79
New cards

Very high level understanding of history/development: Ratings Based

80
New cards

Very high level understanding of history/development: Choice Based – aggregate logit model, one equation for everyone

81
New cards

Very high level understanding of history/development: Latent class – segments and estimates one logit model per segment (a way to represent respondent heterogeneity)

82
New cards

Very high level understanding of history/development: CBC-HB… individual level heterogeneity, one model (part worth utilities) per participant

83
New cards

Relative variable importance: Regression, Logistic, 2 group discriminant, 3 group discriminant, conjoint