The categorical distribution is analagous to a Bernoulli distribution with k > 2 outcomes:
The Binomial, Bernoulli, and categorical distributions are each special cases of the multinomial distribution.
Bin(n, p) = Multinom(n, (1–p, p))
Bernoulli(p) = Multinom(1, (1–p, p))
Cat(p1, p2, … pk) =
Multinom(1, (p1, p2, … pk))
1 trial | >1 trial | |
---|---|---|
2 categories | Bernoulli | Binomial |
>2 categories | Categorical | Multinomial |
But we still need a link function
Untransformed (μ)
Transformed (p)
Putting it all together gives us the multinomial logistic regression model (a.k.a. categorical regression model)
Note that with only two categories, the multinomial logistic regression reduces to the standard logistic regression:
Estimate | Q2.5 | Q97.5 | |
---|---|---|---|
αm | -5.40 | -6.16 | -4.67 |
βm | 0.59 | 0.52 | 0.67 |
αd | -4.81 | -5.80 | -3.86 |
βd | 0.41 | 0.32 | 0.51 |
αw | -3.43 | -4.68 | -2.23 |
βw | 0.19 | 0.07 | 0.32 |
Interpreting these cofficients on their own is complex — the results are analagous to a series of logistic regressions against the reference category, conditional on the other outcomes.
E.g. αm and βm determine the probability of a person being married rather than single, assuming that they are neither divorced nor widowed.
Assessing the sign of the α and β estimates is the only straightforward interpretation.
Figures by Peter McMahan (source code)
Clip from Wheel of Fortune (1983–)
Still from Crazy Rich Asians (2018)
Adapted from image by Jake Clark
Stills adapted from Senhora do Destino (2004)