@cite{kao-levy-goodman-2016} — A Computational Model of Linguistic Humor in Puns #
@cite{kao-levy-goodman-2016}
Kao, J.T., Levy, R., & Goodman, N.D. (2016). A Computational Model of Linguistic Humor in Puns. Cognitive Science, 40, 1270–1285.
Findings #
- Ambiguity (entropy of meaning distribution) distinguishes puns from non-puns
- Distinctiveness (KL divergence of supporting words) predicts funniness within puns
- Both meanings must be plausible AND supported by different parts of the sentence
Data #
The study used 435 sentences:
- 65 identical-homophone puns (e.g., hare/hair)
- 80 near-homophone puns (e.g., tooth/truth)
- 290 non-pun control sentences
Funniness rated on 1–7 scale, z-scored across participants.
See Phenomena.Polysemy.Studies.ErkHerbelot2024 for the SDS↔Kao bridge.
A phonetic pun with two meanings
- sentence : String
The pun sentence
- ambiguousWord : String
The ambiguous word (as written)
- homophone : String
The homophone/near-homophone
- isIdentical : Bool
Whether it's an identical or near homophone
- funniness : ℚ
Mean funniness rating (z-scored)
- ambiguity : ℚ
Ambiguity score (entropy of P(m|w))
- distinctiveness : ℚ
Distinctiveness score (symmetrized KL divergence)
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Key Examples #
These examples from the paper illustrate the ambiguity/distinctiveness measures.
"The magician got so mad he pulled his hare out"
- hare supported by: magician
- hair supported by: mad, pulled High ambiguity (both meanings plausible) + High distinctiveness (different support)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Control: "The hare ran rapidly across the field" Only hare meaning is supported; hair is implausible. Low ambiguity, moderate distinctiveness.
Equations
- One or more equations did not get rendered due to their size.
Instances For
"A dentist has to tell a patient the whole tooth"
- tooth supported by: dentist, patient
- truth supported by: tell, whole High ambiguity + High distinctiveness
Equations
- One or more equations did not get rendered due to their size.
Instances For
Control: "A dentist examines one tooth at a time" Only tooth meaning plausible.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Key Statistical Results #
From Table 2 and Results section:
Puns have significantly higher ambiguity than non-puns (t = 7.89, p < .0001)
Equations
Instances For
Puns have significantly higher distinctiveness than non-puns (t = 6.11, p < .0001)
Equations
Instances For
Within puns, distinctiveness correlates with funniness (r = .28, p < .001)
Equations
Instances For
Within puns, ambiguity does NOT correlate with funniness (r = .03, p = .697)
Instances For
Model R² for predicting funniness from ambiguity + distinctiveness
Equations
Instances For
Regression intercept
Equations
Instances For
Ambiguity coefficient (significant predictor)
Equations
Instances For
Distinctiveness coefficient (significant predictor)
Equations
Instances For
Theoretical Framework #
Ambiguity (Entropy) #
Amb(M) = -Σ_k P(m_k|w) log P(m_k|w)
High ambiguity means both meanings are near-equally likely given the words. This is necessary but not sufficient for humor.
Distinctiveness (Symmetrized KL Divergence) #
Dist(F_a, F_b) = D_KL(F_a||F_b) + D_KL(F_b||F_a)
= Σ_i [ln(F_a(i)/F_b(i)) · F_a(i) + ln(F_b(i)/F_a(i)) · F_b(i)]
Where F_a = P(f|m_a, w) is the distribution over which words are semantically relevant given meaning m_a.
High distinctiveness means different words support different meanings. This predicts fine-grained funniness within puns.
Connection to Incongruity-Resolution Theory #
The paper argues:
- Ambiguity ≈ presence of incongruous meanings (incongruity detection)
- Distinctiveness ≈ each meaning has coherent support (incongruity resolution)
Both are needed for humor: incongruity alone is puzzling, not funny.
More Examples from Supplementary Materials #
The full dataset is available at: http://web.stanford.edu/~justinek/punpaper/results.html
Example puns with their ratings (representative sample)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Summary: @cite{kao-levy-goodman-2016} #
Main Contributions #
- First computational model predicting fine-grained funniness in puns
- Formal measures (ambiguity, distinctiveness) derived from language processing model
- Empirical validation with 435 sentences and human ratings
Insight #
Puns are funny when:
- Both meanings are plausible (high ambiguity)
- Different words support different meanings (high distinctiveness)
Neither alone is sufficient:
- High ambiguity + low distinctiveness → confusing, not funny
- Low ambiguity + high distinctiveness → one meaning clearly wins, not funny
Relevance to SDS #
The distinctiveness measure captures the same intuition as SDS conflict detection: different sources of evidence point to different interpretations.
In Kao's model: different words → different meanings In SDS: selectional vs scenario → different concepts