Documentation

Linglib.Phenomena.Nonliteral.Humor.Studies.KaoEtAl2016

@cite{kao-levy-goodman-2016} — A Computational Model of Linguistic Humor in Puns #

@cite{kao-levy-goodman-2016}

Kao, J.T., Levy, R., & Goodman, N.D. (2016). A Computational Model of Linguistic Humor in Puns. Cognitive Science, 40, 1270–1285.

Findings #

  1. Ambiguity (entropy of meaning distribution) distinguishes puns from non-puns
  2. Distinctiveness (KL divergence of supporting words) predicts funniness within puns
  3. Both meanings must be plausible AND supported by different parts of the sentence

Data #

The study used 435 sentences:

Funniness rated on 1–7 scale, z-scored across participants.

See Phenomena.Polysemy.Studies.ErkHerbelot2024 for the SDS↔Kao bridge.

A phonetic pun with two meanings

  • sentence : String

    The pun sentence

  • ambiguousWord : String

    The ambiguous word (as written)

  • homophone : String

    The homophone/near-homophone

  • isIdentical : Bool

    Whether it's an identical or near homophone

  • funniness :

    Mean funniness rating (z-scored)

  • ambiguity :

    Ambiguity score (entropy of P(m|w))

  • distinctiveness :

    Distinctiveness score (symmetrized KL divergence)

Instances For
    Equations
    • One or more equations did not get rendered due to their size.
    Instances For

      A non-pun control sentence

      • sentence : String

        The sentence

      • ambiguousWord : String

        The phonetically ambiguous word

      • homophone : String

        The homophone

      • funniness :

        Mean funniness rating (z-scored)

      • ambiguity :

        Ambiguity score

      • distinctiveness :

        Distinctiveness score

      Instances For
        Equations
        • One or more equations did not get rendered due to their size.
        Instances For

          Key Examples #

          These examples from the paper illustrate the ambiguity/distinctiveness measures.

          "The magician got so mad he pulled his hare out"

          • hare supported by: magician
          • hair supported by: mad, pulled High ambiguity (both meanings plausible) + High distinctiveness (different support)
          Equations
          • One or more equations did not get rendered due to their size.
          Instances For

            Control: "The hare ran rapidly across the field" Only hare meaning is supported; hair is implausible. Low ambiguity, moderate distinctiveness.

            Equations
            • One or more equations did not get rendered due to their size.
            Instances For

              "A dentist has to tell a patient the whole tooth"

              • tooth supported by: dentist, patient
              • truth supported by: tell, whole High ambiguity + High distinctiveness
              Equations
              • One or more equations did not get rendered due to their size.
              Instances For

                Control: "A dentist examines one tooth at a time" Only tooth meaning plausible.

                Equations
                • One or more equations did not get rendered due to their size.
                Instances For

                  Key Statistical Results #

                  From Table 2 and Results section:

                  Puns have significantly higher ambiguity than non-puns (t = 7.89, p < .0001)

                  Equations
                  Instances For

                    Puns have significantly higher distinctiveness than non-puns (t = 6.11, p < .0001)

                    Equations
                    Instances For

                      Within puns, ambiguity does NOT correlate with funniness (r = .03, p = .697)

                      Equations
                      Instances For

                        Model R² for predicting funniness from ambiguity + distinctiveness

                        Equations
                        Instances For

                          Theoretical Framework #

                          Ambiguity (Entropy) #

                          Amb(M) = -Σ_k P(m_k|w) log P(m_k|w)
                          

                          High ambiguity means both meanings are near-equally likely given the words. This is necessary but not sufficient for humor.

                          Distinctiveness (Symmetrized KL Divergence) #

                          Dist(F_a, F_b) = D_KL(F_a||F_b) + D_KL(F_b||F_a)
                                        = Σ_i [ln(F_a(i)/F_b(i)) · F_a(i) + ln(F_b(i)/F_a(i)) · F_b(i)]
                          

                          Where F_a = P(f|m_a, w) is the distribution over which words are semantically relevant given meaning m_a.

                          High distinctiveness means different words support different meanings. This predicts fine-grained funniness within puns.

                          Connection to Incongruity-Resolution Theory #

                          The paper argues:

                          Both are needed for humor: incongruity alone is puzzling, not funny.

                          More Examples from Supplementary Materials #

                          The full dataset is available at: http://web.stanford.edu/~justinek/punpaper/results.html

                          Example puns with their ratings (representative sample)

                          Equations
                          • One or more equations did not get rendered due to their size.
                          Instances For

                            Summary: @cite{kao-levy-goodman-2016} #

                            Main Contributions #

                            1. First computational model predicting fine-grained funniness in puns
                            2. Formal measures (ambiguity, distinctiveness) derived from language processing model
                            3. Empirical validation with 435 sentences and human ratings

                            Insight #

                            Puns are funny when:

                            1. Both meanings are plausible (high ambiguity)
                            2. Different words support different meanings (high distinctiveness)

                            Neither alone is sufficient:

                            Relevance to SDS #

                            The distinctiveness measure captures the same intuition as SDS conflict detection: different sources of evidence point to different interpretations.

                            In Kao's model: different words → different meanings In SDS: selectional vs scenario → different concepts