Documentation

Linglib.Theories.Pragmatics.RSA.Implementations.BergenGoodman2015

Noisy Channel Model #

A noisy channel is a stochastic matrix P_N(u_p | u_i) giving the probability that the listener perceives utterance u_p when the speaker intended u_i.

Key properties:

A noisy channel over utterances.

transmit u_i u_p is P_N(u_p | u_i): probability of perceiving u_p given intended u_i.

  • transmit : UU

    Transition probability P_N(u_p | u_i)

  • nonneg (ui up : U) : 0 self.transmit ui up

    All probabilities are non-negative

Instances For

    Identity channel: no noise, perfect transmission

    Equations
    Instances For
      def RSA.BergenGoodman2015.NoisyChannel.symmetric {U : Type} [DecidableEq U] (allU : List U) (ε : ) ( : 0 ε ε 1) :

      Symmetric noise channel: each word independently corrupted with probability ε

      Equations
      Instances For

        Ellipsis: Sentence Fragments #

        Example from paper:

        The listener infers "Bob went to the movies" because:

        1. "Bob" is not grammatical on its own
        2. Listener assumes noise deleted the rest
        3. "Bob went to the movies" → "Bob" via deletion is most probable

        Model Setup #

        Meanings: who went to the movies

        Instances For
          Equations
          • One or more equations did not get rendered due to their size.
          Instances For
            Equations
            • One or more equations did not get rendered due to their size.
            Instances For

              Utterances: full sentences and fragments

              Instances For
                Equations
                • One or more equations did not get rendered due to their size.
                Instances For
                  Equations
                  • One or more equations did not get rendered due to their size.
                  Instances For
                    Equations
                    • One or more equations did not get rendered due to their size.
                    Instances For

                      Noise model: probability of perceiving u_p given intended u_i.

                      Simplified model with base noise rate δ:

                      • Same utterance: high probability (1 - δ)
                      • Full sentence → fragment (subject): probability δ (plausible deletion)
                      • Everything else: small probability
                      Equations
                      Instances For

                        L0: Literal listener with noise inference (Eq. 6 from paper)

                        L0(m | u_p) ∝ P(m) Σ_{u_i : m ∈ [[u_i]]} P(u_i) P_N(u_p | u_i)

                        The listener:

                        1. Considers all utterances u_i that could have produced perceived u_p
                        2. Weights by prior P(u_i) and noise probability P_N(u_p | u_i)
                        3. Assigns meaning based on literal semantics of u_i
                        Equations
                        • One or more equations did not get rendered due to their size.
                        Instances For

                          S1: Speaker considering noise (simplified)

                          The speaker chooses utterances that will likely be interpreted correctly. Utility = P(listener gets correct meaning) - cost

                          Equations
                          • One or more equations did not get rendered due to their size.
                          Instances For

                            L1: Pragmatic listener reasoning about noisy speaker

                            L1(m | u_p) ∝ P(m) Σ_{u_i} S1(u_i | m) P_N(u_p | u_i)

                            Equations
                            • One or more equations did not get rendered due to their size.
                            Instances For

                              Noise rate for experiments

                              Equations
                              Instances For

                                Ellipsis Theorem: The fragment "Bob" is interpreted as "Bob went".

                                Even though "Bob" has no literal meaning, the listener infers it via:

                                1. "Bob" most likely came from "Bob went to the movies" via deletion
                                2. "Bob went to the movies" means Bob went
                                3. Therefore "Bob" → Bob went

                                Prosody: Strategic Noise Reduction #

                                Prosodic stress (BOB vs Bob) reduces noise rate on stressed words. This has pragmatic consequences:

                                Example:

                                Mechanism #

                                Speaker S1 with exhaustive knowledge (only Bob went):

                                Speaker with non-exhaustive knowledge (Bob went, maybe others too):

                                Listener infers: stress → speaker had reason to protect that word → speaker has exhaustive knowledge

                                Meanings for exhaustivity example

                                Instances For
                                  Equations
                                  • One or more equations did not get rendered due to their size.
                                  Instances For

                                    Utterances with optional prosodic stress

                                    Instances For
                                      Equations
                                      • One or more equations did not get rendered due to their size.
                                      Instances For
                                        Equations
                                        • One or more equations did not get rendered due to their size.
                                        Instances For

                                          Literal meaning: lower-bound semantics

                                          Equations
                                          Instances For

                                            Noise channel with prosody.

                                            Base noise rate ε. Stress reduces noise by factor of 2.

                                            Model: probability of subject being misheard.

                                            • No stress: ε chance of mishearing
                                            • Stress: ε/2 chance of mishearing
                                            Equations
                                            Instances For

                                              Speaker's knowledge state

                                              Instances For
                                                Equations
                                                • One or more equations did not get rendered due to their size.
                                                Instances For

                                                  L0 with noise

                                                  Equations
                                                  • One or more equations did not get rendered due to their size.
                                                  Instances For

                                                    S1: Speaker choosing utterance given meaning and knowledge

                                                    Equations
                                                    • One or more equations did not get rendered due to their size.
                                                    Instances For

                                                      L1: Pragmatic listener

                                                      Equations
                                                      • One or more equations did not get rendered due to their size.
                                                      Instances For

                                                        Prosody Theorem: Stress increases exhaustive interpretation.

                                                        "BOB went" is more likely to mean "only Bob" than "Bob went" (no stress).

                                                        Connection to Other Noise Models #

                                                        @cite{bergen-goodman-2015} (This File) #

                                                        @cite{degen-etal-2020} (DegenEtAl2020.lean) #

                                                        Potential Unification #

                                                        Could Degen's semantic noise emerge from channel noise?

                                                        Hypothesis: If utterances are descriptions of features, and features can be "misheard", then:

                                                        That is, graded truth = expected Boolean truth over noisy channel.

                                                        This would show:

                                                        Information-Theoretic View #

                                                        Both noise models reduce mutual information I(M; U):

                                                        Higher noise → less informative communication → worse utility.

                                                        This suggests a unified "informativeness" measure applicable to both.

                                                        Channel noise reduces channel capacity.

                                                        With noise rate ε, the mutual information I(U; U') between intended and perceived utterance decreases.

                                                        At ε = 0: I(U; U') = H(U) (perfect channel) At ε = 1: I(U; U') = 0 (completely noisy, no information transmitted)

                                                        Equations
                                                        Instances For
                                                          theorem RSA.BergenGoodman2015.TheoreticalConnections.noise_reduces_capacity (ε₁ ε₂ : ) (h : ε₁ < ε₂) (hε₂ : ε₂ 1) (n : ) :

                                                          Higher noise → lower capacity

                                                          Summary #

                                                          From the Paper #

                                                          1. Ellipsis (Section 3)

                                                            • Fragment "Bob" → meaning "Bob went to the movies"
                                                            • Works via noise inference: listener assumes words were deleted
                                                            • fragment_interpretation: proven
                                                          2. Prosody (Section 4)

                                                            • Stress reduces noise rate
                                                            • "BOB went" → exhaustive (only Bob)
                                                            • "Bob went" → non-exhaustive (maybe others too)
                                                            • stress_increases_exhaustivity: proven

                                                          Innovation #

                                                          The noisy channel model explains phenomena that standard RSA cannot:

                                                          Both emerge from strategic reasoning about noise.

                                                          Relation to @cite{degen-etal-2020} #

                                                          PropertyBergen & GoodmanDegen et al.
                                                          Noise locationChannelSemantics
                                                          TypeP_N(u_p|u_i)φ(u,o) ∈ [0,1]
                                                          EffectWord corruptionGraded match

                                                          Potential unification: semantic noise as expectation over channel noise.