Noisy Channel Model #
A noisy channel is a stochastic matrix P_N(u_p | u_i) giving the probability that the listener perceives utterance u_p when the speaker intended u_i.
Key properties:
- Rows sum to 1 (it's a conditional distribution)
- Diagonal entries are typically high (usually perceived correctly)
- Off-diagonal entries represent noise (corruption)
Identity channel: no noise, perfect transmission
Equations
Instances For
Symmetric noise channel: each word independently corrupted with probability ε
Equations
Instances For
Ellipsis: Sentence Fragments #
Example from paper:
- A: "Who went to the movies?"
- B: "Bob" (fragment, not full sentence)
The listener infers "Bob went to the movies" because:
- "Bob" is not grammatical on its own
- Listener assumes noise deleted the rest
- "Bob went to the movies" → "Bob" via deletion is most probable
Model Setup #
- Meanings M = {Alice went, Bob went, Nobody went}
- Full utterances U = {"Alice went", "Bob went", "Nobody went"}
- Fragments F = {"Alice", "Bob", "Nobody",...}
- Noise: word deletion with probability δ
Equations
- RSA.BergenGoodman2015.Ellipsis.instBEqMeaning.beq x✝ y✝ = (x✝.ctorIdx == y✝.ctorIdx)
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- RSA.BergenGoodman2015.Ellipsis.instBEqUtterance.beq x✝ y✝ = (x✝.ctorIdx == y✝.ctorIdx)
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Literal meaning: only full sentences have truth conditions
Equations
- RSA.BergenGoodman2015.Ellipsis.literalMeaning RSA.BergenGoodman2015.Ellipsis.Utterance.aliceWentToMovies RSA.BergenGoodman2015.Ellipsis.Meaning.aliceWent = true
- RSA.BergenGoodman2015.Ellipsis.literalMeaning RSA.BergenGoodman2015.Ellipsis.Utterance.bobWentToMovies RSA.BergenGoodman2015.Ellipsis.Meaning.bobWent = true
- RSA.BergenGoodman2015.Ellipsis.literalMeaning RSA.BergenGoodman2015.Ellipsis.Utterance.nobodyWentToMovies RSA.BergenGoodman2015.Ellipsis.Meaning.nobodyWent = true
- RSA.BergenGoodman2015.Ellipsis.literalMeaning x✝¹ x✝ = false
Instances For
Prior over utterances (speaker's production probability)
Equations
- RSA.BergenGoodman2015.Ellipsis.utterancePrior RSA.BergenGoodman2015.Ellipsis.Utterance.aliceWentToMovies = 1
- RSA.BergenGoodman2015.Ellipsis.utterancePrior RSA.BergenGoodman2015.Ellipsis.Utterance.bobWentToMovies = 1
- RSA.BergenGoodman2015.Ellipsis.utterancePrior RSA.BergenGoodman2015.Ellipsis.Utterance.nobodyWentToMovies = 1
- RSA.BergenGoodman2015.Ellipsis.utterancePrior x✝ = 0
Instances For
Noise model: probability of perceiving u_p given intended u_i.
Simplified model with base noise rate δ:
- Same utterance: high probability (1 - δ)
- Full sentence → fragment (subject): probability δ (plausible deletion)
- Everything else: small probability
Equations
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.aliceWentToMovies RSA.BergenGoodman2015.Ellipsis.Utterance.aliceWentToMovies = 1 - δ
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.aliceWentToMovies RSA.BergenGoodman2015.Ellipsis.Utterance.alice = δ
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.bobWentToMovies RSA.BergenGoodman2015.Ellipsis.Utterance.bobWentToMovies = 1 - δ
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.bobWentToMovies RSA.BergenGoodman2015.Ellipsis.Utterance.bob = δ
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.nobodyWentToMovies RSA.BergenGoodman2015.Ellipsis.Utterance.nobodyWentToMovies = 1 - δ
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.nobodyWentToMovies RSA.BergenGoodman2015.Ellipsis.Utterance.nobody = δ
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.alice RSA.BergenGoodman2015.Ellipsis.Utterance.alice = 1
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.bob RSA.BergenGoodman2015.Ellipsis.Utterance.bob = 1
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.nobody RSA.BergenGoodman2015.Ellipsis.Utterance.nobody = 1
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ RSA.BergenGoodman2015.Ellipsis.Utterance.wentToMovies RSA.BergenGoodman2015.Ellipsis.Utterance.wentToMovies = 1
- RSA.BergenGoodman2015.Ellipsis.noiseChannel δ x✝¹ x✝ = 0
Instances For
Utterance cost (longer = more costly)
Equations
- RSA.BergenGoodman2015.Ellipsis.cost RSA.BergenGoodman2015.Ellipsis.Utterance.aliceWentToMovies = 5
- RSA.BergenGoodman2015.Ellipsis.cost RSA.BergenGoodman2015.Ellipsis.Utterance.bobWentToMovies = 5
- RSA.BergenGoodman2015.Ellipsis.cost RSA.BergenGoodman2015.Ellipsis.Utterance.nobodyWentToMovies = 5
- RSA.BergenGoodman2015.Ellipsis.cost RSA.BergenGoodman2015.Ellipsis.Utterance.alice = 1
- RSA.BergenGoodman2015.Ellipsis.cost RSA.BergenGoodman2015.Ellipsis.Utterance.bob = 1
- RSA.BergenGoodman2015.Ellipsis.cost RSA.BergenGoodman2015.Ellipsis.Utterance.nobody = 1
- RSA.BergenGoodman2015.Ellipsis.cost RSA.BergenGoodman2015.Ellipsis.Utterance.wentToMovies = 4
Instances For
L0: Literal listener with noise inference (Eq. 6 from paper)
L0(m | u_p) ∝ P(m) Σ_{u_i : m ∈ [[u_i]]} P(u_i) P_N(u_p | u_i)
The listener:
- Considers all utterances u_i that could have produced perceived u_p
- Weights by prior P(u_i) and noise probability P_N(u_p | u_i)
- Assigns meaning based on literal semantics of u_i
Equations
- One or more equations did not get rendered due to their size.
Instances For
S1: Speaker considering noise (simplified)
The speaker chooses utterances that will likely be interpreted correctly. Utility = P(listener gets correct meaning) - cost
Equations
- One or more equations did not get rendered due to their size.
Instances For
Noise rate for experiments
Equations
Instances For
Ellipsis Theorem: The fragment "Bob" is interpreted as "Bob went".
Even though "Bob" has no literal meaning, the listener infers it via:
- "Bob" most likely came from "Bob went to the movies" via deletion
- "Bob went to the movies" means Bob went
- Therefore "Bob" → Bob went
L0 correctly interprets fragment
Prosody: Strategic Noise Reduction #
Prosodic stress (BOB vs Bob) reduces noise rate on stressed words. This has pragmatic consequences:
Example:
- "Bob went to the movies" - compatible with others also going
- "BOB went to the movies" - exhaustive: ONLY Bob went
Mechanism #
Speaker S1 with exhaustive knowledge (only Bob went):
- Needs listener to hear "Bob" correctly
- If listener mishears "Alice", wrong inference
- Therefore: stress "Bob" to reduce noise
Speaker with non-exhaustive knowledge (Bob went, maybe others too):
- Less critical if listener mishears "Bob" as "Alice"
- Both are compatible with speaker's knowledge
- Therefore: less likely to stress
Listener infers: stress → speaker had reason to protect that word → speaker has exhaustive knowledge
Equations
- RSA.BergenGoodman2015.Prosody.instBEqMeaning.beq x✝ y✝ = (x✝.ctorIdx == y✝.ctorIdx)
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- RSA.BergenGoodman2015.Prosody.instBEqUtterance.beq x✝ y✝ = (x✝.ctorIdx == y✝.ctorIdx)
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Literal meaning: lower-bound semantics
Equations
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.aliceWent RSA.BergenGoodman2015.Prosody.Meaning.onlyAlice = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.aliceWent RSA.BergenGoodman2015.Prosody.Meaning.both = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.ALICE_went RSA.BergenGoodman2015.Prosody.Meaning.onlyAlice = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.ALICE_went RSA.BergenGoodman2015.Prosody.Meaning.both = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.bobWent RSA.BergenGoodman2015.Prosody.Meaning.onlyBob = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.bobWent RSA.BergenGoodman2015.Prosody.Meaning.both = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.BOB_went RSA.BergenGoodman2015.Prosody.Meaning.onlyBob = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.BOB_went RSA.BergenGoodman2015.Prosody.Meaning.both = true
- RSA.BergenGoodman2015.Prosody.literalMeaning RSA.BergenGoodman2015.Prosody.Utterance.aliceAndBobWent RSA.BergenGoodman2015.Prosody.Meaning.both = true
- RSA.BergenGoodman2015.Prosody.literalMeaning x✝¹ x✝ = false
Instances For
Noise channel with prosody.
Base noise rate ε. Stress reduces noise by factor of 2.
Model: probability of subject being misheard.
- No stress: ε chance of mishearing
- Stress: ε/2 chance of mishearing
Equations
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.aliceWent RSA.BergenGoodman2015.Prosody.Utterance.aliceWent = 1 - ε
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.aliceWent RSA.BergenGoodman2015.Prosody.Utterance.bobWent = ε
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.bobWent RSA.BergenGoodman2015.Prosody.Utterance.bobWent = 1 - ε
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.bobWent RSA.BergenGoodman2015.Prosody.Utterance.aliceWent = ε
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.ALICE_went RSA.BergenGoodman2015.Prosody.Utterance.ALICE_went = 1 - ε / 2
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.ALICE_went RSA.BergenGoodman2015.Prosody.Utterance.aliceWent = ε / 2
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.BOB_went RSA.BergenGoodman2015.Prosody.Utterance.BOB_went = 1 - ε / 2
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.BOB_went RSA.BergenGoodman2015.Prosody.Utterance.bobWent = ε / 2
- RSA.BergenGoodman2015.Prosody.noiseChannel ε RSA.BergenGoodman2015.Prosody.Utterance.aliceAndBobWent RSA.BergenGoodman2015.Prosody.Utterance.aliceAndBobWent = 1
- RSA.BergenGoodman2015.Prosody.noiseChannel ε x✝¹ x✝ = 0
Instances For
Speaker's knowledge state
Instances For
Equations
- RSA.BergenGoodman2015.Prosody.instBEqKnowledge.beq x✝ y✝ = (x✝.ctorIdx == y✝.ctorIdx)
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Speaker's belief given observation (for partial knowledge model)
Equations
- RSA.BergenGoodman2015.Prosody.speakerBelief RSA.BergenGoodman2015.Prosody.Knowledge.full m = 1
- RSA.BergenGoodman2015.Prosody.speakerBelief RSA.BergenGoodman2015.Prosody.Knowledge.incomplete RSA.BergenGoodman2015.Prosody.Meaning.onlyBob = 1 / 2
- RSA.BergenGoodman2015.Prosody.speakerBelief RSA.BergenGoodman2015.Prosody.Knowledge.incomplete RSA.BergenGoodman2015.Prosody.Meaning.both = 1 / 2
- RSA.BergenGoodman2015.Prosody.speakerBelief RSA.BergenGoodman2015.Prosody.Knowledge.incomplete RSA.BergenGoodman2015.Prosody.Meaning.onlyAlice = 0
Instances For
Equations
Instances For
Prosody Theorem: Stress increases exhaustive interpretation.
"BOB went" is more likely to mean "only Bob" than "Bob went" (no stress).
Both utterances are compatible with the "both" meaning
Connection to Other Noise Models #
@cite{bergen-goodman-2015} (This File) #
- Channel noise: P_N(u_p | u_i)
- Words can be deleted, inserted, replaced
- Noise is in transmission, not semantics
- Prosody reduces channel noise
@cite{degen-etal-2020} (DegenEtAl2020.lean) #
- Semantic noise: φ(u, o) ∈ [0,1]
- Feature matching is probabilistic
- Noise is in perception of features
- No explicit channel model
Potential Unification #
Could Degen's semantic noise emerge from channel noise?
Hypothesis: If utterances are descriptions of features, and features can be "misheard", then:
- φ(u, o) = E_{u' ~ P_N(·|u)}[[u']]
That is, graded truth = expected Boolean truth over noisy channel.
This would show:
- Degen's continuous φ = Bergen's discrete channel + Boolean semantics
- Similar to: Degen φ = Boolean × noise channel (proven in DegenEtAl2020.lean)
Information-Theoretic View #
Both noise models reduce mutual information I(M; U):
- Channel noise: I(U_intended; U_perceived) decreases
- Semantic noise: I(U; O) decreases (less discrimination)
Higher noise → less informative communication → worse utility.
This suggests a unified "informativeness" measure applicable to both.
Channel noise reduces channel capacity.
With noise rate ε, the mutual information I(U; U') between intended and perceived utterance decreases.
At ε = 0: I(U; U') = H(U) (perfect channel) At ε = 1: I(U; U') = 0 (completely noisy, no information transmitted)
Equations
Instances For
Higher noise → lower capacity
Summary #
From the Paper #
Ellipsis (Section 3)
- Fragment "Bob" → meaning "Bob went to the movies"
- Works via noise inference: listener assumes words were deleted
fragment_interpretation: proven
Prosody (Section 4)
- Stress reduces noise rate
- "BOB went" → exhaustive (only Bob)
- "Bob went" → non-exhaustive (maybe others too)
stress_increases_exhaustivity: proven
Innovation #
The noisy channel model explains phenomena that standard RSA cannot:
- Fragments have NO literal meaning, yet are interpreted
- Prosody has NO semantic content, yet affects meaning
Both emerge from strategic reasoning about noise.
Relation to @cite{degen-etal-2020} #
| Property | Bergen & Goodman | Degen et al. |
|---|---|---|
| Noise location | Channel | Semantics |
| Type | P_N(u_p|u_i) | φ(u,o) ∈ [0,1] |
| Effect | Word corruption | Graded match |
Potential unification: semantic noise as expectation over channel noise.