Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
- HawkinsGweonGoodman2021.instBEqUtterance.beq x✝¹ x✝ = false
Instances For
All possible utterances (2³ = 8)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Utterance cost: number of features mentioned
Equations
Instances For
Equations
Equations
- One or more equations did not get rendered due to their size.
Instances For
Does utterance literally apply to object?
Equations
- One or more equations did not get rendered due to their size.
Instances For
Extension of utterance: objects it applies to
Equations
- HawkinsGweonGoodman2021.extension u targetFeatures objects = List.filter (HawkinsGweonGoodman2021.utteranceApplies u targetFeatures) objects
Instances For
Utterance u₀ is more specific than u₁ if its extension is a subset
Equations
- One or more equations did not get rendered due to their size.
Instances For
Literal listener probability: P(o | u, C) ∝ L(u, o)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Egocentric informativity: listener success rate in visible context only
Equations
- HawkinsGweonGoodman2021.egocentricInformativity u target visibleObjects = HawkinsGweonGoodman2021.literalListenerProb u target.features target visibleObjects
Instances For
Egocentric utility: informativity minus cost U_ego(u; o, C) = I_ego(u; o, C_visible) - λ · cost(u)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Possible hidden objects: all feature combinations
Equations
- One or more equations did not get rendered due to their size.
Instances For
Asymmetric informativity: marginalizes over possible hidden objects
I_asym(u; o, C) = Σ_{o_h} P(o_h) · P_L0(o | u, C ∪ {o_h})
This captures the speaker's expected listener success rate under uncertainty.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Asymmetric utility: informativity minus cost
U_asym(u; o, C) = I_asym(u; o, C) - λ · cost(u)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Perspective-taking weight: 0 = egocentric, 1 = full perspective-taking
Equations
Instances For
Mixture informativity: interpolates between egocentric and asymmetric
I_mix(u; o, C, w_S) = w_S · I_asym(u; o, C) + (1 - w_S) · I_ego(u; o, C)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Mixture utility: interpolates between egocentric and asymmetric
U_mix(u; o, C, w_S) = w_S · U_asym(u; o, C) + (1 - w_S) · U_ego(u; o, C)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Speaker probability: P(u | o, C, w_S) ∝ exp(α · U_mix)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Normalize speaker scores to get probabilities
Equations
- One or more equations did not get rendered due to their size.
Instances For
Expected communicative accuracy at weight w_S. This is the benefit side of the cost-benefit trade-off.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Resource-rational utility: accuracy - β · w
U_RR(w_S) = E[accuracy] - β · w_S
Equations
- HawkinsGweonGoodman2021.resourceRationalUtility target visibleObjects hiddenPrior wS alpha beta = HawkinsGweonGoodman2021.expectedAccuracy target visibleObjects hiddenPrior wS alpha - beta * wS
Instances For
Example context: target is unique in shape
Equations
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Uniform prior over hidden objects
Equations
- HawkinsGweonGoodman2021.uniformHiddenPrior x✝ = 1 / 64
Instances For
Shape-only utterance
Equations
Instances For
Shape + color utterance
Equations
Instances For
Shape + color + texture utterance
Equations
Instances For
Theorem 1: More specific utterance has higher INFORMATIVITY under asymmetry.
When there may be hidden objects, being more specific guards against more possible distractors. This tests informativity (without cost).
Theorem 2: Asymmetric INFORMATIVITY favors specificity more than egocentric.
This is the key qualitative prediction: the GAIN from being more specific is larger under asymmetric reasoning than egocentric reasoning.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Theorem 3: Full description has highest informativity at high perspective-taking weight.
When w_S = 1 (full perspective-taking), more informative utterances maximize expected listener success rate.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Theorem 4: Shape-only has same informativity as full description at egocentric weight.
When w_S = 0 (pure egocentric), shape alone is equally informative in visible context (target is unique in shape among visible objects).
Equations
- One or more equations did not get rendered due to their size.
Instances For
Compositional Grounding #
The utterance semantics derive from predicate modification (H&K Ch. 4):
⟦α β⟧ = λx. ⟦α⟧(x) ∧ ⟦β⟧(x)
Each feature mention (shape, color, texture) is an intersective adjective
that denotes a characteristic function of type e → t:
- ⟦square⟧ = λx. shape(x) = square
- ⟦blue⟧ = λx. color(x) = blue
- ⟦checked⟧ = λx. texture(x) = checked
Composing via predicate modification: ⟦blue checked square⟧ = λx. blue(x) ∧ checked(x) ∧ square(x)
This is exactly Semantics.Montague.Modification.intersectiveMod applied iteratively.
Feature predicates are Montague-style intersective adjectives (e → t).
Each feature denotes a characteristic function from entities (Objects) to truth values. These are the basic building blocks for compositional utterance semantics.
Equations
- HawkinsGweonGoodman2021.MontaguGrounding.shapePred targetShape o = (o.features.shape == targetShape)
Instances For
Equations
- HawkinsGweonGoodman2021.MontaguGrounding.colorPred targetColor o = (o.features.color == targetColor)
Instances For
Equations
- HawkinsGweonGoodman2021.MontaguGrounding.texturePred targetTexture o = (o.features.texture == targetTexture)
Instances For
Compositionally derived utterance denotation.
An utterance mentions some subset of {shape, color, texture}.
The denotation is the conjunction of all mentioned feature predicates,
using predMod from Semantics.Montague.Modification:
⟦blue checked square⟧ = predMod (predMod ⟦blue⟧ ⟦checked⟧) ⟦square⟧ = λx. blue(x) ∧ checked(x) ∧ square(x)
Equations
- One or more equations did not get rendered due to their size.
Instances For
Direct (ad-hoc) utterance denotation from Part 2
Equations
- One or more equations did not get rendered due to their size.
Instances For
Grounding theorem: Direct denotation equals compositional derivation.
The ad-hoc semantics in utteranceApplies are exactly what we get from applying predicate modification (from Semantics.Montague.Modification) to individual feature predicates.
Grounding theorem: utteranceApplies = compositional denotation
The RSA meaning function φ is grounded in compositional semantics
Equations
Instances For
Grounding: utteranceApplies equals compositional denotation
Asymmetric Case via Unified API #
The full perspective-taking case (w_S = 1) maps to RSAConfig with latent variables:
- Latent = Speaker's visual access (which objects they)
- World = Full context (visible objects + hidden object features)
- speakerCredence = P(world | speaker's visual access)
The mixture model (w_S ∈ (0,1)) and resource-rational optimization (finding w*) are implementation-specific extensions that sit outside the unified API.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
All world states: each possible hidden object configuration
Equations
- One or more equations did not get rendered due to their size.
Instances For
Speaker credence: uniform over hidden objects given visual access.
P(world | access) = 1/64 if world.visible matches access, else 0. This encodes that speaker knows what's visible but is uncertain about hidden.
Equations
- HawkinsGweonGoodman2021.visualAccessCredence access world = if (world.visible == access.visibleObjects && world.target == access.targetObject) = true then 1 / 64 else 0
Instances For
Literal meaning: utterance applies to target in this world context
Equations
Instances For
Grounding: The unified API's worldMeaning computes the same listener probability
as our manual literalListenerProb for each world configuration.
Grounding: Speaker credence in unified API marginalizes uniformly over hidden objects,
matching the manual uniformHiddenPrior.
Mixture Model (Implementation-Specific) #
The mixture model w_S · U_asym + (1-w_S) · U_ego and resource-rational
optimization for finding optimal w* are handled in Parts 6-8 above.
These are implementation-specific extensions that:
- Blend two reasoning modes (asymmetric vs egocentric)
- Find optimal effort allocation via cost-benefit analysis
The unified API handles the asymmetric case directly; the mixture and meta-cognitive choice of w* sit outside the core RSA loop.
Listener's belief about speaker's weight after observing utterances
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Initial uniform belief about speaker's weight
Equations
- HawkinsGweonGoodman2021.initialBeliefs = { wS_expectation := 1 / 2, observations := 0 }
Instances For
Update beliefs after observing speaker use short utterances. If speaker consistently uses minimal descriptions, listener infers low w_S.
Equations
- One or more equations did not get rendered due to their size.
Instances For
After seeing short utterances, listener expects lower w_S
Equations
- One or more equations did not get rendered due to their size.
Instances For
Resource-rational listener response: increase own perspective-taking when speaker is under-informative.
Equations
- HawkinsGweonGoodman2021.optimalListenerWeight speakerWS beta = 1 ⊓ (0 ⊔ (1 - speakerWS + beta))
Instances For
Listener increases effort when speaker decreases theirs
Key Predictions from Paper (Section 2.4.1) #
The paper identifies four key qualitative predictions, which we verify as theorems:
- speakersHedgeUnknowns: Speakers increase informativity with occlusions
- divisionDependsOnPartner: Optimal effort depends on expected partner effort
- listenersAdaptOverTime: Listeners update beliefs about speaker from observations
- intermediateWeightsOptimal: Partial perspective-taking when cost > 0
Paper Prediction 1: Speakers hedge against known unknowns.
From the paper: "speakers will anticipate possible confusion from the listener's perspective, and produce additional information beyond what would be necessary from their own viewpoint."
Verified by: asymmetric informativity favors more specific utterances.
Paper Prediction 2: Division of labor depends on partner's expected effort.
From the paper: "The effort one participant ought to exert depends on how much effort they expect others to exert."
Verified by: at different listener weights, speaker utility differs. This shows speaker decisions depend on beliefs about listener.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Paper Prediction 3: Listeners adapt over time.
From the paper: "listeners used violations to adaptively make fewer errors over time" (z = 2.6, p < 0.01)
Verified by: beliefs about speaker weight decrease when observing short utterances.
Resource-rational utility at a given perspective weight and cost coefficient
Equations
- One or more equations did not get rendered due to their size.
Instances For
Paper Prediction 4: Intermediate weights are optimal when β > 0.
From the paper (Figure 2): "Above a certain β (i.e., if perspective-taking is sufficiently effortful), an intermediate weighting of perspective-taking is boundedly optimal."
At β = 0.2: w*_S = 0.36, w*_L = 0.51
Note: At β = 0, egocentric may have higher raw informativity (since it doesn't average over hidden distractors). But at β > 0, the cost term creates a trade-off where intermediate weights become optimal. The key insight is that speakers should choose MORE INFORMATIVE utterances (like fullDescription) rather than shapeOnly when doing perspective-taking - that's where the benefit comes.
Equations
Instances For
Equations
Instances For
Equations
Instances For
Paper Prediction 4 (continued): Intermediate weights optimal.
When cost is moderate, the optimal weight is strictly between 0 and 1. This matches Figure 2 of the paper where w*_S ≈ 0.36 at β = 0.2.
Empirical Findings from Paper #
Experiment 1 (Speaker Production, N=83 dyads) #
- Occlusion effect: +1.3 words, t(120.3) = 8.8, p < .001
- Distractor effect: +0.6 words, t(206) = 5.7, p < .001
Experiment 2 (Listener Comprehension, N=116 dyads) #
- Scripted: 51% critical errors
- Unscripted: 20% critical errors
- χ²(1) = 43, p < .001
Equations
Instances For
Equations
Instances For
Model correctly predicts that more informative speakers lead to fewer errors
Informativity-error correlation from paper: ρ = -0.81
Equations
Instances For
Model Summary #
Key model predictions verified as theorems:
more_specific_higher_asymmetric_informativity: More specific utterances have higher informativity when considering hidden objectsasymmetry_increases_specificity_gain: The asymmetric model predicts LARGER informativity gain from specificity than egocentricfull_description_preferred_at_wS1: At full perspective-taking, more specific utterances maximize listener successshape_only_sufficient_at_wS0: At pure egocentric, minimal description is equally informative (target unique in shape)listener_infers_low_wS_from_short_utterances: Listeners infer speaker's low effort from under-informative utteranceslistener_compensates_for_low_speaker_effort: Optimal listener effort increases when speaker effort is lowsemantics_grounded: Utterance semantics grounded in compositional (Montague) denotations