Documentation

Linglib.Phenomena.ScalarImplicatures.Studies.CremersWilcoxSpector2023

@cite{cremers-wilcox-spector-2023}: Exhaustivity and Anti-Exhaustivity in RSA #

"Exhaustivity and Anti-Exhaustivity in the RSA Framework: Testing the Effect of Prior Beliefs." Cognitive Science 47(5), e13286.

The Symmetry Problem #

In baseline RSA, hearing "A" can increase belief in w_ab (where both A and B are true) when priors are biased toward w_ab. This "anti-exhaustive" behavior contradicts human interpretation — humans always exhaust "A" to mean "A and not B."

Key Result #

Models with grammatical exhaustification (EXH-LU, RSA-LI) block anti-exhaustivity regardless of prior bias. The EXH parse makes A false in w_ab, so S1 never uses A to convey w_ab under that parse. The prior bias is overridden by the structural asymmetry in meaning.

The paper's experimental data (comprehension task) found no evidence of anti-exhaustivity: listeners consistently infer "A and not B" from "A", regardless of prior bias. This supports grammatical models (EXH-LU, RSA-LI, svRSA) over baseline RSA and wRSA.

Domain #

Prior Placement #

The paper's L0 includes the prior (eq. 1): L0(w|u) ∝ P(w) · ⟦u⟧(w). The paper also has P(w) at L1 (eq. 3): L1(w|u) ∝ P(w) · S1(u|w). So P(w) enters twice — in L0 and L1.

For models where prior biases L0 (baseline, EXH-LU, wRSA, FREE-LU), we bake P(w) into meaning and set worldPrior = 1. This is equivalent to double-counting for anti-exhaustivity classification (Appendix A.1, eq. A.4): L1(w_ab|A) > P(w_ab) ⟺ T(w_ab) > T(w_a) where T(w) = Σ_l P(l)·S1(A|w,l), regardless of worldPrior. With worldPrior = 1, L1_score(w) = T(w), so the comparison L1(.A, .w_ab) > L1(.A, .w_a) IS the anti-exhaustivity test.

For svRSA, the prior does NOT enter meaning. Under Q_A (coarse QUD), QUD projection aggregates all A-true worlds into one cell, making L0(Q_A(w)|A) = 1 for all w. This neutralizes the prior's effect on S1, giving world-independent S1 under Q_A (Appendix A.3). Our encoding captures this by using uniform weights in meaning under both QUDs: L0(w|A, coarse) = 1/2 for both worlds.

Anti-Exhaustivity Definition #

The paper's definition (Eq. 6b, equal costs): L1(w_ab|A) > P(w_ab), equivalently T(w_ab) > T(w_a) (Appendix A.1, eq. A.4). With our encoding (worldPrior = 1), this reduces to L1(.A, .w_ab) > L1(.A, .w_a).

Utterance Costs #

The paper includes costs c(A) = 0, c(A∧B) = c(A∧¬B) = c (eq. 2): S1(u|w) ∝ exp(λ(log L0(w|u) - c(u))) = L0(w|u)^λ · exp(-λc(u)). With equal costs for A∧B and A∧¬B, the cost terms cancel in the anti-exhaustivity condition (eq. 6b): the classification depends only on the log-prior ratio. We set all costs to 0 in our configs; the analytic condition in antiExhaustivityCondition handles the general case.

Model Classification #

The paper's 9 models fall into two groups by anti-exhaustivity behavior:

#ModelMechanismAnti-exhaustive?Config
1Baseline RSAYesbaselineBiased
2wRSAPrior-in-L0 via backgroundYeswRSABiased
3BwRSABayesian wRSAYes (= wRSA at L1)wRSABiased
4svRSA1QUD blocks at S1NosvRSABiased
5svRSA2QUD blocks at S1NosvRSABiased
6FREE-LUFree parse choiceYesfreeLUBiased
7EXH-LUEXH blocks at L0NoexhLUBiased
8RSA-LI1EXH blocks at L0No (= EXH-LU at L1)rsaLIBiased
9RSA-LI2EXH blocks at L0No (= EXH-LU at L1)rsaLIBiased

All 9 models are encoded as RSAConfig instances. Each model's distinctive mechanism maps to a different Latent type and meaning. Costs enter S1 via utteranceCost (= 0, so exp(-λ·0) = 1 is implicit); the analytic condition antiExhaustivityCondition handles general costs.

RSA-LI (Models 8–9) is @cite{franke-bergen-2020}'s Lexical Intentions model; at L1 with uniform P(i) and equal costs it equals EXH-LU (Table 1). wRSA and BwRSA are identical at L1 (BwRSA adds a Bayesian S2 layer). svRSA uses no prior in L0 meaning — QUD projection neutralizes it (Appendix A.3), giving categorical blocking.

Connection to Other Formalizations #

Two-world model for exhaustivity.

  • w_a: A true, B false (A ∧ ¬B)
  • w_ab: A and B both true (A ∧ B)

No w_b or w_none because we condition on A being true. The question is whether B is also true.

Instances For
    Equations
    • One or more equations did not get rendered due to their size.
    Instances For
      Equations
      • One or more equations did not get rendered due to their size.

      Three utterances in the minimal exhaustivity domain.

      Instances For
        Equations
        • One or more equations did not get rendered due to their size.
        Instances For
          Equations
          • One or more equations did not get rendered due to their size.
          Equations
          • One or more equations did not get rendered due to their size.

          Literal truth: which utterance is true in which world.

          Equations
          • One or more equations did not get rendered due to their size.
          Instances For

            All worlds for IE enumeration.

            Equations
            • One or more equations did not get rendered due to their size.
            Instances For

              Alternatives for each utterance: A has scale-mate A∧B.

              Equations
              • One or more equations did not get rendered due to their size.
              Instances For

                Exhaustified meaning derived via Innocent Exclusion. IE negates A∧B (the non-entailed stronger alternative to A), giving EXH(A) = A ∧ ¬(A∧B) = A ∧ ¬B.

                Equations
                • One or more equations did not get rendered due to their size.
                Instances For

                  Parse = whether EXH is applied.

                  Instances For
                    Equations
                    • One or more equations did not get rendered due to their size.
                    Instances For
                      Equations
                      • One or more equations did not get rendered due to their size.

                      Meaning with parse-dependent exhaustification.

                      Equations
                      • One or more equations did not get rendered due to their size.
                      Instances For

                        Prior weight for each world: 1:3 bias toward w_ab.

                        Baked into meaning so that L0(w|u) ∝ P(w) · ⟦u⟧(w), matching the paper's eq. (1). This makes S1 asymmetric for utterances true in both worlds, which is the source of anti-exhaustivity.

                        Equations
                        Instances For

                          The paper's analytic condition for when baseline RSA is anti-exhaustive.

                          Full condition (Eq. 6b): log P(w_ab) - log P(w_a) > c(A∧¬B) - c(A∧B). When the cost of the exhaustive paraphrase A∧¬B exceeds the conjunctive alternative A∧B, the speaker's dispreference for the paraphrase makes anti-exhaustivity easier to trigger. With equal costs, any prior bias toward w_ab suffices.

                          Equations
                          Instances For

                            Equal costs: anti-exhaustivity iff P(w_ab) > P(w_a) (i.e., log ratio > 0).

                            Asymmetric costs can block anti-exhaustivity even with biased prior: if c(A∧¬B) is much more expensive than c(A∧B), the cost gap c(A∧¬B) - c(A∧B) exceeds the log prior ratio.

                            Baseline RSA (Model 1) with biased prior (1:3 toward w_ab).

                            Prior enters L0 via meaning: meaning(u,w) = P(w) · ⟦u⟧(w). Since A is true in both worlds, L0(w_a|A) = 1/4, L0(w_ab|A) = 3/4. S1 is asymmetric: S1(A|w_ab) = 9/25 > S1(A|w_a) = 1/17 (at α=2). This drives genuine anti-exhaustivity: the speaker preferentially uses A when w_ab is true (because L0 already favors w_ab given A).

                            Equations
                            • One or more equations did not get rendered due to their size.
                            Instances For

                              EXH-LU RSA (Model 7) with biased prior.

                              Parse is a latent variable; meaning under each parse includes the prior: meaning(p,u,w) = P(w) · ⟦u⟧_p(w). Under the exh parse, EXH(A) is false at w_ab, so meaning = P(w_ab) · 0 = 0 and S1(A|w_ab, exh) = 0. The exh parse's blocking contribution (S1(A|w_a, exh) = 1/2) outweighs the literal parse's prior amplification, giving T(w_a) > T(w_ab) despite the 3:1 bias.

                              Equations
                              • One or more equations did not get rendered due to their size.
                              Instances For

                                Baseline RSA is genuinely anti-exhaustive: L1(w_ab|A) > L1(w_a|A).

                                With prior-in-L0, S1 is asymmetric: S1(A|w_ab) = 9/25 > 1/17 = S1(A|w_a) because L0(w_ab|A) = 3/4 > 1/4 = L0(w_a|A). The speaker preferentially uses the cheap utterance A when w_ab is true, since L0 already assigns high probability to w_ab. This is the paper's central problematic prediction (Eq. 6b, Fig. 1).

                                EXH-LU blocks anti-exhaustivity: even with prior-in-L0 and 3:1 bias, L1 correctly infers w_a from "A". The exh parse contributes S1(A|w_a, exh) = 1/2 but S1(A|w_ab, exh) = 0, which outweighs the literal parse's prior amplification (S1(A|w_ab, lit) = 9/25 vs S1(A|w_a, lit) = 1/17). Total: T(w_a) = 19/34 > 9/25 = T(w_ab).

                                Background assumption for wRSA (@cite{degen-etal-2015}).

                                • wonky: unusual situation, uniform prior over worlds
                                • default_: default assumption, prior follows the bias
                                Instances For
                                  Equations
                                  • One or more equations did not get rendered due to their size.
                                  Instances For
                                    Equations
                                    • One or more equations did not get rendered due to their size.
                                    Equations
                                    • One or more equations did not get rendered due to their size.

                                    QUD for svRSA (@cite{spector-2017}).

                                    • coarse: Q_A — is A true? Corresponds to Discourse.QUD.trivial. All A-true worlds are equivalent; QUD projection gives L0 = 1.
                                    • fine: Q_fine — which world? Corresponds to Discourse.QUD.exact. Each world is its own cell; S1(A|w_ab) = 0 under exh interpretation.
                                    Instances For
                                      Equations
                                      • One or more equations did not get rendered due to their size.
                                      Instances For
                                        Equations
                                        • One or more equations did not get rendered due to their size.

                                        Interpretation for FREE-LU (@cite{bergen-levy-goodman-2016}).

                                        • literal: A true in both worlds
                                        • exh: A ∧ ¬B (true only in w_a)
                                        • antiExh: A ∧ B (true only in w_ab)
                                        Instances For
                                          Equations
                                          • One or more equations did not get rendered due to their size.
                                          Equations
                                          • One or more equations did not get rendered due to their size.
                                          Instances For
                                            Equations
                                            • One or more equations did not get rendered due to their size.

                                            Anti-exhaustive meaning: A means A ∧ B (true only in w_ab). A∧B and A∧¬B are unaffected (already fully specified).

                                            Equations
                                            Instances For

                                              Meaning under each interpretation (for FREE-LU and EXH-LU).

                                              Equations
                                              • One or more equations did not get rendered due to their size.
                                              Instances For

                                                wRSA (Models 2–3) with biased prior. BwRSA is identical at L1.

                                                Background is a latent variable. Under the default background, the prior P(w|default) = P(w) is baked into meaning (matching the paper's eq. 1 in §4.1: L0(w|u,b) ∝ P(w|b) · ⟦u⟧(w)). Under the wonky background, P(w|wonky) = uniform, so meaning uses unit weights.

                                                The world-dependent latentPrior encodes P(w|b) × P(b): under default_, w_ab gets 3× the weight of w_a; under wonky, both worlds get equal weight. worldPrior is uniform since the prior enters through meaning (for L0) and latentPrior (for L1).

                                                Equations
                                                • One or more equations did not get rendered due to their size.
                                                Instances For

                                                  svRSA (Models 4–5) — @cite{spector-2017} supervaluationist RSA.

                                                  The paper's svRSA (§4.2, eqs 3–8) has S1 compute expected utility over interpretations i, parameterized by QUD Q. The key structural properties (Appendix A.3, B.3) are:

                                                  1. Under Q_A (coarse): QUD projection makes L0(Q_A(w)|A) = 1 for all w (all A-true worlds are in one cell). So S1(A|w, Q_A) is the same at both worlds — the prior's effect on S1 is neutralized.

                                                  2. Under Q_fine: the exh interpretation makes A false at w_ab, so S1(A|w_ab, Q_fine) = 0 (log(0) = -∞ kills the expected utility).

                                                  Our encoding approximates this by NOT including the prior in L0 meaning. Under coarse QUD, L0(w|A) = 1/2 for both worlds → S1 world-independent. Under fine QUD, exhMeaning makes A false at w_ab → S1(A|w_ab) = 0.

                                                  This gives categorical blocking: T(w_a) = T(w_ab) + S1(A|w_a, fine)

                                                  T(w_ab) for any positive S1(A|w_a, fine), matching Appendix A.3. Concretely: T(w_a) = 1/5 + 1/2 = 7/10 > 1/5 = T(w_ab).

                                                  The QUD values correspond to Discourse.QUD.trivial (coarse, Q_A) and Discourse.QUD.exact (fine, Q_fine) from Core/Discourse/QUD.lean.

                                                  svRSA1 and svRSA2 differ only at S2 (production); both block at L1.

                                                  Equations
                                                  • One or more equations did not get rendered due to their size.
                                                  Instances For

                                                    FREE-LU (Model 6) with biased prior, prior-in-L0.

                                                    Three interpretations as latent variables: literal, exhaustive, and anti-exhaustive. With prior-in-L0, the literal parse is asymmetric (S1(A|w_ab, lit) = 9/25 > 1/17 = S1(A|w_a, lit)), making the anti-exh contribution dominant at w_ab. Total: T(w_ab) = 9/25 + 0 + 1/2 = 43/50 > 19/34 = T(w_a). This is genuine anti-exhaustivity, matching the paper's prediction.

                                                    Equations
                                                    • One or more equations did not get rendered due to their size.
                                                    Instances For

                                                      RSA-LI (Models 8–9) — @cite{franke-bergen-2020} Lexical Intentions.

                                                      The paper's RSA-LI (§4.4, eqs 1–6 on p.18) has S1 jointly choose (utterance, interpretation): S1(u,i|w) ∝ exp(λ·U1(u,i|w)), normalized over all (u',i') pairs. EXH-LU instead has S1(u|w,i) normalized per-i.

                                                      At L1, both give L1(w|u) ∝ P(w)·Σ_i S1(A|w,i) — the difference is only in how S1 is normalized. With uniform P(i) and equal costs (c(A∧B) ≤ c(A∧¬B)), both block anti-exhaustivity at L1 (Appendix B.5).

                                                      The structural reason (same for both): S1 never prefers A over A∧¬B at w_a (A∧¬B is at least as informative and has exh-compatible meaning), so S1(A|w_ab) ≤ S1(A|w_a) under both normalizations.

                                                      We alias rsaLIBiased := exhLUBiased. The models differ at S2 (RSA-LI's S2 cannot select a specific meaning, unlike EXH-LU's).

                                                      Equations
                                                      Instances For

                                                        wRSA is anti-exhaustive: the default background bakes the biased prior into L0, making S1 asymmetric. Even with the wonky background (uniform L0) averaging in, the biased background dominates.

                                                        svRSA blocks anti-exhaustivity categorically (Appendix A.3).

                                                        Under Q_A (coarse), S1(A|w, coarse) = 1/5 for both worlds (L0 is uniform since no prior in meaning). Under Q_fine, S1(A|w_ab, fine) = 0 (exh(A) false at w_ab) while S1(A|w_a, fine) = 1/2. So T(w_a) = 1/5 + 1/2 = 7/10 > 1/5 = T(w_ab). The blocking is structural: T(w_a) = T(w_ab) + S1(A|w_a, fine) > T(w_ab) for any parameters.

                                                        FREE-LU is genuinely anti-exhaustive: with prior-in-L0, the literal parse's S1 asymmetry makes the anti-exh interpretation dominant at w_ab. T(w_ab) = 9/25 + 0 + 1/2 = 43/50 > 19/34 = 1/17 + 1/2 + 0 = T(w_a).

                                                        The paper's central result: 9 RSA models fall into two groups.

                                                        Anti-exhaustive (L1(w_ab|A) > L1(w_a|A), i.e., L1(w_ab|A) > P(w_ab)): baseline RSA, wRSA (+ BwRSA at L1), FREE-LU. These models allow the prior bias to amplify through S1, predicting that listeners hearing "A" will infer w_ab — contradicting experimental data.

                                                        Exhaustive (L1(w_a|A) > L1(w_ab|A) despite biased prior): EXH-LU, RSA-LI1/2 (= EXH-LU at L1), svRSA1/2. These models have grammatical exhaustification or QUD-based gating that blocks anti-exhaustivity, correctly predicting the experimental data.