Documentation

Linglib.Phenomena.WordOrder.Studies.SchlotterbeckWang2023

@cite{schlotterbeck-wang-2023} — Incremental RSA for Adjective Ordering #

@cite{cohn-gordon-goodman-potts-2019} @cite{degen-etal-2020}

Schlotterbeck, F. & Wang, H. (2023). An incremental RSA model for adjective ordering preferences in referential visual context. Proceedings of the Society for Computation in Linguistics (SCiL) 6, 121–132.

The Model #

The incremental sequence speaker (S1^inc) produces adjective–noun sequences word-by-word. At each step the utility is the incremental listener's posterior. The trajectory score accumulates utility across all prefixes:

S1^inc(w₁,...,wₙ | r) ∝ ∏ₖ U(w₁,...,wₖ; r)

where U(w⃗; r) = exp(β · log L0^inc(r | w⃗)) and the paper sets β = 1 in all reported simulations. With β = 1, no cost, and uniform language prior, this simplifies to:

S1^inc(w₁,...,wₙ | r) = ∏ₖ L0(r | w₁,...,wₖ)

The model uses continuous/noisy semantics (@cite{degen-etal-2020}) where each word applies with reliability v (correct application) or 1 − v (noise).

Key insight: With strictly positive noisy semantics, the prefix meaning is a product of per-word terms, and multiplication commutes. Therefore the full-sequence L0 posterior is order-independent: L0(r | w₁, w₂) = L0(r | w₂, w₁). In the paper's batch-normalized model, where S1^inc scores are normalized once over all trajectories, the ordering preference ratio S1^inc(adj₁,adj₂,n|r) / S1^inc(adj₂,adj₁,n|r) reduces entirely to the first-word L0 posterior ratio L0(r|adj₁) / L0(r|adj₂).

Formalization #

This uses RSAConfig's sequential infrastructure (following @cite{cohn-gordon-goodman-potts-2019} and @cite{waldon-degen-2021}):

Predictions use trajectoryProb for ordering preferences and S1_at for first-word informativity, proved via rsa_predict.

Findings #

#FindingTheorem
1Prefix meaning is order-independentprefix_meaning_swap
2Size discriminatory → size-first preferredsize_first_when_size_discriminates
3Equal discrimination + color reliable → color-firstcolor_first_when_color_reliable
4Both orderings identify the target (A)both_orderings_identify_target_A
5Both orderings identify the target (B)both_orderings_identify_target_B

Connections #

Simplifications #

The paper's full model includes components not formalized here:

  1. Gaussian+binomial perception: The paper models size via Gaussian distributions with Weber fractions and color via binomial noise ε (@cite{degen-etal-2020}). Core.Agent.Psychophysics formalizes the Stevens power law and multidimensional decomposition that underlie Weber's law; a future integration could derive size reliability from this framework. We currently use a simpler noise model with flat reliability parameters sRel and cRel.
  2. Language model P_Lang: The paper constrains the S1 vocabulary at each step to grammatically valid continuations (noun vs adjective). Our S1 distributes over all 6 words at each step. This does not affect the qualitative ordering predictions.
  3. S1^{inc_utt} vs S1^inc: The paper defines both a word-level speaker (S1^inc, used for data fitting with β = 1) and an utterance-level speaker (S1^{inc_utt}). We formalize S1^inc.
  4. Bias parameter b: The paper includes a prior bias b for size-first ordering (to account for language-specific defaults). We omit this.

The specific reliability values (sRel, cRel) are chosen to demonstrate the qualitative predictions — they are not taken from the paper's fitted values.

Referents in the reference game. Flat enum with Fintype for RSAConfig.

Instances For
    Equations
    • One or more equations did not get rendered due to their size.
    Equations
    • One or more equations did not get rendered due to their size.
    Instances For

      Words available to the speaker: size adjectives, color adjectives, noun.

      Instances For
        Equations
        • One or more equations did not get rendered due to their size.
        Equations
        • One or more equations did not get rendered due to their size.
        Instances For

          Noisy word meaning: returns reliability if the word truly applies, noise floor (1 − reliability) otherwise. Simplified from @cite{degen-etal-2020}'s continuous semantics.

          Equations
          • One or more equations did not get rendered due to their size.
          Instances For

            Prefix meaning: product of noisy word meanings over a word sequence. This implements the Product of Experts model from @cite{degen-etal-2020}: each word contributes an independent noisy channel value.

            Equations
            • One or more equations did not get rendered due to their size.
            Instances For
              theorem Phenomena.WordOrder.Studies.SchlotterbeckWang2023.lexContinuousQ_pos (sRel cRel : ) (w : Word) (r : Referent) (hs : 0 < sRel sRel < 1) (hc : 0 < cRel cRel < 1) :
              0 < lexContinuousQ sRel cRel w r

              Strict positivity: with reliability strictly between 0 and 1, every word–referent pair has a strictly positive noisy meaning value. This ensures the incremental L0 is well-defined (no zero denominators).

              noncomputable def Phenomena.WordOrder.Studies.SchlotterbeckWang2023.mkIncRSA (scene : ReferentBool) (sRel cRel : ) (hs : 0 sRel sRel 1) (hc : 0 cRel cRel 1) :

              Incremental RSA for adjective ordering, parameterized by scene and perceptual reliability. Uses RSAConfig's sequential infrastructure:

              • L0 uses product-of-experts noisy semantics
              • S1 uses identity scoring (β = 1, no cost)
              • trajectoryProb chains word-by-word S1 probabilities
              Equations
              • One or more equations did not get rendered due to their size.
              Instances For

                Scene A config: sizeRel = 99/100, colorRel = 95/100.

                Equations
                • One or more equations did not get rendered due to their size.
                Instances For

                  Scene B config: sizeRel = 80/100, colorRel = 95/100.

                  Equations
                  • One or more equations did not get rendered due to their size.
                  Instances For

                    Size-first ordering for the big-blue target.

                    Equations
                    • One or more equations did not get rendered due to their size.
                    Instances For

                      Color-first ordering for the big-blue target.

                      Equations
                      • One or more equations did not get rendered due to their size.
                      Instances For

                        Prefix meaning for two words is order-independent. This follows from commutativity of ℚ multiplication: foldl(*lex) 1 [a,b] = lex(a)·lex(b) = lex(b)·lex(a) = foldl(*lex) 1 [b,a].

                        Prefix meaning for three words is independent of the first two words' order. Swapping the adjectives before the noun does not change the product semantics.

                        Two-word prefix meaning decomposes as a product of per-word noisy meanings. This is the Product of Experts (PoE) structure from @cite{degen-etal-2020}: each word contributes an independent noisy channel value.

                        Finding: When both properties discriminate equally but color is more reliable (Scene B), S1^inc prefers color-first ordering.

                        The ordering preference flips between scenes: Scene A prefers size-first, Scene B prefers color-first. This captures @cite{schlotterbeck-wang-2023}'s key prediction: the preferred ordering depends on the discriminatory structure of the scene, not a fixed ordering rule.

                        After hearing both adjectives, the meaning function assigns highest value to the target among Scene A members.

                        After hearing both adjectives, the meaning function assigns highest value to the target among Scene B members.

                        lexContinuousQ is an instance of the unified noise channel from RSA.Core.Noise. The continuous lexical semantics is exactly the noise channel with onMatch = reliability, onMismatch = 1 − reliability.

                        This connects @cite{schlotterbeck-wang-2023} to the @cite{degen-etal-2020} parameterization where mismatch = 1 − match.

                        Qualitative findings from the incremental RSA adjective ordering model.

                        • prefix_order_independent : Finding

                          Prefix meaning is order-independent for any two words.

                        • size_first_when_size_discriminates : Finding

                          When size has high discriminatory power, size-first ordering is preferred: trajectoryProb(size-first) > trajectoryProb(color-first).

                        • color_first_when_color_reliable : Finding

                          When both properties discriminate equally but color is more reliable, color-first is preferred.

                        • both_orderings_identify_target_A : Finding

                          The meaning function correctly identifies the target (scene A).

                        • both_orderings_identify_target_B : Finding

                          The meaning function correctly identifies the target (scene B).

                        Instances For
                          Equations
                          • One or more equations did not get rendered due to their size.
                          Instances For

                            Map each finding to the model prediction that accounts for it.

                            Equations
                            • One or more equations did not get rendered due to their size.
                            Instances For