Documentation

Linglib.Phenomena.Agreement.Studies.PrasertsonSmithCulbertson2026

Prasertsom, Smith & Culbertson (2026) #

@cite{prasertsom-smith-culbertson-2026}

Domain-general categorisation explains constrained cross-linguistic variation in noun classification. Cognition 271, 106411.

Key Claims #

@cite{prasertsom-smith-culbertson-2026} ask why animacy — but not colour — appears in noun classification systems, despite colour being perceptually salient. They argue the answer lies in a domain-general categorisation principle: features with higher predictive power (the ability to predict other features of category members) yield more compact, more distinct categories, which are easier to learn. Animacy is more predictive than colour, explaining its cross-linguistic prevalence via cultural transmission amplification (@cite{griffiths-kalish-2007}, @cite{kirby-et-al-2007}, @cite{smith-2011}).

Argument Chain #

  1. Typological observation (§1): Animacy is universal in noun categorization (@cite{aikhenvald-2000}); colour is absent from every attested system — despite high perceptual salience.
  2. Learning bias (Exp 1a–b): Participants learn animacy-based artificial noun classes better than colour-based ones.
  3. Domain generality (Exp 2a–c): Non-linguistic sorting also shows animacy preference, ruling out grammar-specific mechanisms.
  4. Mechanism (§4.1, Table 6): Predictive power — animacy predicts other features (shape, horn, appendages) better than colour does, yielding more compact categories (Table 2) and higher classifier accuracy (Table 3).
  5. Causal evidence (Exp 3a–b): Manipulating predictive power modulates sorting preferences, though the effect on noun class learning is indirect — participants who notice colour's predictive structure and sort by colour learn animacy-based noun classes worse.

Formalization #

The typological observation (§1) connects to existing NounCategorization infrastructure. Predictive power is formalized via Table 6's conditional entropy data from the experimental stimuli. The connection to Core.Prominence.AnimacyLevel — the Silverstein hierarchy — makes explicit that animacy's grammatical privilege spans both case marking and noun classification.

Whether a semantic parameter is attested in any system in our typology. Derived from allSystems data rather than stipulated.

Equations
  • One or more equations did not get rendered due to their size.
Instances For

    Features in the experimental stimuli (Fig. 10, Tables 4–5).

    Instances For
      Equations
      • One or more equations did not get rendered due to their size.
      Instances For

        Conditional entropy H(row|col) from Table 6. Lower values mean the column feature is more predictive of the row feature. Values are in bits; the maximum for a binary feature is 1.0, for a 4-valued feature is 2.0.

        Instances For
          Equations
          • One or more equations did not get rendered due to their size.
          Instances For

            Table 6 data, predictive-animacy stimulus set. In this set, animacy is the predictive feature: knowing animacy reduces uncertainty about horn/shape/appendages more than knowing colour does. H(Horn|Animacy) = 0.406 < H(Horn|Colour) = 0.906.

            Equations
            • One or more equations did not get rendered due to their size.
            Instances For
              Equations
              • One or more equations did not get rendered due to their size.
              Instances For
                Equations
                • One or more equations did not get rendered due to their size.
                Instances For
                  Equations
                  • One or more equations did not get rendered due to their size.
                  Instances For
                    Equations
                    • One or more equations did not get rendered due to their size.
                    Instances For
                      Equations
                      • One or more equations did not get rendered due to their size.
                      Instances For

                        Table 6: animacy is more predictive of horn than colour is. H(horn|animacy) = 0.406 < H(horn|colour) = 0.906.

                        Table 6, "Both stimulus sets" column (reverse direction): Horn/shape/appendages predict animacy better than they predict colour. H(Animacy|Horn) = 1.451 < H(Colour|Horn) = 1.951. This is consistent: predictive power is symmetric — if animacy predicts horn well, horn also predicts animacy well.

                        Equations
                        • One or more equations did not get rendered due to their size.
                        Instances For
                          Equations
                          • One or more equations did not get rendered due to their size.
                          Instances For

                            Semantic basis for categorization (both experimental and typological).

                            Instances For
                              Equations
                              • One or more equations did not get rendered due to their size.
                              Instances For

                                Intra-category similarity from Word2Vec embeddings of 472 frequent physical nouns from CHILDES (§4.1.1). Higher InSim means category members are more similar to each other in distributional semantic space.

                                Instances For
                                  Equations
                                  • One or more equations did not get rendered due to their size.
                                  Instances For
                                    Equations
                                    • One or more equations did not get rendered due to their size.
                                    Instances For
                                      Equations
                                      • One or more equations did not get rendered due to their size.
                                      Instances For
                                        Equations
                                        • One or more equations did not get rendered due to their size.
                                        Instances For
                                          Equations
                                          • One or more equations did not get rendered due to their size.
                                          Instances For

                                            InSim(Avg) × 1000: weighted average across categories. Animacy: 0.149, Colour: 0.147.

                                            Equations
                                            Instances For

                                              Animacy-based categories are more compact on average (Table 2). InSim(Avg) animacy (0.149) > InSim(Avg) colour (0.147).

                                              5-fold cross-validated logistic classifier accuracy from normalised Word2Vec embeddings (§4.1.2, Table 3).

                                              Instances For
                                                Equations
                                                • One or more equations did not get rendered due to their size.
                                                Instances For

                                                  Result of a noun class learning experiment.

                                                  Instances For
                                                    Equations
                                                    • One or more equations did not get rendered due to their size.
                                                    Instances For

                                                      Experiment 1a: Ease of learning (§2.1). Animacy condition: 0.915 mean accuracy. Colour condition: 0.841 mean accuracy. β_SemanticBasis = 1.859, p = 0.023.

                                                      Equations
                                                      Instances For

                                                        Experiment 1b: Extrapolation from ambiguous data (§2.2). When animacy and colour are confounded in training, 77.5% of participants (62/80) generalise to animacy-based classification at test.

                                                        Equations
                                                        Instances For

                                                          Result of an image sorting experiment (§3). Participants sort 16 images into two groups; sorting strategy is inferred from AMI with reference sorts. Proportions are expressed as parts per thousand.

                                                          • animacy_x1000 :

                                                            Proportion ×1000 who sorted by animacy

                                                          • colour_x1000 :

                                                            Proportion ×1000 who sorted by colour

                                                          • other_x1000 :

                                                            Proportion ×1000 who sorted by another strategy

                                                          Instances For
                                                            Equations
                                                            • One or more equations did not get rendered due to their size.
                                                            Instances For

                                                              Exp 2a: original stimuli (same as Exp 1).

                                                              Equations
                                                              Instances For

                                                                Exp 2b: increased within-category colour similarity.

                                                                Equations
                                                                Instances For

                                                                  Exp 2c: further decreased within-category animacy similarity.

                                                                  Equations
                                                                  Instances For

                                                                    The animacy bias persists even when visual similarity is manipulated to favour colour-based sorting (Exp 2b, 2c).

                                                                    The bias weakens as colour similarity increases — evidence that it is not absolute but can be modulated.

                                                                    Experiment 3a results (§4.2): predictive power manipulation had a main effect of semantic basis (β = 1.251, p = 0.003) but the interaction with predictive feature was NOT significant (β = 0.276, p = 0.519). The prediction was "not straightforwardly confirmed."

                                                                    However, the sorting strategy × semantic basis interaction WAS significant (β = 1.058, p < 0.001): participants who sorted by animacy learned animacy-based noun classes better.

                                                                    • animacyBasis_x1000 :

                                                                      Accuracy ×1000 for animacy-basis condition

                                                                    • colourBasis_x1000 :

                                                                      Accuracy ×1000 for colour-basis condition

                                                                    Instances For
                                                                      Equations
                                                                      • One or more equations did not get rendered due to their size.
                                                                      Instances For

                                                                        Exp 3b (§4.3): among participants who sorted by colour (successful manipulation), there was NO significant difference between animacy-based and colour-based noun class learning (0.800 vs 0.803). β = −0.510, p = 0.314.

                                                                        Equations
                                                                        Instances For

                                                                          Exp 3b: when the animacy bias is neutralised by selecting colour-sorters, the learning advantage for animacy disappears. This is consistent with the domain-general account — the bias is in categorisation, not in grammar.

                                                                          The animacy hierarchy that governs noun classification (@cite{aikhenvald-2000}, @cite{silverstein-1976}) is the same hierarchy that governs differential argument marking (Core.Prominence).

                                                                          @cite{prasertsom-smith-culbertson-2026} provide a domain-general explanation for WHY animacy is grammatically privileged in both domains: animacy is highly predictive of other referent features, making it a good basis for categorization in general.

                                                                          This theorem connects the prominence hierarchy to the noun categorization typology: the levels of AnimacyLevel correspond to the SemanticParameter.animacy / .humanness distinction.

                                                                          The paper's core argument, assembled from the evidence above:

                                                                          1. Colour is absent from all attested noun categorization systems
                                                                          2. Animacy is present in all attested systems
                                                                          3. Animacy is more predictive of other features than colour (Table 6)
                                                                          4. Animacy-based categories are more compact in distributional space (Table 2)
                                                                          5. Animacy-based categories are easier for classifiers to learn (Table 3)
                                                                          6. Humans learn animacy-based noun classes better (Exp 1a)
                                                                          7. Humans prefer animacy-based sorting even non-linguistically (Exp 2a–c)

                                                                          The connection is: predictive power → compactness → learnability → typological prevalence (via cultural transmission).