Prasertsom, Smith & Culbertson (2026) #
@cite{prasertsom-smith-culbertson-2026}
Domain-general categorisation explains constrained cross-linguistic variation in noun classification. Cognition 271, 106411.
Key Claims #
@cite{prasertsom-smith-culbertson-2026} ask why animacy — but not colour — appears in noun classification systems, despite colour being perceptually salient. They argue the answer lies in a domain-general categorisation principle: features with higher predictive power (the ability to predict other features of category members) yield more compact, more distinct categories, which are easier to learn. Animacy is more predictive than colour, explaining its cross-linguistic prevalence via cultural transmission amplification (@cite{griffiths-kalish-2007}, @cite{kirby-et-al-2007}, @cite{smith-2011}).
Argument Chain #
- Typological observation (§1): Animacy is universal in noun categorization (@cite{aikhenvald-2000}); colour is absent from every attested system — despite high perceptual salience.
- Learning bias (Exp 1a–b): Participants learn animacy-based artificial noun classes better than colour-based ones.
- Domain generality (Exp 2a–c): Non-linguistic sorting also shows animacy preference, ruling out grammar-specific mechanisms.
- Mechanism (§4.1, Table 6): Predictive power — animacy predicts other features (shape, horn, appendages) better than colour does, yielding more compact categories (Table 2) and higher classifier accuracy (Table 3).
- Causal evidence (Exp 3a–b): Manipulating predictive power modulates sorting preferences, though the effect on noun class learning is indirect — participants who notice colour's predictive structure and sort by colour learn animacy-based noun classes worse.
Formalization #
The typological observation (§1) connects to existing NounCategorization
infrastructure. Predictive power is formalized via Table 6's conditional
entropy data from the experimental stimuli. The connection to
Core.Prominence.AnimacyLevel — the Silverstein hierarchy —
makes explicit that animacy's grammatical privilege spans both case marking
and noun classification.
Whether a semantic parameter is attested in any system in our typology.
Derived from allSystems data rather than stipulated.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Animacy is attested in the typological data.
Humanness is attested in the typological data.
Shape is attested in the typological data.
Colour is NOT attested in any system in our typology.
The central asymmetry: animacy is attested, colour is not.
Features in the experimental stimuli (Fig. 10, Tables 4–5).
- animacy : StimulusFeature
- colour : StimulusFeature
- horn : StimulusFeature
- shape_ : StimulusFeature
- appendages : StimulusFeature
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Conditional entropy H(row|col) from Table 6. Lower values mean the column feature is more predictive of the row feature. Values are in bits; the maximum for a binary feature is 1.0, for a 4-valued feature is 2.0.
- row : StimulusFeature
- col : StimulusFeature
- hBits_x1000 : ℕ
H(row|col) in bits, scaled ×1000 for exact rational arithmetic
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Table 6 data, predictive-animacy stimulus set. In this set, animacy is the predictive feature: knowing animacy reduces uncertainty about horn/shape/appendages more than knowing colour does. H(Horn|Animacy) = 0.406 < H(Horn|Colour) = 0.906.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Table 6: animacy is more predictive of horn than colour is. H(horn|animacy) = 0.406 < H(horn|colour) = 0.906.
Table 6: animacy is more predictive of shape than colour is.
Table 6: animacy is more predictive of appendages than colour is.
Table 6, "Both stimulus sets" column (reverse direction): Horn/shape/appendages predict animacy better than they predict colour. H(Animacy|Horn) = 1.451 < H(Colour|Horn) = 1.951. This is consistent: predictive power is symmetric — if animacy predicts horn well, horn also predicts animacy well.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Semantic basis for categorization (both experimental and typological).
- animacy : SemanticBasis
- colour : SemanticBasis
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Intra-category similarity from Word2Vec embeddings of 472 frequent physical nouns from CHILDES (§4.1.1). Higher InSim means category members are more similar to each other in distributional semantic space.
- basis : SemanticBasis
- category : String
- inSim_x1000 : ℕ
Mean cosine similarity × 1000 (for exact arithmetic)
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
InSim(Avg) × 1000: weighted average across categories. Animacy: 0.149, Colour: 0.147.
Instances For
Animacy-based categories are more compact on average (Table 2). InSim(Avg) animacy (0.149) > InSim(Avg) colour (0.147).
The animate category is the most compact single category.
5-fold cross-validated logistic classifier accuracy from normalised Word2Vec embeddings (§4.1.2, Table 3).
- basis : SemanticBasis
- accuracy_x1000 : ℕ
Mean accuracy × 1000
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
Instances For
Equations
Instances For
Animacy categories are easier for a classifier to learn (Table 3).
Result of a noun class learning experiment.
- basis : SemanticBasis
- accuracy_x1000 : ℕ
Mean proportion correct × 1000
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Experiment 1a: Ease of learning (§2.1). Animacy condition: 0.915 mean accuracy. Colour condition: 0.841 mean accuracy. β_SemanticBasis = 1.859, p = 0.023.
Equations
Instances For
Equations
Instances For
Experiment 1b: Extrapolation from ambiguous data (§2.2). When animacy and colour are confounded in training, 77.5% of participants (62/80) generalise to animacy-based classification at test.
Instances For
Result of an image sorting experiment (§3). Participants sort 16 images into two groups; sorting strategy is inferred from AMI with reference sorts. Proportions are expressed as parts per thousand.
- animacy_x1000 : ℕ
Proportion ×1000 who sorted by animacy
- colour_x1000 : ℕ
Proportion ×1000 who sorted by colour
- other_x1000 : ℕ
Proportion ×1000 who sorted by another strategy
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Exp 2a: original stimuli (same as Exp 1).
Equations
- Phenomena.Agreement.Studies.PrasertsonSmithCulbertson2026.exp2a = { animacy_x1000 := 875, colour_x1000 := 25, other_x1000 := 100 }
Instances For
Exp 2b: increased within-category colour similarity.
Equations
- Phenomena.Agreement.Studies.PrasertsonSmithCulbertson2026.exp2b = { animacy_x1000 := 683, colour_x1000 := 150, other_x1000 := 167 }
Instances For
Exp 2c: further decreased within-category animacy similarity.
Equations
- Phenomena.Agreement.Studies.PrasertsonSmithCulbertson2026.exp2c = { animacy_x1000 := 617, colour_x1000 := 183, other_x1000 := 200 }
Instances For
The animacy bias persists even when visual similarity is manipulated to favour colour-based sorting (Exp 2b, 2c).
The bias weakens as colour similarity increases — evidence that it is not absolute but can be modulated.
Experiment 3a results (§4.2): predictive power manipulation had a main effect of semantic basis (β = 1.251, p = 0.003) but the interaction with predictive feature was NOT significant (β = 0.276, p = 0.519). The prediction was "not straightforwardly confirmed."
However, the sorting strategy × semantic basis interaction WAS significant (β = 1.058, p < 0.001): participants who sorted by animacy learned animacy-based noun classes better.
- animacyBasis_x1000 : ℕ
Accuracy ×1000 for animacy-basis condition
- colourBasis_x1000 : ℕ
Accuracy ×1000 for colour-basis condition
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Exp 3b (§4.3): among participants who sorted by colour (successful manipulation), there was NO significant difference between animacy-based and colour-based noun class learning (0.800 vs 0.803). β = −0.510, p = 0.314.
Instances For
Exp 3b: when the animacy bias is neutralised by selecting colour-sorters, the learning advantage for animacy disappears. This is consistent with the domain-general account — the bias is in categorisation, not in grammar.
The animacy hierarchy that governs noun classification
(@cite{aikhenvald-2000}, @cite{silverstein-1976}) is the same hierarchy
that governs differential argument marking (Core.Prominence).
@cite{prasertsom-smith-culbertson-2026} provide a domain-general explanation for WHY animacy is grammatically privileged in both domains: animacy is highly predictive of other referent features, making it a good basis for categorization in general.
This theorem connects the prominence hierarchy to the noun categorization
typology: the levels of AnimacyLevel correspond to the
SemanticParameter.animacy / .humanness distinction.
The paper's core argument, assembled from the evidence above:
- Colour is absent from all attested noun categorization systems
- Animacy is present in all attested systems
- Animacy is more predictive of other features than colour (Table 6)
- Animacy-based categories are more compact in distributional space (Table 2)
- Animacy-based categories are easier for classifiers to learn (Table 3)
- Humans learn animacy-based noun classes better (Exp 1a)
- Humans prefer animacy-based sorting even non-linguistically (Exp 2a–c)
The connection is: predictive power → compactness → learnability → typological prevalence (via cultural transmission).