Giulianelli, Wallbridge, Cotterell & Fernández (2026) #
@cite{giulianelli-etal-2026}
Incremental alternative sampling as a lens into the temporal and representational resolution of linguistic prediction. Journal of Memory and Language, 148, 104715.
Key contributions #
- Defines incremental information value — the expected representational distance between alternatives sampled before vs. after observing a word
- Shows that different psycholinguistic measures are best predicted by different (forecast horizon, representational level) configurations
- The full IAS model (all h × all l) outperforms standard surprisal for most measures
- Surprisal implicitly integrates across multiple temporal and representational resolutions — its predictability is closest to a best-case (d^{min}) notion rather than average discrepancy (d^{mean})
Datasets #
- Aligned: 1,726 target–context pairs from 205 English sentences (predictability ratings, cloze, eye-tracking, ERPs, self-paced RT)
- Natural Stories: 10 English narratives, ~1,000 words each (self-paced RT via dashed moving-window display)
Language models #
GPT-2 in four sizes (Small 12L / Medium 24L / Large 36L / XL 48L), forecast horizons h ∈ 1..10, k = 50 sampled alternatives per timestep.
Connection to ProcessingModel #
Editorial note: The following bridge is our interpretation connecting
IAS's findings to linglib's ProcessingModel.ProcessingProfile. The paper
itself does not reference ProcessingProfile.
IAS's layer-wise decomposition reveals that different representational
levels predict different psycholinguistic measures. This parallels
ProcessingProfile's separation of locality, boundaries,
referentialLoad, and ease into orthogonal cognitive dimensions —
both architectures recognize that processing cost is irreducibly
multi-dimensional.
Equations
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
For each psycholinguistic measure, the forecast horizon at which incremental information value is most predictive (highest Δ_{Adj.R²}) in the sentence-level Aligned dataset. N400 and P600 peak at h = 2 (two-word lookahead); all other measures peak at h = 1 (next word).
Equations
Instances For
For each psycholinguistic measure, the representational level at which incremental information value is most predictive (primary peak).
- Explicit predictability (cloze, ratings): embedding layer (lexical identity)
- N400: embedding layer — shallow lexical-semantic prediction error
- Eye-tracked RT: early-to-intermediate layers — structural processing
- Self-paced RT and P600: intermediate layers — syntactic integration
Note: most measures show bimodal layer patterns; see layerPattern
for the full picture.
Equations
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.predictabilityRating = Core.GeneralisedSurprisal.RepLevel.lexical
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.clozeProbability = Core.GeneralisedSurprisal.RepLevel.lexical
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.clozeSurprisal = Core.GeneralisedSurprisal.RepLevel.lexical
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.n400 = Core.GeneralisedSurprisal.RepLevel.lexical
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.firstFixationRT = Core.GeneralisedSurprisal.RepLevel.shallowSyntactic
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.firstPassRT = Core.GeneralisedSurprisal.RepLevel.shallowSyntactic
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.rightBoundedRT = Core.GeneralisedSurprisal.RepLevel.shallowSyntactic
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.goPastRT = Core.GeneralisedSurprisal.RepLevel.shallowSyntactic
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.selfPacedRT = Core.GeneralisedSurprisal.RepLevel.syntactic
- Phenomena.Processing.Studies.GiulianelliEtAl2026.peakLevel Core.GeneralisedSurprisal.PsychMeasure.p600 = Core.GeneralisedSurprisal.RepLevel.syntactic
Instances For
Layer-activation pattern across representational depth.
The paper's central finding is that predictive power as a function of layer depth has distinct shapes for different measure classes.
- unimodal : LayerPattern
Single peak at one depth region (N400: embedding; P600: intermediate)
- uShaped : LayerPattern
Peaks at both boundary layers (embedding + final); characteristic of explicit predictability measures (cloze, ratings)
- sShaped : LayerPattern
Peaks at early-to-intermediate layers with secondary rise at final layer; characteristic of reading time measures
Instances For
Equations
Instances For
Equations
- One or more equations did not get rendered due to their size.
Instances For
Layer-activation pattern for each psycholinguistic measure.
- Explicit measures (cloze, ratings): U-shaped — peak at embedding and final layer, the two layers closest to lexical identity
- Reading times: S-shaped — peak at early-to-intermediate layers with secondary rise at the final layer
- N400: unimodal at embedding layer
- P600: unimodal at intermediate (syntactic) layers
Equations
- Phenomena.Processing.Studies.GiulianelliEtAl2026.layerPattern Core.GeneralisedSurprisal.PsychMeasure.predictabilityRating = Phenomena.Processing.Studies.GiulianelliEtAl2026.LayerPattern.uShaped
- Phenomena.Processing.Studies.GiulianelliEtAl2026.layerPattern Core.GeneralisedSurprisal.PsychMeasure.clozeProbability = Phenomena.Processing.Studies.GiulianelliEtAl2026.LayerPattern.uShaped
- Phenomena.Processing.Studies.GiulianelliEtAl2026.layerPattern Core.GeneralisedSurprisal.PsychMeasure.clozeSurprisal = Phenomena.Processing.Studies.GiulianelliEtAl2026.LayerPattern.uShaped
- Phenomena.Processing.Studies.GiulianelliEtAl2026.layerPattern Core.GeneralisedSurprisal.PsychMeasure.n400 = Phenomena.Processing.Studies.GiulianelliEtAl2026.LayerPattern.unimodal
- Phenomena.Processing.Studies.GiulianelliEtAl2026.layerPattern Core.GeneralisedSurprisal.PsychMeasure.p600 = Phenomena.Processing.Studies.GiulianelliEtAl2026.LayerPattern.unimodal
- Phenomena.Processing.Studies.GiulianelliEtAl2026.layerPattern x✝ = Phenomena.Processing.Studies.GiulianelliEtAl2026.LayerPattern.sShaped
Instances For
In Natural Stories (multi-sentence stimuli), self-paced reading time predictive power increases with forecast horizon up to h = 7, unlike sentence-level stimuli where h = 1 is optimal. The extended context and dashed moving-window display promote longer-range predictions.
Instances For
The full IAS model (all h × all l simultaneously) outperforms standard surprisal in predictive power for most psycholinguistic measures.
Exceptions:
- Predictability ratings: surprisal slightly better (not significant)
- Self-paced RT in Aligned: difference not significant
Equations
- One or more equations did not get rendered due to their size.
- Phenomena.Processing.Studies.GiulianelliEtAl2026.iasOutperformsSurprisal Core.GeneralisedSurprisal.PsychMeasure.predictabilityRating x✝ = false
- Phenomena.Processing.Studies.GiulianelliEtAl2026.iasOutperformsSurprisal x✝¹ x✝ = true
Instances For
Observed sign of the relationship between IAS information value and each psycholinguistic measure, as actually found in the data.
Differs from PsychMeasure.expectedSign for P600: the paper predicted a
positive relationship but found a negative one ("for P600, however, our
directional hypothesis is refuted").
Equations
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.predictabilityRating = -1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.clozeProbability = -1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.clozeSurprisal = 1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.firstFixationRT = 1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.firstPassRT = 1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.rightBoundedRT = 1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.goPastRT = 1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.selfPacedRT = 1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.n400 = -1
- Phenomena.Processing.Studies.GiulianelliEtAl2026.observedSign Core.GeneralisedSurprisal.PsychMeasure.p600 = -1
Instances For
Whether IAS and surprisal show high complementarity (joint model substantially exceeds either alone) for a given measure and dataset.
Complementarity is highest for predictability measures and self-paced RT in multi-sentence stimuli. For other measures, the joint model ≈ IAS alone, meaning IAS subsumes surprisal's predictive contribution.
Equations
- One or more equations did not get rendered due to their size.
- Phenomena.Processing.Studies.GiulianelliEtAl2026.highComplementarity Core.GeneralisedSurprisal.PsychMeasure.predictabilityRating x✝ = true
- Phenomena.Processing.Studies.GiulianelliEtAl2026.highComplementarity Core.GeneralisedSurprisal.PsychMeasure.clozeProbability x✝ = true
- Phenomena.Processing.Studies.GiulianelliEtAl2026.highComplementarity Core.GeneralisedSurprisal.PsychMeasure.clozeSurprisal x✝ = true
- Phenomena.Processing.Studies.GiulianelliEtAl2026.highComplementarity x✝¹ x✝ = false
Instances For
Under which distance summary statistic surprisal's implicit predictions are best captured.
Under d^{mean}, surprisal correlates most with the final layer at h = 1 (lexical expectations for the next word). Under d^{min}, surprisal correlates most with intermediate layers at h = 3–5 (closest-hypothesis tracking), with correlation coefficients up to 0.81.
The d^{min} finding reveals that surprisal's predictability is closest to a best-case (closest-hypothesis) notion rather than average discrepancy, which may explain its strong predictive power despite lacking explicit temporal-representational resolution.
Equations
Instances For
Explicit predictability measures all peak at horizon 1 (next word).
N400 and P600 both peak at horizon 2, not horizon 1 — prediction of ERP components benefits from two-word lookahead.
N400 and P600 are predicted by different representational levels: N400 by embedding (lexical-semantic), P600 by intermediate (syntactic). This dissociation mirrors the established functional distinction between these components.
Self-paced RT in multi-sentence stimuli benefits from substantially longer forecast horizons than sentence-level stimuli.
IAS outperforms surprisal for both ERP components (Aligned dataset).
IAS outperforms surprisal for all eye-tracking measures (Aligned).
IAS outperforms surprisal for self-paced RT in Natural Stories but not in the sentence-level Aligned dataset.
Explicit measures and implicit measures are predicted by different representational levels: explicit at lexical, implicit at syntactic or shallow-syntactic.
Explicit measures show U-shaped layer patterns while reading times show S-shaped patterns — different measure classes engage different representational geometries.
P600's observed sign is negative, contradicting the predicted positive relationship. This is the only measure where the directional hypothesis is refuted.
All other measures confirm their predicted signs.
For most measures, IAS subsumes surprisal (low complementarity), but for explicit predictability measures, they complement each other.
Surprisal's implicit predictions are best characterized by the minimum distance statistic, not mean distance.