NPIs in Questions: Entropy as Strength #
@cite{van-rooy-2003} @cite{krifka-1995a} @cite{kadmon-landman-1993} @cite{shannon-1948}
@cite{van-rooy-2003}: for assertions, strength = informativity; for questions, strength = entropy. NPIs are licensed when they increase entropy by reducing bias.
Architecture #
All entropy definitions use ℝ-valued Real.negMulLog from Mathlib, which gives
negMulLog(x) = -x * log(x) with negMulLog(0) = 0. Shannon entropy of a
question H(Q) = ∑_cell negMulLog(P(cell)) where cell probabilities are computed
via Fintype-based sums.
- @cite{van-rooy-2003}. Negative Polarity Items in Questions: Strength as Relevance.
- @cite{shannon-1948}. Mathematical Theory of Communication.
- @cite{krifka-1995a}. The semantics and pragmatics of polarity items.
- @cite{kadmon-landman-1993}. Any.
Shannon entropy of a question: H(Q) = ∑_cell negMulLog(P(cell)).
Each cell contributes negMulLog(P(cell)) = -P(cell) · log(P(cell)) to the total.
Degenerate cells (P = 0 or P = 1) contribute 0 by negMulLog boundary behavior.
Equations
- Semantics.Questions.EntropyNPIs.questionEntropy prior q = List.foldl (fun (acc : ℝ) (cell : W → Bool) => acc + (Semantics.Questions.EntropyNPIs.cellProb prior cell).negMulLog) 0 q
Instances For
Binary entropy function: H(p) = negMulLog(p) + negMulLog(1 - p).
Instances For
A question has zero entropy iff it is already settled (one cell has probability 1).
Equations
- Semantics.Questions.EntropyNPIs.isSettled prior q = ∃ c ∈ q, Semantics.Questions.EntropyNPIs.cellProb prior c = 1
Instances For
Binary entropy is maximized at p = 1/2.
Proof: By strict concavity of negMulLog on [0, 1]. For any p ∈ [0, 1],
negMulLog(p) + negMulLog(1-p) ≤ 2 · negMulLog(1/2) by Jensen's inequality,
with equality iff p = 1/2.
Binary entropy is maximal at equiprobable: for any binary partition,
H([½, ½]) ≥ H([p, 1−p]).
Proof: H([pos, neg]) = binaryEntropy(P(pos)) ≤ binaryEntropy(½) by concavity.
Binary entropy increases when probability moves closer to ½.
If p ≤ q ≤ 1 - p (q is between p and its "mirror" 1-p), then
binaryEntropy(q) ≥ binaryEntropy(p). This is the mathematical core of
van Rooy's argument: reducing bias increases entropy.
Proof: write q = (1−t)p + t(1−p) where t = (q−p)/(1−2p).
By concavity of negMulLog, applied to both q and 1−q:
negMulLog(q) ≥ (1−t)·negMulLog(p) + t·negMulLog(1−p)negMulLog(1−q) ≥ (1−t)·negMulLog(1−p) + t·negMulLog(p)Summing givesbinaryEntropy(q) ≥ binaryEntropy(p).
A polar question is biased toward negative if P(neg) > P(pos).
Equations
- Semantics.Questions.EntropyNPIs.isBiasedNegative prior positive negative = (Semantics.Questions.EntropyNPIs.cellProb prior negative > Semantics.Questions.EntropyNPIs.cellProb prior positive)
Instances For
Degree of bias: |P(pos) - P(neg)|.
Equations
- Semantics.Questions.EntropyNPIs.biasDegree prior positive negative = |Semantics.Questions.EntropyNPIs.cellProb prior positive - Semantics.Questions.EntropyNPIs.cellProb prior negative|
Instances For
NPI effect on a polar question: widens the positive answer's domain
- posWithoutNPI : W → Bool
Positive answer without NPI (e.g., "someone called")
- posWithNPI : W → Bool
Positive answer with NPI (e.g., "anyone called")
- widens (w : W) : self.posWithoutNPI w = true → self.posWithNPI w = true
NPI widens domain: withNPI ⊇ withoutNPI
Instances For
Negative answer is complement of positive
Equations
- e.negWithoutNPI w = !e.posWithoutNPI w
Instances For
Equations
- e.negWithNPI w = !e.posWithNPI w
Instances For
Question without NPI
Equations
Instances For
Question with NPI
Equations
Instances For
NPI increases entropy when question is negatively biased.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Van Rooy's Key Result: When negatively biased, domain widening that reduces bias increases entropy.
The key hypotheses are:
- Partition: both questions' cells sum to 1 (proper probability distributions)
- Negative bias: P(pos) < ½ (negative answer is more likely)
- Widening increases positive prob: P(pos') ≥ P(pos)
- Bounded widening: P(pos') ≤ 1 - P(pos) (widening doesn't overshoot past the "mirror" of the original bias)
Under these conditions, |P(pos') - ½| ≤ |P(pos) - ½|, so by concavity of
negMulLog, binaryEntropy(P(pos')) ≥ binaryEntropy(P(pos)).
The proof applies binaryEntropy_mono_of_closer_to_half: since P(pos) < ½ and
P(pos) ≤ P(pos') ≤ 1 - P(pos), the widened question has higher binary entropy.
Converse: In positively biased questions, widening the positive cell moves further from balance, DECREASING entropy. NPIs are not licensed.
Proof: By symmetry of binaryEntropy, this reduces to the negatively biased
case on the complementary probabilities. Since P(neg) < ½ and P(neg') ≤ P(neg),
the neg probabilities move AWAY from ½, decreasing entropy.
Entropy Properties #
Non-negativity and reduction to binary entropy.
Note: @cite{van-rooy-2003} shows that entropy equals expected utility
for the log-scoring decision problem, grounding the entropy measure in decision
theory. A formal bridge to the ℚ-valued questionUtility in Core.DecisionTheory
requires ℝ-valued decision theory infrastructure.
Question entropy is non-negative when cell probabilities are in [0, 1].
Each cell contributes negMulLog(P(c)) ≥ 0 by negMulLog_nonneg, and the
foldl sum of non-negative terms is non-negative.
For binary partitions, question entropy reduces to the binary entropy function.
H([pos, neg]) = binaryEntropy(P(pos)) when P(pos) + P(neg) = 1.
Rhetorical Questions and Even-NPIs #
Strong NPIs (lift a finger, bat an eye) create rhetorical readings because:
- They denote MINIMAL scalar values
- They share a presupposition with EVEN: "For all non-minimal alternatives, the question is already settled"
Van Rooy's analysis:
- If alternatives are settled, only the minimal value is questionable
- But if minimal value is least likely to be true, only negative answer reasonable
- This creates rhetorical force
A strong NPI has EVEN-like presupposition.
- posWithoutNPI : W → Bool
- posWithNPI : W → Bool
- isMinimal : True
The NPI denotes a minimal scalar value
Set of non-minimal alternatives
- alternativesSettled (alt : W → Bool) : alt ∈ self.alternatives → ∀ (w : W), alt w = true ∨ ¬alt w = true
Presupposition: all alternatives are settled (known false or true)
Instances For
A question is rhetorical if it has near-zero entropy.
Equations
- Semantics.Questions.EntropyNPIs.isRhetorical prior q threshold = (Semantics.Questions.EntropyNPIs.questionEntropy prior q < threshold)
Instances For
Theorem: Strong NPIs create rhetorical readings.
When a strong NPI is used in a question and the positive cell has probability
P(pos) ≤ ε ≤ ½, the question entropy is bounded by binaryEntropy(ε).
Since binaryEntropy(0) = 0 and binaryEntropy is continuous and increasing
on [0, ½], small ε gives small entropy — a rhetorical reading.
The proof reduces to binaryEntropy being monotone on [0, ½]:
questionEntropy [pos, neg] = binaryEntropy(P(pos)) ≤ binaryEntropy(ε).
The proof applies binaryEntropy_mono_of_closer_to_half when P(pos) < ½,
and handles the degenerate case P(pos) = ½ (forcing ε = ½) separately.
K&L's Question Strengthening #
@cite{kadmon-landman-1990} propose that stressed "any" strengthens questions:
Q' strengthens Q iff Q is already answered but Q' is still unanswered.
Van Rooy: This is a special case of entropy increase.
- If Q is settled (entropy 0) but Q' is not (entropy > 0)
- Then Q' has higher entropy
- Domain widening achieves this when Q is biased toward negative
K&L's notion: Q' strengthens Q if Q is settled but Q' is open.
Equations
- Semantics.Questions.EntropyNPIs.klStrengthens prior q q' = (Semantics.Questions.EntropyNPIs.isSettled prior q ∧ ¬Semantics.Questions.EntropyNPIs.isSettled prior q')
Instances For
K&L strengthening implies higher entropy.
When Q is settled (all cells have P ∈ {0,1}), its entropy is 0. When Q' is not settled, it has some cell with 0 < P < 1, giving entropy > 0.
Stressed "any" achieves K&L strengthening via domain widening.
If the question without NPI is settled, and widening adds a new world to the positive cell, then the question with NPI is not settled (the new world moves probability mass, breaking the P = 1 condition).
Strength as Relevance: The Unification #
Van Rooy's key contribution: a UNIFIED notion of strength.
| Speech Act | Strength Measure | NPI Effect |
|---|---|---|
| Assertion | Informativity (entailment) | Wider domain → stronger in DE |
| Question | Entropy (avg informativity) | Wider domain → less biased |
Both are instances of utility maximization:
- Assertions: direct informativity maximization
- Questions: expected informativity (entropy) maximization
The connection to decision theory (Section 5.3) shows this is rational: agents should choose utterances that maximize expected utility.
The unified strength measure.
Instances For
Strength for assertions: surprisal = −log P(φ).
@cite{van-rooy-2003}: informativity of an assertion is its surprisal, -log P(φ).
In DE contexts, the assertion is negated, so strength should be computed on the
negated form. Surprisal is strictly decreasing for P > 0, so narrower
propositions (lower probability) have higher informativity.
Equations
Instances For
Strength for questions: entropy.
Equations
Instances For
NPI licensing condition: increases strength under appropriate polarity/bias.
For assertions in DE contexts (polarity = false), strength is computed on the NEGATED form: wider domain under negation = narrower negation = more informative. For questions with negative bias (polarity = false), wider domain increases entropy.
Equations
- One or more equations did not get rendered due to their size.
Instances For
NPI assertion licensing in DE contexts: wider domain under negation narrows the negated proposition, making it more informative (higher surprisal).
Proof: Since cellProb negWithNPI ≤ cellProb negWithoutNPI and -log is
order-reversing on (0, ∞), the narrower negation has higher surprisal.
The Grand Unification: NPI licensing follows the same principle for assertions and questions—maximize strength under current polarity/bias.
For assertions: DE context → wider domain under negation is MORE informative → NPI licensed. Wider positive domain ⟹ narrower negation ⟹ lower P(neg) ⟹ higher surprisal. For questions: Negative bias → wider domain increases entropy → NPI licensed.
The hypotheses suffice for both cases:
- Assertion:
hNegPosandhWider+ partition givecellProb neg↓ ≤ cellProb neg, thennpi_assertion_licensed_deapplies. - Question:
npi_increases_entropy_when_negatively_biasedapplies directly.
Wh-Questions (Section 3.1) #
Van Rooy notes that wh-questions ARE downward entailing in subject position:
"Who of John, Mary and Sue are sick?" entails "Who of John and Mary are sick?"
(Every complete answer to the wider question entails an answer to the narrower.)
But this DOESN'T explain NPI licensing in predicate position or polar questions. That's why we need the entropy-based account.
Wh-question entailment: Q entails Q' if every complete answer to Q determines a complete answer to Q'.
When Q' has a subset domain of Q (Q'.domain ⊆ Q.domain), knowing the predicate's extension over Q.domain determines its extension over Q'.domain. This is the sense in which wider wh-questions are stronger.
Equations
Instances For
Domain widening in wh-subject position is DE when predicates agree.
When Q and Q' share the same predicate and Q'.domain ⊇ Q.domain,
knowing the predicate for all of Q'.domain determines it for Q.domain.
The hSamePred hypothesis ensures the predicate is the same function.
NPIs licensed in wh-subject position via standard DE reasoning.
The NPI widens the domain (Q'.domain ⊇ Q.domain). In subject position this is downward-entailing: the wider question entails the narrower one, because every complete answer to the wider question determines an answer to the narrower question. Requires predicates to agree.