Note: This is a work in progress, being actively developed for a Summer 2025 workshop at the Stanford Social Interaction Lab.

This course introduces a modern computational framework for understanding language use that bridges formal models of pragmatic reasoning with the rich representational capabilities of neural language models. As large language models (LLMs) demonstrate increasingly sophisticated linguistic behaviors, the need for principled, interpretable models of pragmatic reasoning becomes ever more critical. This course explores how neuro-symbolic approaches can provide both the formal rigor needed to understand the computational principles underlying communication and the flexibility to operate over natural language at scale.

Building on the foundations of the Rational Speech Act (RSA) framework, we develop models that integrate:

The course employs memo, a modern probabilistic programming language designed for recursive reasoning about reasoning, which offers significant improvements in both expressiveness and computational efficiency over traditional approaches.

Main content

  1. Introducing the Rational Speech Act framework
    An introduction to language understanding as Bayesian inference

  2. Modeling pragmatic inference
    Enriching the literal interpretations

  3. Inferring the Question-Under-Discussion
    Non-literal language

  4. Combining RSA and compositional semantics
    Jointly inferring parameters and interpretations

  5. Fixing free parameters
    Vagueness

  6. Expanding our ontology
    Plural predication

  7. Extending our models of predication
    Generic language

  8. Reasoning about literal meanings
    Lexical uncertainty

  9. Social reasoning about social reasoning
    Politeness

  10. Summary and outlook
    Questions about RSA

Appendix

  1. Probabilities & Bayes rule (in memo)
    A quick and gentle introduction to probability and Bayes rule (in memo)

  2. More on speaker utility
    Derivation of suprisal-based utilities from KL-divergence

  3. Utterance costs and utterance priors
    More on utterance costs and utterance priors

  4. Bayesian data analysis
    BDA for the RSA reference game model

  5. Quantifier choice & approximate number
    Speaker choice of quantifiers for situations where perception of cardinality is uncertain

  6. Introduction to WebPPL
    A brief introduction.

  7. Glossary
    WebPPL functions used in this book

Citation

[Robert D. Hawkins]. ProbLang v2: A Computational Approach to Pragmatic Reasoning. Retrieved from https://hawkrobe.github.io/probLang-memo.

Based on the original Probabilistic Language Understanding by G. Scontras, M. H. Tessler, and M. Franke.

Useful resources

Acknowledgments

This webbook builds closely on the foundation laid by Scontras, Tessler, and Franke in their Probabilistic Language Understanding course. We are grateful for their pioneering work in making formal pragmatics accessible through probabilistic programming. We also thank Kartik Chandra and other memo developers for creating a tool that makes neuro-symbolic modeling both expressive and efficient.