Note: This is a work in progress, being actively developed for a Summer 2025 workshop at the Stanford Social Interaction Lab.
This course introduces a modern computational framework for understanding language use that bridges formal models of pragmatic reasoning with the rich representational capabilities of neural language models. As large language models (LLMs) demonstrate increasingly sophisticated linguistic behaviors, the need for principled, interpretable models of pragmatic reasoning becomes ever more critical. This course explores how neuro-symbolic approaches can provide both the formal rigor needed to understand the computational principles underlying communication and the flexibility to operate over natural language at scale.
Building on the foundations of the Rational Speech Act (RSA) framework, we develop models that integrate:
- Symbolic reasoning about speaker and listener intentions through recursive Bayesian inference
- Neural representations that capture real-world priors, contextual nuances, and the full complexity of natural language
- Probabilistic programming techniques that make these hybrid models both expressible and computationally tractable
The course employs memo
, a modern probabilistic programming language designed for recursive reasoning about reasoning, which offers significant improvements in both expressiveness and computational efficiency over traditional approaches.
Main content
-
Introducing the Rational Speech Act framework
An introduction to language understanding as Bayesian inference -
Modeling pragmatic inference
Enriching the literal interpretations -
Inferring the Question-Under-Discussion
Non-literal language -
Combining RSA and compositional semantics
Jointly inferring parameters and interpretations -
Fixing free parameters
Vagueness -
Expanding our ontology
Plural predication -
Extending our models of predication
Generic language -
Reasoning about literal meanings
Lexical uncertainty -
Social reasoning about social reasoning
Politeness -
Summary and outlook
Questions about RSA
Appendix
-
Probabilities & Bayes rule (in memo)
A quick and gentle introduction to probability and Bayes rule (in memo) -
More on speaker utility
Derivation of suprisal-based utilities from KL-divergence -
Utterance costs and utterance priors
More on utterance costs and utterance priors -
Bayesian data analysis
BDA for the RSA reference game model -
Quantifier choice & approximate number
Speaker choice of quantifiers for situations where perception of cardinality is uncertain -
Introduction to WebPPL
A brief introduction. -
Glossary
WebPPL functions used in this book
Citation
[Robert D. Hawkins]. ProbLang v2: A Computational Approach to Pragmatic Reasoning. Retrieved from https://hawkrobe.github.io/probLang-memo.
Based on the original Probabilistic Language Understanding by G. Scontras, M. H. Tessler, and M. Franke.
Useful resources
- Probabilistic Models of Cognition: An introduction to computational cognitive science
- The ProbLang book: The WebPPL-based predecessor to this course
- Pragmatic language interpretation as probabilistic inference: A review of the RSA framework
- memo documentation: The probabilistic programming language used throughout this book
Acknowledgments
This webbook builds closely on the foundation laid by Scontras, Tessler, and Franke in their Probabilistic Language Understanding course. We are grateful for their pioneering work in making formal pragmatics accessible through probabilistic programming. We also thank Kartik Chandra and other memo developers for creating a tool that makes neuro-symbolic modeling both expressive and efficient.