Compositional Imprecise Probability
Liell-Cock, Staton ยท POPL 2025 ยท
arXiv:2405.09391
Prereqs: ๐ Fritz 2020 (Markov categories). 5 min.
When you don't know the exact probability โ only that it lies in some set of distributions โ you have imprecise probability. Liell-Cock shows these form a Markov category (nLab) via the Para construction, and that graded monads organize the imprecision levels.
Imprecise probability โ sets of distributions
A precise Markov kernel maps an input to one distribution. An imprecise kernel maps to a set of distributions: a credal set. This is stronger than a single distribution (you commit less) and weaker than full ignorance (you commit something).
The Para construction
The Para construction takes a Markov category C and builds a new category Para(C) where morphisms carry a hidden parameter. A morphism in Para(C) from X to Y is a morphism P ร X โ Y in C, where P is the parameter space. Imprecise kernels arise when you existentially quantify over P: the credal set is the image of the kernel across all parameter values.
Confidence: Simplified. Real Para construction works over arbitrary Markov categories with tensor products for parameters. Same idea: hidden parameter generates a family.
Lower and upper expectations
With a credal set, you compute bounds on expectations. The lower expectation is the minimum expected value over all distributions in the set. The upper expectation is the maximum. These brackets tell you how uncertain you are.
Graded monads โ imprecision levels
Different levels of imprecision form a graded monad. The grade tracks how wide the credal set is: grade 0 = precise (one distribution), higher grades = wider sets. Composition accumulates grades, just like effect grades in ๐ Gaboardi 2021.
The bridge โ graded monads = Markov categories
Liell-Cock's main result: the Kleisli category of a graded monad on a Markov category is itself a Markov category. Imprecise probability inherits all the structure โ copy, discard, composition โ from the precise case. You can do everything Fritz does in FinStoch, but with credal sets instead of single distributions.
Confidence: Simplified. Real composition requires careful treatment of the parameter space tensor. Same compositional structure.
Notation reference
| Paper | Scheme | Meaning |
|---|---|---|
| Para(C) | (make-para params kernel) | Para construction |
| credal set | (make-credal d1 d2 ...) | Set of distributions |
| Eฬฒ[f] | (lower-exp credal f) | Lower expectation |
| Eฬ [f] | (upper-exp credal f) | Upper expectation |
| T_e | # credal set at grade e | Graded monad at imprecision level e |
Neighbors
Other paper pages
- ๐ Fritz 2020 โ the Markov category that Para generalizes
- ๐ Gaboardi 2021 โ graded monads for effects (same grading idea)
- ๐ Fritz, Perrone 2021 โ support is the maximally imprecise shadow
- ๐ Capucci 2021 โ Para construction appears in cybernetics too
Foundations (Wikipedia)
Translation notes
All examples use finite lists of finite distributions as credal sets. The paper works with convex sets of probability measures over measurable spaces, the Para construction on arbitrary Markov categories, and graded monads indexed by a quantale of imprecision levels.
For example: this page's lower/upper expectation iterates over three coin distributions. In the paper, the same construction applies to credal sets of Gaussian measures over โโฟ, where the lower expectation becomes an optimization problem over an infinite-dimensional convex set. The bracket structure (min/max over the set) is identical. The measure theory and optimization are not.
Read the paper. Start at ยง3 for the Para construction, ยง5 for graded monads and Markov categories.
Framework connection: Imprecise probability via the Para construction generalizes the Natural Framework's ambient category from precise to credal-set-valued stages. (
Ambient Category)