Linear Contracts Are Optimally Robust
August 20, 2025
nb: attempting a daily posting cadence. adjust quality priors accordingly
Consider the following game:
- Alice offers a contract
to Bob. - Bob, knowing his compact action space over lotteries
chooses action - The output
is realized (sampling from the lottery chosen) - Alice receives
payoff; Bob receives payoff.
Importantly, Alice receives limited information about Bob's action space (interchangeable with "technology"). What is the optimal contract Alice should give Bob, if she wants to maximize her worst case outcome? (We assume Alice and Bob are rational actors: their dynamics will be given shortly). This is an identical problem to studying the structure of the optimal
[Car15]1 proves that the optimal
Bob's behavior is quite simple, given that Bob has all information. The set of actions
and Alice searches over expected payoffs as
optimality in the zero-shot game
Motivating example:
Proof: Rewrite
which gives
Is there a sense in which linear contracts are "the best you can do?" Carroll shows that any contract
The full technical details can be found the paper, and I will not discuss them now. However, I would like to discuss the generalization of this lemma to include more observables to the principal.
Let
for real numbers
learnability
It seems like the linear optimality result for robust contracts is pretty general and not too sensitive to assumptions: load-bearing here is that
One obvious consideration: consider the possibility of unbounded risk to the principal (as suggested in UK AISI's Economics and Game Theory research agenda). It is difficult to construct contracts that then robustly protect against these scenarios, even with partial information. What are the minimum viable assumptions necessary to get guarantees in this regime?
Another consideration that I am interested in: this feels very similar to classical bandit problems in RL. Heck, the agent is engaging in the optimal bandit policy given a set of lotteries! Unifying the two literatures (perhaps [KZ25] is of interest) might tell us something interesting about certain classes of decision problems.
Also, learnable agents would (I think) perform better than static ones on a kind of mixed-objective performance metric! Perhaps one which mixes expected average and expected worst-case reward, with some weighting. Generally, intuitive restrictions & extensions on this result are things I'm excited about.
The majority of this post is a distillation of this paper. All credit to Gabriel Carroll.
This one in particular has some philosophical implications.