Linear Contracts Are Optimally Robust

August 20, 2025

nb: attempting a daily posting cadence. adjust quality priors accordingly

Consider the following game:

  1. Alice offers a contract w:YR+ to Bob.
  2. Bob, knowing his compact action space over lotteries AΔ(Y)×R+, chooses action (F,c)A,
  3. The output yF is realized (sampling from the lottery chosen)
  4. Alice receives yw(y) payoff; Bob receives w(y)c payoff.

Importantly, Alice receives limited information about Bob's action space (interchangeable with "technology"). What is the optimal contract Alice should give Bob, if she wants to maximize her worst case outcome? (We assume Alice and Bob are rational actors: their dynamics will be given shortly). This is an identical problem to studying the structure of the optimal w in this scenario.

[Car15]1 proves that the optimal w is linear, subject to the following assumptions: YR+, A must be compact; Alice (the "principal") knows A0A possible actions for Bob (the "agent") such that there exists (F,c)A0 such that EF[y]c>0 (the principal should have some reason for hiring the agent); w must be continuous.

Bob's behavior is quite simple, given that Bob has all information. The set of actions (F,c)A Bob will consider are those which maximize EF(y)c.2 We denote this by A(w|A), and we denote by VA(w|A) the expected payoff of Bob given rational behavior. Alice's expected payoff is then

VP(w|A)=max(F,c)A(w|A)EF[yw(y)],

and Alice searches over expected payoffs as

VP(w)=infAA0VP(w|A). We are interested in w such that VP(w) is maximized.

optimality in the zero-shot game

Motivating example: w(y)=αy always guarantees the principal Alice positive worst-case payoff, for α[0,1]. This analysis holds independently of the possible technology A, due to the nontriviality assumption we impose on A0.

Proof: Rewrite yw(y) as w(y)/αw(y)=1ααw(y). Lower bound EF[w(y)] (the expected payment of Bob the agent) with EF[w(y)]EF[w(y)]c=VA(w|A), which is VA(w|A0) (because adding more actions to the agent can't decrease the optimal outcome given A0). Combining the two gives

EF[yw(y)]1ααEF[w(y)]1ααVA(w|A0),

which gives VP(w)1ααVA(w|A0). Given the nontrivality assumption VA(w|A0)>0, this means that this gives Alice a positive lower bound on the worst case outcome independent of the choice of technology A!

Is there a sense in which linear contracts are "the best you can do?" Carroll shows that any contract w(y) can be improved to a linear contract, even if w(y) is pathological. The gist of the argument is as follows: consider the convex hull of the curve w(y) that lies above the value VA(w|A0). Consider the Q=(y,w(y)) point which minimizes EF[y]EF[w(y)]. This is the worst case for Alice, the principal. Q will typically be where VA(w|A0) intersects the left side of the convex hull. Note that the line parametrizing the intersecting boundary of this convex hull is itself a contract w(y) which dominates w(y)! Repeating this process and considering some technical details gives you the optimality result.

cfg1

The full technical details can be found the paper, and I will not discuss them now. However, I would like to discuss the generalization of this lemma to include more observables to the principal.

Let z=(z1,,zk) range over a compact set ZRk. Define a cost function b:RkR+ such that actions depend on b: an action is then (F,c) such that FΔ(Z) and cb(EF(z)). A contract is now a function w:ZR+. Changing the definitions of A0,VA appropriately etc., it turns out the optimally robust contract is linear in the observables available to the principal. Precisely, the optimal contract is of the form

w(z)=α1z1++αkzk+β

for real numbers αi,β.

learnability

It seems like the linear optimality result for robust contracts is pretty general and not too sensitive to assumptions: load-bearing here is that Y has a minimum that we normalize to be zero, some relaxed version of the nontrivality assumption, and that the risk associated with any particular action is quantifiable in a shared manner by both the principal and the agent.3

One obvious consideration: consider the possibility of unbounded risk to the principal (as suggested in UK AISI's Economics and Game Theory research agenda). It is difficult to construct contracts that then robustly protect against these scenarios, even with partial information. What are the minimum viable assumptions necessary to get guarantees in this regime?

Another consideration that I am interested in: this feels very similar to classical bandit problems in RL. Heck, the agent is engaging in the optimal bandit policy given a set of lotteries! Unifying the two literatures (perhaps [KZ25] is of interest) might tell us something interesting about certain classes of decision problems.

Also, learnable agents would (I think) perform better than static ones on a kind of mixed-objective performance metric! Perhaps one which mixes expected average and expected worst-case reward, with some weighting. Generally, intuitive restrictions & extensions on this result are things I'm excited about.

1

The majority of this post is a distillation of this paper. All credit to Gabriel Carroll.

2

A(w|A) is guaranteed to be nonempty by continuity and compactness.

3

This one in particular has some philosophical implications.