Cooperation substrate ↑ · ai artefact
Cooperative AI
The binding constraint is equilibrium dynamics of many AI systems, not individual alignment; commitment and verification tech make cooperative equilibria reachable.
Mechanism
Design AI systems, protocols, and institutions for stable cooperative equilibria between AI systems.
If it succeeds: what binds next
AI agents cooperate. The binding problem becomes exclusion from the cooperating group, which humans, which AIs, which goals are in.
A strategy that produces a worse next problem than the one it solved has not done durable work.
Falsification signal
AI systems defect in deployments where commitment technology exists and cooperation was available.
A strategy held without a falsification signal is not strategy; it is affiliation. Continued support after this signal lands is identity, not bet. See the identity diagnostic.
Self-undermining threshold
overshoot riskBefore commitment and verification technology is mature enough for stable cooperation.
Unstable multi-agent regimes inherit the failure modes of imperfect cooperation without the benefits of robust cooperation.
Every strategy has a stable region where it reinforces itself and an unstable region where pursuit defeats it. The threshold between them is usually narrower than advocates acknowledge.
People on the record
6Profiled figures appear first, with their tier in small caps. Each face links to the person and their full quote record. Tag: cooperative-ai.
Andrew Critch
Deep ML / safety technical · Recognised inside subfield
Vincent Conitzer
Deep ML / safety technical · Recognised inside subfield
Igor Mordatch
Google DeepMind; multi-agent and embodied AI

Jakob Foerster
Oxford FLAIR lab; multi-agent RL
Kate Sills
AI economic systems and multi-agent markets researcher
Lewis Hammond
Cooperative AI Foundation co-director
Load-bearing commitments
Worldview positions this strategy quietly assumes. If the claim fails empirically or philosophically, the strategy loses its target or its premise.
AI has sufficient agency for reciprocal arrangement.
Fails if: If AI agency is instrumental, cooperation is a design choice humans make, not a two-sided arrangement.
Coordinates
Conflicts, grouped by mechanism
0No strict conflicts catalogued. This strategy pulls a lever that nothing else pulls in the opposite direction.
Complements, grouped by mechanism
5Same-lever reinforce
same lever, same pull, different mechanismBoth strategies pull the same lever in the same direction by different means. They stack: doing both amplifies the pull, at the cost of double-counting in portfolio audits.
Same phase, different layer
same stage, distinct leversBoth are active in the same phase of the transition but act on different layers (model vs institution vs culture). They cover different failure modes inside the same window.
Axis position
Source note: Cooperative AI strategy.md