AGI Strategies
← overview

Control mechanism · ai artefact

Safe by construction AI

Safety is a property that can be mathematically specified and mechanically verified for the class of systems being built.

Mechanism

Formally verify safety properties of AI systems, build provably aligned optimisation targets, and mathematically guarantee corrigibility.

Coordinates

Acts onai artefact
Coercionconsent
Actor in controlhumans
Time horizonpre transition
Legitimacy sourcetechnical

Conflicts, grouped by mechanism

0

No strict conflicts catalogued. This strategy pulls a lever that nothing else pulls in the opposite direction.

Complements, grouped by mechanism

4

Same-lever reinforce

same lever, same pull, different mechanism

Both strategies pull the same lever in the same direction by different means. They stack: doing both amplifies the pull, at the cost of double-counting in portfolio audits.

Alignment firstInterpretability first

Same phase, different layer

same stage, distinct levers

Both are active in the same phase of the transition but act on different layers (model vs institution vs culture). They cover different failure modes inside the same window.

Narrow AI preservation

Same-side diversification

same side, different lever

Both act on the same side (AI or world) but pull distinct levers. They cover several failure modes on that side while leaving the other side uncovered.

Abandon superintelligence

Same-lever twins

3

Both use the same lever in the same direction. Usually redundant inside a portfolio: each dollar or effort unit only buys one lever pull, even if two strategies are named.

AI containmenttwinAI for safetytwinCounter AI AItwin

Axis position

What the strategy acts onAI artefact
Coercion levelConsent
Actor in controlHumans as principals
Time horizonPre-transition
Legitimacy sourceTechnical

Source note: Safe by construction AI strategy.md