person
Geoffrey Irving
Chief Scientist of UK AI Safety Institute; debate-protocol researcher
Now chief scientist at the UK AI Safety Institute after DeepMind, OpenAI, and Google Brain. Advances 'debate' scalable-oversight protocols and has proved properties about when honesty can be incentivised.
current Chief Scientist, UK AI Safety Institute
past Staff Research Scientist, Google DeepMind; Research scientist, OpenAI
Strategy positions
Alignment firstendorses
Solve technical alignment before capability thresholds closeAdvances formal scalable-oversight protocols (debate, prover-estimator debate) that aim to make honest behaviour the equilibrium strategy.
“We present a new scalable oversight protocol (prover-estimator debate) and a proof that honesty is incentivised at equilibrium.”
Closest strategy neighbours
by jaccard overlapOther people whose strategy tags overlap with Geoffrey Irving's. Overlap is on tag identity, not stance; opposites can show up if they reference the same tags.
Record last updated 2026-04-24.