AGI Strategies

person

Martin Rees

Martin Rees

Astronomer Royal; CSER co-founder

Former Astronomer Royal who co-founded the Centre for the Study of Existential Risk at Cambridge with Huw Price and Jaan Tallinn. Frames AI alongside bioengineering as the most serious civilisational-scale risks this century.

current Emeritus Professor, University of Cambridge; Co-founder, Centre for the Study of Existential Risk

Profile

expertise

External-domain expert

Recognised expert outside AI (philosophy, economics, biology, journalism) who weighs in on AI consequences from that vantage.

UK Astronomer Royal. Co-founded Cambridge Centre for the Study of Existential Risk. Cosmologist; engages AI through x-risk lens.

recognition

Household name

Name recognition outside the AI/CS community. Featured by mainstream press, a Wikipedia page in many languages, a published bestseller, or holds a position the lay public knows.

Member of the House of Lords; mainstream UK press regular.

vintage

Pre-deep-learning

Active before AlexNet. The existential-risk frame matures (FHI, OpenPhil, EA). Public AI commentary still rare; deep learning not yet dominant.

CSER co-founded ~2012. Astronomer Royal long before. Pre-deep-learning x-risk frame.

Hand-classified. See the board for the criteria and the full grid.

Strategy positions

Existential primacyendorses

Extinction/disempowerment risk overrides ordinary cost-benefit

Argues AI is one of a small set of 21st-century technologies with genuine civilisational-scale downside risk.

“Since we can't understand what's going on inside them, we have to be cautious about handing over power to them.”
articleBuckle up: We are in for a bumpy ride. An interview with Royal Astronomer Martin Rees· Bulletin of the Atomic Scientists· 2020-12· direct quote

Closest strategy neighbours

by jaccard overlap

Other people whose strategy tags overlap with Martin Rees's. Overlap is on tag identity, not stance; opposites can show up if they reference the same tags.

  • Alan Robock

    Alan Robock

    shared 1 · J=1.00

    Rutgers climate scientist; nuclear winter researcher

  • Andy Jones

    shared 1 · J=1.00

    Anthropic researcher; scaling inference laws

  • Avital Balwit

    shared 1 · J=1.00

    Anthropic communications lead; public-facing AI safety voice

  • Bill McKibben

    Bill McKibben

    shared 1 · J=1.00

    Environmental writer; Middlebury scholar

  • Cade Metz

    shared 1 · J=1.00

    NYT AI reporter; Genius Makers author

  • Clay Graubard

    shared 1 · J=1.00

    Forecaster; RAND and Good Judgment contributor

Record last updated 2026-04-24.