AGI Strategies

person

Stephen Casper

Stephen Casper

MIT PhD researcher; red-teaming and model audit

MIT algorithmic alignment researcher focused on red-teaming, auditing, and interpretability. Has documented how safeguards at current frontier labs are reliably broken by determined red-teamers.

current PhD candidate, Algorithmic Alignment Group, MIT CSAIL

Strategy positions

Evals-drivenendorses

Capability/risk evals gate deployment; evals are the load-bearing artefact

Argues empirical red-teaming reveals that current safeguards are not robust; auditing must become standard infrastructure.

Example after example of state-of-the-art safeguards get pretty reliably broken. That's the empirical reality.
podcastStephen Casper at Center for AI Policy Podcast· Center for AI Policy Podcast· 2024· faithful paraphrase

Closest strategy neighbours

by jaccard overlap

Other people whose strategy tags overlap with Stephen Casper's. Overlap is on tag identity, not stance; opposites can show up if they reference the same tags.

  • Aleksander Mądry

    shared 1 · J=1.00

    MIT; ex-OpenAI head of preparedness

  • Alex Meinke

    Alex Meinke

    shared 1 · J=1.00

    Apollo Research; deceptive alignment evaluations

  • Ali Rahimi

    Ali Rahimi

    shared 1 · J=1.00

    Google Brain ML researcher; 'Alchemy' speech

  • Anna Rogers

    Anna Rogers

    shared 1 · J=1.00

    IT University of Copenhagen; LLM benchmarking critique

  • Arati Prabhakar

    Arati Prabhakar

    shared 1 · J=1.00

    White House OSTP director (2022–2025)

  • Beth Barnes

    Beth Barnes

    shared 1 · J=1.00

    Founder of METR; dangerous capability evaluations

Record last updated 2026-04-25.