person
Asma Ghandeharioun
Google DeepMind; 'Patchscopes' for LLM interpretability
Senior research scientist at Google DeepMind; lead author of Patchscopes, a unifying framework for using language models to inspect their own internal representations.
current Senior Research Scientist, Google DeepMind
Strategy positions
Interpretability betendorses
Mechanistic interpretability is necessary and sufficient to know models are safeArgues language models can be turned into interpretability tools for themselves; reframes mechanistic interpretation as a translation problem between hidden states and natural language.
“Patchscopes leverage the model's own ability to generate text to inspect its hidden representations, unifying many prior interpretability methods.”
Closest strategy neighbours
by jaccard overlapOther people whose strategy tags overlap with Asma Ghandeharioun's. Overlap is on tag identity, not stance; opposites can show up if they reference the same tags.
Record last updated 2026-04-25.