person

Cynthia Rudin
Duke professor; interpretable ML pioneer
Computer scientist who has been the most consistent public voice against black-box ML in high-stakes domains. Argues interpretable models should always be preferred to post-hoc explanations of black boxes.
Profile
expertise
Deep technical
Sustained peer-reviewed contribution to ML, alignment, interpretability, or safety techniques. Could review a frontier paper.
Duke professor. Pioneered interpretable ML (as opposed to post-hoc explanation). Won the AAAI Squirrel AI Award 2022 for AI for the Benefit of Humanity.
recognition
Field-leading
Widely known inside the AI and AI-safety community. Appears repeatedly in top venues, podcasts, or governance forums. Not a household name to outsiders.
Recognised in interpretable-ML community; less mainstream press than DeepMind/Anthropic interpretability leads.
vintage
Deep-learning rise
Came up post-AlexNet. ImageNet, AlphaGo, transformer paper. DeepMind, Google Brain, FAIR establish the modern lab template.
PhD 2004 (Princeton). Interpretable-ML programme matures in 2010s as a counter to deep-learning opacity.
Hand-classified. See the board for the criteria and the full grid.
Strategy positions
Interpretability betmixed
Mechanistic interpretability is necessary and sufficient to know models are safeArgues for inherently interpretable models over post-hoc explanations, a different flavour of interpretability than the mechanistic-interpretability school.
“Stop explaining black box machine learning models for high-stakes decisions and use interpretable models instead.”
Closest strategy neighbours
by jaccard overlapOther people whose strategy tags overlap with Cynthia Rudin's. Overlap is on tag identity, not stance; opposites can show up if they reference the same tags.
Record last updated 2026-04-24.