AGI Strategies

person

Bo Li

UChicago / UIUC; AI safety evaluations

University of Chicago associate professor specializing in safety, robustness, and trustworthiness of ML systems. Lead developer of DecodingTrust, a comprehensive trustworthiness benchmark for LLMs.

current Associate Professor of Computer Science, University of Chicago; Co-founder, Virtue AI

Strategy positions

Evals-drivenendorses

Capability/risk evals gate deployment; evals are the load-bearing artefact

Argues comprehensive multi-dimensional safety benchmarks, covering toxicity, fairness, privacy, robustness, ethics, are needed to characterize AI risks empirically before deployment.

“Despite the impressive capabilities of GPT-4, we identify significant trustworthiness gaps in dimensions including toxicity, stereotype bias, robustness, privacy, and ethics.”
§ paperDecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models· arXiv / NeurIPS· 2023-06· direct quote

Closest strategy neighbours

by jaccard overlap

Other people whose strategy tags overlap with Bo Li's. Overlap is on tag identity, not stance; opposites can show up if they reference the same tags.

  • Aleksander Mądry

    shared 1 · J=1.00

    MIT; ex-OpenAI head of preparedness

  • Alex Meinke

    Alex Meinke

    shared 1 · J=1.00

    Apollo Research; deceptive alignment evaluations

  • Ali Rahimi

    Ali Rahimi

    shared 1 · J=1.00

    Google Brain ML researcher; 'Alchemy' speech

  • Anna Rogers

    Anna Rogers

    shared 1 · J=1.00

    IT University of Copenhagen; LLM benchmarking critique

  • Arati Prabhakar

    Arati Prabhakar

    shared 1 · J=1.00

    White House OSTP director (2022–2025)

  • Beth Barnes

    Beth Barnes

    shared 1 · J=1.00

    Founder of METR; dangerous capability evaluations

Record last updated 2026-04-25.