person
Lilian Weng
Thinking Machines; former OpenAI VP of Research
Former VP of Research at OpenAI who helped lead safety research there. Wrote widely-read technical blog posts on RLHF and alignment. Joined Mira Murati's Thinking Machines Lab in 2024.
current Researcher, Thinking Machines Lab
past VP of Research, Safety Systems, OpenAI
Strategy positions
Alignment firstendorses
Solve technical alignment before capability thresholds closePublic educator on alignment technical work. Ran OpenAI's Safety Systems team before leaving for Thinking Machines.
Reward modelling, RLHF, red-teaming, and eval frameworks are the concrete instruments of alignment today.
Closest strategy neighbours
by jaccard overlapOther people whose strategy tags overlap with Lilian Weng's. Overlap is on tag identity, not stance; opposites can show up if they reference the same tags.
Record last updated 2026-04-25.