AGI Strategies

strategy tag

Open source.

Release weights widely; transparency beats closed safety

also known as: open weights

stated endorsers

37

no opposers yet

profiled endorsers

8

248 on the board total

endorser mean p(doom)

25%

n=2 · median 25%

quotes by endorsers

37

just for this tag

principal voices

Highest-recognition profiled endorsers, broken ties by quote count. Inclusion is not endorsement of the position, it's recognition of who the discourse turns to when the bet is debated.

  • Yann LeCunYann LeCun

    Household name

  • Andrew NgAndrew Ng

    Household name

  • Mark ZuckerbergMark Zuckerberg

    Household name

  • Tim Berners-LeeTim Berners-Lee

    Household name

  • Emad MostaqueEmad Mostaque

    Field-leading

where the endorsers sit on the board

8 of 248 profiled · 3% of the board

expertise ↓ · recognition →Household nameField-leadingEstablishedEmerging
Frontier builder
  • Yann LeCun
  • Tim Berners-Lee
··
Deep technical
  • Andrew Ng
  • Jeremy Howard
  • Joëlle Pineau
··
Applied technical····
Policy / meta
  • Mark Zuckerberg
···
External-domain expert····
Commentator·
  • Emad Mostaque
··

Each face is one profiled person. Cell shade intensifies with endorser density. Faces with × are profiled opposers, same tier, opposite position. Empty cells mark tier combinations the field has not produced for this bet.

also held by these endorsers

What other strategies the same people endorse. Behavioural signal of compatibility, not a declared rule. A high share means the two positions are routinely held together.

Compare this list to the declared relations matrix. Where they differ, the data reveals a pairing the framework doesn't name yet, the global co-endorsement view ranks all pairs.

Tier mix counts only endorsers (endorses, mixed, conditional, evolved-toward).

expertise mix of endorsers · 8 profiled of 37

Builds frontier systems
3
Deep ML / safety technical
3
Applied or adjacent technical
0
Governance, policy, strategy
1
Expert in another field
0
Public-square commentator
1

recognition mix of endorsers

Mass-public recognition
4
Known across the AI/safety field
4
Recognised inside subfield
0
Newer or less central voice
0

vintage mix · n=8 of 8 profiled with era assigned

Pioneer
0
Symbolic era
2
Pre-deep-learning
0
Deep-learning rise
4
Scaling era
2
Post-ChatGPT
0

Vintage is the era when this person's AI worldview formed, pioneer through post-ChatGPT. A bet held mostly by post-ChatGPT entrants is in a different epistemic state from one held by pre-deep-learning veterans.

People on the record.

37

Ada Rose Cannon

W3C web standards advocate; AR/VR engineer

endorses

Argues open standards are the structural antidote to AI-driven platform consolidation.

If the open web doesn't have an answer to AI, AI will swallow the open web.
blogAda Rose Cannon homepage· ada.is· 2024· loose paraphrase

Ali Farhadi

Allen Institute for AI CEO

endorses

Argues fully open frontier models, including data and training recipes, are necessary for the field's scientific integrity and for democratizing access to advanced AI.

We released OLMo with everything: weights, training data, training code, evaluation suites. Open-source AI cannot be a marketing word, it has to be all the way down.
articleOLMo: Accelerating the Science of Language Models· Allen Institute for AI· 2024-02· faithful paraphrase
Ali Ghodsi

Ali Ghodsi

Databricks co-founder and CEO

endorses

Argues every enterprise will need to build custom AI on its own data; closed APIs are a temporary architecture and the durable form is open-source models trained against private data.

The future of enterprise AI is custom models trained on the company's own data. That requires open foundation models you can adapt, not closed APIs you have to depend on.
articleAli Ghodsi on Databricks AI· Databricks· 2024· faithful paraphrase
Andrew Ng

Andrew Ng

Coursera co-founder; former Baidu Chief Scientist

endorses

Supports open-source models and has warned against regulatory capture that would lock them out.

Bad actors can use open source models; good actors can use them too. The real risk is if regulation effectively bans open source.
blogAndrew Ng on open source AI· The Batch· 2023· loose paraphrase

Anjney Midha

Andreessen Horowitz general partner; AI investor

endorses

Argues open-weight foundation models are the principal path to keeping frontier AI from being captured by a small number of incumbents; supports portfolio companies building on open weights.

The open-source AI ecosystem is the only credible counterweight to closed-frontier capture. Backing it is an act of structural diversification of the AI economy.
articleAnjney Midha, Andreessen Horowitz· a16z· 2024· faithful paraphrase
Arthur Mensch

Arthur Mensch

CEO of Mistral AI; French frontier-model founder

endorses

Argues open-weight frontier models are both a safety benefit and a sovereignty imperative for Europe.

Open models are the only way to ensure that AI development remains a shared enterprise, not an oligopoly.
articleMistral AI· Mistral AI· 2024· loose paraphrase

Ce Zhang

ETH Zürich → University of Chicago; ML systems

endorses

Argues distributed and open ML infrastructure is necessary to keep frontier capabilities accessible to academic research and small organizations; co-founded Together AI on this thesis.

Open and decentralized infrastructure for foundation models is the only way to make sure academic research and small companies can keep contributing to the frontier.
articleTogether AI, about· Together AI· 2023· faithful paraphrase
Clément Delangue

Clément Delangue

CEO of Hugging Face; open-source AI advocate

endorses

Argues open-source AI gives civil society, academia, and governments a counterbalance to frontier-lab concentration.

“Open science and open source create a safer path for development of the technology by giving civil society, nonprofits, academia and policymakers the capabilities they need to counterbalance the power of big private companies.”

Context: Testimony to the US House Science, Space, and Technology Committee.

testimonyHugging Face CEO tells US House open-source AI is 'extremely aligned' with American interests· US House Science, Space, and Technology Committee· 2023-06-22· direct quote

Colin Raffel

UofT; Hugging Face; T5 author

endorses

Argues open-source ML research, datasets, weights, training code, is the principle that lets science around foundation models actually accumulate; closed releases break that mechanism.

“We propose a unified framework that converts every text-based language problem into a text-to-text format. The simple objective lets us study transfer learning systematically.”
§ paperExploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer· arXiv / Google Research· 2019-10· direct quote
Emad Mostaque

Emad Mostaque

Former CEO of Stability AI; open-source frontier advocate

endorses

Built Stability AI on open-weight principles.

Open models are a public good.
tweetEmad Mostaque on open-source AI· X/Twitter· 2023· loose paraphrase

Illia Polosukhin

NEAR Protocol co-founder; Transformer co-author

endorses

Argues open and decentralized infrastructure is the natural counterweight to corporate AI concentration; founded NEAR partly on this thesis.

Centralized AI poses risks no single regulator can solve. We need open infrastructure that lets users own their data and models.
articleIllia Polosukhin on open AI· NEAR Foundation· 2024· faithful paraphrase
Jeremy Howard

Jeremy Howard

Co-founder of fast.ai; former Kaggle president

endorses

Argues open-weight AI is necessary for safety, research, and access.

We need open research and open models. The closed-lab approach is a danger to civil society.
articlefast.ai homepage· fast.ai· 2023· loose paraphrase
Joëlle Pineau

Joëlle Pineau

Cohere Chief AI Officer; former Meta VP of AI Research

endorses

Long-time public face of Meta's open-source AI strategy. Argues open research accelerates the field.

“Science advances more quickly when we share our knowledge. When we put all of our knowledge in common, we push that boundary faster.”
podcastBuilding Connections Through Open Research· MIT Sloan Management Review· 2024· direct quote

Lewis Tunstall

Hugging Face; LLM post-training

endorses

Argues open post-training tooling is a precondition for the open-source AI ecosystem to keep pace with frontier labs; built TRL and DPO/PPO scaling primitives toward this end.

If RLHF is opaque and proprietary, the open-source ecosystem is locked out of the most important phase of model development. Building open post-training tooling is the answer.
articleLewis Tunstall, Hugging Face· Hugging Face· 2024· faithful paraphrase

Liang Wenfeng

Founder of DeepSeek; Chinese frontier AI

endorses

Released DeepSeek R1 and V3 as open-weight models with cost-efficient training. Resets cost expectations for Chinese frontier AI.

Open-weight models are the path to China's leadership in AI.
articleDeepSeek· DeepSeek· 2025-01· loose paraphrase

Luis Ceze

OctoML CEO; UW computer architecture

endorses

Argues open-source compiler-and-runtime ecosystems (TVM, MLIR) are the technical foundation for diversified AI hardware.

Open compilers like TVM are the only path to a diverse AI hardware ecosystem.
articleOctoML· OctoML· 2024· loose paraphrase
Mark Zuckerberg

Mark Zuckerberg

CEO of Meta; open-weight frontier AI proponent

endorses

Argues open-weight models are essential and that the long-term risks of closed concentration exceed the near-term risks of open weights.

I think most developers want to use open-source AI. I think most of the AI ecosystem is going to be built on open-source models.
blogOpen Source AI Is the Path Forward· About Meta· 2024-07-23· faithful paraphrase

Martin Casado

Andreessen Horowitz general partner; infrastructure investor

endorses

Argues open-source AI is essential for both market structure and safety; framed regulatory capture by frontier labs as the principal AI risk to oppose.

Regulating AI right now means regulating math. Open source is the antidote to the regulatory capture being attempted by the largest labs.
articleMartin Casado on AI regulation· a16z· 2023· faithful paraphrase
Matei Zaharia

Matei Zaharia

Databricks CTO and co-founder; Apache Spark creator

endorses

Argues open and customizable foundation models, deployed on enterprise data, are how AI actually creates value at scale; closed APIs alone are insufficient.

The biggest value of generative AI for most businesses comes from training and customizing models on their own data, in their own infrastructure. That requires open foundation models.
articleDatabricks AI Strategy· Databricks· 2023· faithful paraphrase

Mike Lewis

Meta FAIR; BART, Llama 2 lead

endorses

Co-architect of Meta's open-weight Llama family; argues open foundation models accelerate research and broaden the safety discussion beyond the frontier labs.

“BART pretrains by corrupting text and learning a model to reconstruct the original text. The denoising objective generalises across many language tasks.”
§ paperBART: Denoising Sequence-to-Sequence Pre-training· arXiv / Facebook AI· 2019-10· direct quote

Nathan Lambert

Allen Institute for AI; 'Interconnects' newsletter

endorses

Argues open post-training is a prerequisite to credible AI safety claims; the closed nature of frontier RLHF makes external evaluation impossible.

We can't talk seriously about AI safety while the post-training of frontier models is opaque to outsiders. Open-source post-training is the precondition for everything else.
blogInterconnects by Nathan Lambert· Substack· 2024· faithful paraphrase
Nick Clegg

Nick Clegg

Former Meta President of Global Affairs (2018–2025)

endorses

Argued during his Meta tenure that open-weight models are the right answer to concerns about AI concentration; positioned Llama's release as a structural counterweight to closed labs.

Open-source AI is the path to keeping innovation accessible and trustworthy. The alternative, a small number of opaque models controlling the future of the technology, is the worse outcome for both users and societies.
articleNick Clegg on open-source AI· Meta Newsroom· 2024· faithful paraphrase
Nigel Shadbolt

Nigel Shadbolt

Oxford / Open Data Institute co-founder

endorses

Argues open data and open infrastructure are the prerequisites for AI that benefits society broadly rather than entrenching incumbent platforms.

Open data is the raw material of an AI future that is more equitable. Without it, AI development concentrates in whoever already controls the largest private datasets.
bookThe Digital Ape: how to live (in peace) with smart machines· Scribe· 2018· faithful paraphrase

Omar Khattab

Stanford / Databricks; DSPy creator

endorses

Argues programmatic LLM pipelines, not prompt engineering, are how production AI systems should be built; this requires open infrastructure to be fully practical.

DSPy lets you treat LLM calls as differentiable building blocks, optimized end-to-end against a metric. The shift is from prompt-tuning by hand to compiling LLM programs.
§ paperDSPy: Compiling Declarative Language Model Calls into Self-Improving Pipelines· arXiv / Stanford· 2023· faithful paraphrase
Pavel Durov

Pavel Durov

Telegram founder; arrested in France 2024

endorses

Argues open and uncensored communication infrastructure is foundational to civil society and should not be sacrificed to AI-content-moderation imperatives.

Outdated laws should not be used to force platforms to police speech automatically. The principle that providers of communication tools are not responsible for individual user actions has to survive the AI era.
articlePavel Durov statement on Telegram and platform liability· Telegram· 2024-09· faithful paraphrase
Peter Wang

Peter Wang

Co-founder of Anaconda; scientific Python and AI

endorses

Argues open-source infrastructure is a public good; AI companies building on top of it should contribute back.

The AI revolution runs on open-source that nobody is funding sustainably.
articleAnaconda· Anaconda· 2024· loose paraphrase

Robin Rombach

Black Forest Labs co-founder; Stable Diffusion lead

endorses

Argues open-weight image-generation models, not just open APIs, are necessary for the field to keep developing safety properties externally; co-built Black Forest Labs on this thesis.

Latent diffusion models compress the image generation problem into a learned latent space, making high-resolution generation accessible to consumer hardware.
§ paperHigh-Resolution Image Synthesis with Latent Diffusion Models· arXiv / CVPR· 2022· faithful paraphrase

Sasha Rush

Cornell Tech professor; HuggingFace research scientist

endorses

Champions open replication of frontier research; argues open replication and tutorials are how the field grows.

Open replication is the lifeblood of an honest research field.
articleSasha Rush, Cornell Tech· Cornell Tech· 2024· loose paraphrase

Sebastian Raschka

Lightning AI; ML educator and author

endorses

Argues open educational materials and open-source ML tooling produce the most resilient version of the field; an industry that depends entirely on closed APIs has lost its capacity to teach itself.

Open-source tooling and open educational materials are how the ML field reproduces itself. We can't outsource that to a few frontier labs without losing something important.
blogSebastian Raschka, Ahead of AI· Substack· 2024· faithful paraphrase

Soumith Chintala

PyTorch creator; Meta AI

endorses

Built the most-used open-source deep-learning framework; advocates open infrastructure as the foundation of AI research.

Open infrastructure is what allowed AI research to scale. It is what allows new entrants to compete.
articlePyTorch· PyTorch· 2024· loose paraphrase

Stella Biderman

Executive Director of EleutherAI

endorses

Argues open research and open models are essential; the field cannot be governed if only a handful of closed labs can audit models.

Science doesn't work in secret. Open research is how safety work gets done.
articleEleutherAI· EleutherAI· 2023· loose paraphrase

Tianqi Chen

CMU professor; XGBoost and TVM creator

endorses

Built foundational open-source AI tools used industry-wide. Argues the AI ecosystem depends on open infrastructure.

Open infrastructure is what allowed many actors to build AI. Closing it now would be a fundamental mistake.
articleTianqi Chen, CMU· tqchen.com· 2024· loose paraphrase
Tim Berners-Lee

Tim Berners-Lee

Inventor of the World Wide Web

endorses

Argues decentralised data architecture (Solid pods) is the structural response to AI and platform exploitation.

“The web has evolved into an engine of inequity and division; swayed by powerful forces who use it for their own agendas.”
articleTim Berners-Lee urges decentralized web to counter AI exploitation· TechSpot· 2024· direct quote

Tim Dettmers

Efficient-training and quantization researcher

endorses

Develops efficiency techniques that broaden access to frontier-scale training.

If only frontier labs can run frontier models, then only frontier labs will shape what they do.
blogTim Dettmers, efficiency research· timdettmers.com· 2024· loose paraphrase
Vukosi Marivate

Vukosi Marivate

Univ Pretoria; African NLP / Masakhane

endorses

Argues open and community-led NLP is the only credible path for African languages; built Masakhane as a distributed research network independent of frontier labs.

Frontier-lab tokenizers and pretraining mixtures barely cover African languages. The community has had to build its own datasets, evaluations, and benchmarks, often from scratch.
articleMasakhane Research Community· Masakhane· 2024· faithful paraphrase
Wes McKinney

Wes McKinney

Pandas creator; Posit/RStudio data infrastructure

endorses

Argues open-source data infrastructure is the foundation of AI; without open data tooling, AI access is gated by industry.

Open data infrastructure is the precondition for democratised AI work.
articleWes McKinney homepage· wesmckinney.com· 2024· loose paraphrase
Yann LeCun

Yann LeCun

Chief AI Scientist at Meta; outspoken AI-doom skeptic

endorses

Has pushed Meta toward open-weighting Llama models and argues open weights are necessary for diverse, safe AI.

The smartest among us do not want to dominate the others. The idea that because a system is intelligent, it wants to take control is completely false.
articleMeta Chief A.I. Scientist Yann LeCun says A.I. doomsayers are 'preposterous'· Fortune· 2023-06-14· faithful paraphrase