Part IV — The Ecosystem

The Labs

OpenAI, Anthropic, DeepMind, Meta AI, Mistral, and the organizational landscape of frontier AI research.

The Institutional Shift

For most of its history, AI research happened primarily at universities. MIT, Stanford, CMU, Toronto, Montreal, Berkeley — these were the places that mattered. Faculty published papers, trained PhD students, and set the research agenda. Industry labs existed (AT&T Bell Labs, Xerox PARC, IBM Research) but generally operated on long time horizons and published openly.

That changed in the 2010s. As deep learning began producing commercially valuable results — image recognition, machine translation, recommendation systems — talent and compute migrated from universities to industry at an accelerating rate. Today, the most consequential AI research comes from a handful of private labs, and the compute required for frontier work is beyond the reach of any university. A single training run for a model like GPT-4 is estimated to cost over $100 million in compute alone.1

Understanding these labs — their histories, incentives, and philosophies — is essential context for reading the field. Their decisions about what to build, what to release, and what to keep private shape what AI looks like for everyone.

OpenAI

Founded in 2015 as a nonprofit research lab with the stated mission to "ensure that artificial general intelligence benefits all of humanity." Early backers included Elon Musk, Sam Altman, Peter Thiel, Reid Hoffman, and others, who collectively pledged $1 billion. The original thesis was that concentrating AGI capabilities in a single company would be dangerous, so an open, nonprofit alternative was needed.

The lab's trajectory since then has been a case study in how compute costs shape organizational structure:

OpenAI's significance is less about any single model and more about proving that scaling language models could produce commercially viable products. They didn't invent transformers (that was Google Brain), they didn't invent RLHF (that was DeepMind), and they didn't invent large language models (GPT-1 was one of several contemporary efforts). What they did was push the scaling curve harder than anyone else and build the first product that made LLMs legible to non-technical users.

Anthropic

Founded in 2021 by Dario Amodei (former VP of Research at OpenAI) and Daniela Amodei (former VP of Operations at OpenAI), along with several other ex-OpenAI researchers including Tom Brown (first author on the GPT-3 paper) and Chris Olah (known for mechanistic interpretability work).

The founding thesis was that AI development needed to be safety-focused from the organizational level, not just the research level. Key distinctions from OpenAI:

The Claude model family is Anthropic's product. Claude 3.5 Sonnet (mid-2024) demonstrated that safety-focused training didn't have to come at the cost of capability — it performed competitively with GPT-4 on most benchmarks while generally refusing more categories of harmful requests. This is the tool you build with daily.

Google DeepMind

The result of a 2023 merger between DeepMind (founded 2010 in London, acquired by Google in 2014 for ~$500 million) and Google Brain (Google's internal AI research division). The merged entity is led by Demis Hassabis, DeepMind's co-founder.

DeepMind's reputation was built on reinforcement learning and game-playing AI:

Google Brain's contribution was different: fundamental research that the field built on. The transformer architecture came from Google Brain (Vaswani et al., 2017). BERT came from Google AI Language. TensorFlow, the first major deep learning framework, came from Google Brain. Word2Vec, which kicked off the word embedding era, came from Google Research.

The merged entity has the deepest research talent, the most compute (TPU pods that no other lab can match), and the broadest product surface (Search, Gmail, YouTube, Android, Cloud). Their challenge is organizational: translating research breakthroughs into products within a large corporate bureaucracy. OpenAI shipped ChatGPT while Google was still debating internally how to release a chatbot without embarrassment.

Meta AI (FAIR)

Meta's AI lab, originally called FAIR (Facebook AI Research), was founded in 2013 and led by Yann LeCun, one of the three "godfathers of deep learning" (with Geoffrey Hinton and Yoshua Bengio). LeCun's leadership gives FAIR a distinctive character: more emphasis on self-supervised learning and less on the autoregressive language model paradigm that dominates the field.

Meta's defining strategic choice is open weights:

Why open weights? Meta's business model doesn't depend on selling API access to models. It depends on social media advertising. For Meta, making AI models freely available commoditizes the model layer, prevents competitors (OpenAI, Google) from building a moat around proprietary models, and creates an ecosystem of developers who build on Meta's stack. It's strategy, not charity — but the effect is that Meta has done more to democratize frontier AI than any other company.

Key idea: The "open" in "open source" and "open weights" are different things. Open-source means you can see and modify the code. Open weights means you can download and run the trained model. But most "open" models don't release training data, training code, or the compute budget needed to reproduce the result. You can use the model, but you can't truly replicate or audit the process that created it. This distinction matters for both science and safety.

Mistral

Founded in 2023 in Paris by Arthur Mensch, Guillaume Lample, and Timothee Lacroix — all formerly at Meta AI/FAIR and Google DeepMind. Mistral represents the European entry into frontier model development.

Mistral's approach is notable for efficiency:

Mistral raised over $600 million within its first year, reaching a $6 billion valuation. Its existence matters for AI geopolitics: the EU's AI Act creates a regulatory environment where European-developed models may have strategic advantages in European markets.

The Rest of the Field

Beyond the five labs above, several others are worth tracking:

Lab Notable For Status
xAI Elon Musk's lab. Grok models. Massive compute investment (100,000 H100 cluster). Integrated with X (Twitter). Well-funded, building infrastructure
Cohere Enterprise-focused. Retrieval-augmented generation. Founded by Aidan Gomez (co-author of the transformer paper). Viable enterprise player
AI21 Labs Israeli lab. Jamba models (SSM-transformer hybrid). Founded by AI pioneers from Hebrew University. Niche but innovative
Stability AI Stable Diffusion (image generation). Open-source focus. Financial and organizational turbulence. Uncertain future, lasting impact
Inflection AI Founded by Mustafa Suleyman (DeepMind co-founder). Built Pi chatbot. Most of team absorbed by Microsoft in 2024. Effectively acquired
DeepSeek Chinese lab. DeepSeek-V2/V3 models competitive with frontier Western models. Very strong on efficient architectures (MLA attention, MoE). Major player, limited Western access
Alibaba (Qwen) Qwen model family. Strong multilingual performance. Open weights for several model sizes. Active, open-weight releases
The AI Lab Landscape Closed Weights More Open More Closed Compute & Scale OpenAI GPT-4, o3, DALL-E Google DeepMind Gemini, AlphaFold, TPUs Anthropic Claude, Const. AI xAI Grok Meta AI LLaMA, PyTorch Mistral Mixtral, efficiency DeepSeek V3, R1, MoE Cohere Enterprise RAG Stability AI Stable Diffusion Qwen Alibaba AI21 Safety-focused / API-gated Open-weight ecosystem

The Open vs. Closed Debate

This is the defining disagreement in AI governance, and it cuts across technical, economic, and ethical lines.

The Case for Closed

Proponents (primarily OpenAI, Anthropic, and some safety researchers) argue:

The Case for Open

Proponents (primarily Meta, Mistral, and the broader open-source community) argue:

The honest answer is that both sides are right about the risks they identify and wrong about the risks they minimize. Concentrated power is genuinely dangerous. Unrestricted access to frontier models is also genuinely dangerous. There's no clean solution, and anyone who tells you otherwise is selling something or hasn't thought it through.

How the Labs Make Money

This matters because it determines what gets built and what gets prioritized.

Revenue Model Who How It Works
API access OpenAI, Anthropic, Cohere, Google Charge per token (input and output) for model access. Simple, scalable, usage-based. OpenAI's API generates significant revenue, though exact figures are not public.
Subscriptions OpenAI (ChatGPT Plus/Pro), Anthropic (Claude Pro/Max) Fixed monthly fee for enhanced access. $20/month for Plus-tier products is the standard price point. Higher tiers ($200/month) for power users.
Enterprise contracts All major labs Custom deployments, fine-tuning, dedicated capacity, compliance guarantees. Higher margins, longer sales cycles.
Cloud integration Google (Vertex AI), Microsoft (Azure OpenAI) Model access bundled with cloud infrastructure. Microsoft resells OpenAI models through Azure; Google offers Gemini through Vertex AI.
Advertising / ecosystem Meta, Google AI improves core products (ad targeting, search, content recommendations) rather than being sold directly. Open-weight models create ecosystem lock-in.

The economics are important: as of early 2025, none of the pure-play AI labs (OpenAI, Anthropic) are profitable. They're subsidized by investor capital, burning cash to build capability and market share in the hope that the market will be enormous. This is the classic Silicon Valley playbook (Amazon lost money for years), but it means the current AI landscape is shaped more by investor confidence than by sustainable business models.

The University Response

Universities haven't disappeared from AI research, but their role has changed. They're less likely to produce frontier models (they can't afford the compute) and more likely to contribute foundational theory, evaluation methodology, safety research, and applications to non-commercial domains.

A few university labs remain highly influential:

The brain drain from universities to industry is real and well-documented. The solution isn't obvious: you can't train a frontier model on a professor's grant budget. Some labs (Google DeepMind, Meta) address this by maintaining research partnerships with universities and allowing dual appointments, but the gravitational pull of industry resources is strong.


The labs build the models, but they don't build the chips. That distinction matters enormously, because the entire AI industry runs on hardware produced by a surprisingly small number of companies. The hardware stack — who builds what, why Nvidia dominates, and what the supply chain looks like — is the subject of the next chapter.