Skip to content

Anthropic

How Dario Amodei built Anthropic by betting that safety and capability are the same problem

January 15, 2025·7 min read

Anthropic's founders left OpenAI over safety concerns and built a competitor. Their bet — that the safest AI would also be the most capable — is the most consequential wager in the history of the technology industry.

The origin

Dario Amodei was the Vice President of Research at OpenAI in 2020 and 2021. He had joined in 2016 after a stint at Google Brain, and he had led the development of GPT-2 and GPT-3 — the models that established OpenAI's position at the frontier of large language model research.

By 2021, Amodei and a group of colleagues had developed serious concerns about the direction of AI development at OpenAI. The concerns were not primarily about the technology. They were about the pace of deployment, the governance structure, and what they saw as insufficient attention to the risks of increasingly capable AI systems.

In the summer of 2021, Amodei and his sister Daniela Amodei — who had been OpenAI's VP of Operations — left the company along with five other senior researchers and engineers. They founded Anthropic in San Francisco with a specific thesis: that the most important problem in AI was not making models more capable, but making them more reliably aligned with human values and intentions.

The founding team raised $124 million in a Series A round led by Spark Capital. The pitch was unusual for a technology startup: Anthropic was not promising to build the most capable AI. It was promising to build the safest one — and arguing that, in the long run, those were the same thing.

The challenge

Anthropic's founding premise created an immediate commercial challenge. Safety research is expensive, slow, and does not produce the kind of product demos that generate press coverage and user growth. OpenAI had ChatGPT. Google had Bard. Anthropic had research papers.

The company's first public product, Claude, launched in March 2023 — four months after ChatGPT. Claude was built on a technique Anthropic had developed called Constitutional AI: a method for training AI models to follow a set of explicit principles, rather than relying solely on human feedback to shape behavior. The constitutional approach made Claude's reasoning more transparent and its refusals more consistent — qualities that mattered less to casual users and enormously to enterprise customers in regulated industries.

The commercial challenge was real. Anthropic was competing against OpenAI, which had a two-year head start, a $10 billion investment from Microsoft, and the most recognized AI brand in the world. It was competing against Google, which had more compute, more data, and more distribution than any other company on earth. It was a startup with a safety thesis in a market that was moving at startup speed.

The fundraising helped. Amazon invested $4 billion in Anthropic in 2023, with a commitment to invest up to $4 billion more — the largest investment in Anthropic's history and a signal that the cloud infrastructure companies saw Anthropic as a strategic partner rather than just a vendor. Google invested $300 million. The total capital raised by 2024 exceeded $7 billion.

The breakthrough

Claude 2, released in July 2023, was the first version of Claude that could compete directly with GPT-4 on capability benchmarks. It had a 100,000-token context window — the ability to process and reason about roughly 75,000 words of text in a single conversation — that was substantially larger than any competing model. For enterprise customers who needed to analyze long documents, contracts, or codebases, the context window was a decisive advantage.

Claude 3, released in March 2024, came in three versions — Haiku, Sonnet, and Opus — designed for different cost and capability tradeoffs. Claude 3 Opus outperformed GPT-4 on several standard benchmarks, marking the first time a non-OpenAI model had clearly led on capability measures. The release established Anthropic as a genuine peer competitor to OpenAI rather than a safety-focused alternative.

The Constitutional AI approach, which had seemed like a constraint in 2021, had become a commercial asset by 2024. Enterprise customers in healthcare, finance, legal services, and government — industries where AI errors have real consequences and where regulatory scrutiny is intense — preferred Claude's more predictable behavior and more transparent reasoning. Anthropic's safety focus had attracted exactly the customers who were willing to pay enterprise prices for AI that they could trust.

Anthropic's revenue grew from essentially zero in early 2023 to an annualized run rate of over $1 billion by late 2024 — a growth trajectory that matched or exceeded OpenAI's early commercial growth.

The safety thesis

Anthropic's founding thesis — that safety and capability are the same problem — has been tested by the company's own research. The Alignment Science team, which Dario Amodei has described as the core of Anthropic's mission, has published research on interpretability (understanding what is happening inside AI models), on Constitutional AI, and on the risks of increasingly capable systems.

The research has produced findings that are both technically important and commercially relevant. Interpretability research — understanding why a model produces a particular output — is directly useful for enterprise customers who need to audit AI decisions. Constitutional AI — training models to follow explicit principles — is directly useful for customers who need AI that behaves consistently across a wide range of inputs.

The safety work has also attracted talent. Researchers who are motivated by the long-term implications of AI development, rather than by product velocity or commercial success, have chosen Anthropic over competitors. The talent concentration in safety-focused research has given Anthropic a genuine advantage in the areas that matter most to its target customers.

The impact

Anthropic's impact on the AI industry has been disproportionate to its size. The company's research on Constitutional AI, interpretability, and AI safety has influenced how every major AI lab thinks about model development. OpenAI, Google DeepMind, and Meta AI have all published research that builds on or responds to Anthropic's work.

The Claude model family has established a genuine alternative to OpenAI's GPT series for enterprise customers. The competition has been good for the market: it has driven both companies to improve their models faster, to be more transparent about capabilities and limitations, and to develop better tools for enterprise deployment.

For the broader technology industry, Anthropic's success has validated a thesis that was not obvious in 2021: that safety-focused AI development is commercially viable, not just morally preferable. The customers who care most about AI safety — regulated industries, government agencies, large enterprises with reputational risk — are also the customers with the largest budgets and the longest contract cycles.

The legacy

Anthropic's story is still in its early chapters. The company's mission — the responsible development and maintenance of advanced AI for the long-term benefit of humanity — is ambitious enough that it cannot be evaluated on a five-year timeline.

But the business lesson is already clear. Anthropic's founders left a more successful company because they believed the most important problem was being underweighted. They built a company around that belief, attracted customers who shared it, and created a product that turned the constraint into an advantage.

Anthropic's safety focus was supposed to be a competitive disadvantage. It became the reason the customers who mattered most chose Claude over every alternative.

The question Anthropic answered — what if the thing we're most worried about is also the thing our best customers are most worried about? — is the right question for any business trying to differentiate in a crowded market. The constraints you take seriously, the standards you refuse to compromise on, the problems you treat as existential rather than manageable — these are not liabilities. They are the foundation of a brand that the right customers will trust.

Continue learning

AmericaOwl

The Owl Brief

One story. One growth lesson. One practical idea — every Sunday.

No spam. Unsubscribe any time.