Anthropic CEO谈生成式AI的商业模式、与OpenAI的竞争、AI的潜力、末日论

Original-VideoinhaltVideo erweitern
  • Dario Amodei expresses frustration at being labeled a "doomer" while addressing the urgency of AI.
  • He highlights advancements in AI's capabilities and the potential economic impact they could have.
  • Dario discusses the importance of scaling laws and the exponential growth of AI models.
  • He reflects on his personal experiences driving his passion for impactful AI.
  • The need for responsible development and caution in AI is emphasized throughout the conversation.

I get very angry when people call me a doomer. You know, when someone's like, this guy's a doomer, he wants to slow things down. You heard what I just said. Like, you know, my father died because of you, cures that, you know, could have happened a few years later. I understand the benefit of this technology.

I'm sure you've heard the criticism from people like Jensen who say, well, Dario thinks he's the only one who can build this safely and therefore wants to control the entire industry. I've never said anything like that. That's an outrageous lie. That's the most outrageous lie I've ever heard.

Anthropic CEO Dario Amodei joins us to talk about the path forward for artificial intelligence, whether generative AI is a good business, and to fire back at those who call him a doomer. And he's here with us in studio at Anthropic headquarters in San Francisco. Dario, it's great to see you again. Welcome to the show. Thank you for having me.

So, let's recap the past couple of months for you. You said AI could wipe out half of entry-level white-collar jobs. You cut off Windsurf's access to Anthropic's top-tier models when you learned that OpenAI was going to acquire them. You asked the government for export controls and annoyed Nvidia CEO Jensen Wong. What's gotten into you?

I think Anthropic myself and Anthropic are always focused on trying to do and say the things that we believe. As we've gotten more close to AI systems that are more powerful, I've wanted to say those things more forcefully, more publicly, to make the point clearer.

I've been saying for many years that we can talk in detail about them, but we have these scaling laws. AI systems are getting more powerful. They're going from the level of a few years ago; they were barely coherent. A couple of years ago, they were at the level of a smart high school student. Now we're getting to smart college student Ph.D.

They're starting to apply across the economy. All the issues related to AI, ranging from national security issues to economic issues, are starting to become quite near to where we're actually going to face them.

Even though in some form, Anthropic has been saying these things for a while, I think the urgency of these things has gone up. I want to make sure that we say what we believe and that we warn the world about the possible downsides, even though no one can say what's going to happen. We're saying what we think might happen, what we think is likely to happen.

We back it up as best we can, although it's often extrapolations about the future where no one can be sure. We see ourselves as having the duty to warn the world about what's going to happen. That's not to say I think there's an incredible number of positive applications of AI. I've continued to talk about that.

I wrote this essay, Machines of Loving Grace. I feel, in fact, that I in Anthropic have often been able to do a better job of articulating the benefits of AI than some of the people who call themselves optimists or accelerationists. We probably appreciate the benefits more than anyone, but for exactly the same reason.

Because we can have such a good world if we get everything right, I feel obligated to warn about the risks. All of this is coming from your timeline. Basically, it seems like you have a shorter timeline than most. So you were feeling a sense of urgency to get out there because you think that this is imminent.

I'm not sure. I think it's very hard to predict, particularly on the societal side. So if you say, when are people going to deploy AI or when are companies going to use X dollars of spend of AI, or when will AI be used in these applications? Or when will it drive these medical cures? That's harder to say. The underlying technology is more predictable, but still no one knows.

On the underlying technology, I've started to become more confident. There is some uncertainty about it. I think the exponential that we're on could kind of totally peter out. I think there's maybe a 20 or 25% chance that sometime in the next two years the models just stop getting better for reasons we don't understand, or maybe reasons we do understand, like data or compute availability.

And then everything I've said just seems totally silly and everyone makes fun of me for all the warnings I've made. And I'm just totally fine with that, given the distribution that I see.

So I should say that this conversation is part of a profile I'm writing about you. I've spoken with more than two dozen people who've worked with you, who know you, who've competed with you, and I'm going to link that in the show notes. If anybody wants to read it, it's free to read. But one of the themes that has come through across everybody I've spoken with is that you have the shortest timeline of any of the major lab leaders.

You just referenced it just now. Why do you have such a short timeline? And why should we believe in yours? It really depends on what you mean by timeline. One thing, and I've been consistent on this over the years, is there are these terms in the AI world like AGI and superintelligence. You'll hear leaders of companies say, we've achieved AGI, we're moving on to superintelligence.

I think these terms are totally meaningless. I don't know what AGI is. I don't know what superintelligence is. It sounds like a marketing term. It is marketing. Yeah. It sounds like something designed to activate people's dopamine.

So you'll see in public, I never use those terms, and I'm actually careful to criticize their use. But despite that, I am indeed one of the most bullish about AI capabilities improving very fast. The thing I think is real, that I've said over and over again, is the exponential.

The idea that every few months we get an AI model that is better than the AI model we got before, and we get that by investing more compute in AI models, more data, and more new types of training models. Initially, this was done by what's called pre-training, which is when you just feed a bunch of data from the Internet into the model.

Now we have a second stage that's reinforcement learning or test-time compute or reasoning, or whatever you want to call it. I think of it as a second stage that involves reinforcement learning. Both of those things are scaling up together, as we've seen with our models and as we've seen with models from other companies. I don't see anything blocking the further scaling of that.

There's some stuff about how do we broaden the tasks. On the RL side of it, we've seen more progress on, say, math and code, where the models are getting pretty close to a high professional level and less on more subjective tasks. But I think that is very much a temporary obstacle.

So when I look at it, I see this exponential and I say, look, people aren't very good at making sense of exponentials. Like, if something is doubling every six months, then two years before it happens, it looks like it's only 1/16 of the way there.

We are sitting here in the middle of 2025 and the models are really starting to explode in terms of the economy. If you look at the capabilities of the model, they're starting to saturate all the benchmarks. If you look at revenue, Anthropic's revenue every year has grown 10x.

We're conservative and we say it can't grow 10x this time. I never assume anything and actually always am very conservative in saying I think it's going to slow down on the business side. But we went from zero to 100 million dollars in 2023. We went from 100 million dollars to a billion in 2024.

And this year, in the first half of the year, we've gone from 1 billion dollars to I think as of speaking today, it's well above 4. It might be 4.5. So if you think about it, suppose that exponential continued for two years.

I'm not saying it will, but suppose it continued for two years, you're like well into the 100 billions. I'm not saying that will happen. I'm saying the situation is that when you're on an exponential, you can really get fooled by it. Two years away from when the exponential goes totally crazy, it looks like it's just starting to be a thing.

That's the fundamental dynamic we see with the Internet. We saw that with the Internet in the 90s where it was like networking speeds and the underlying speed of the computers were getting fast. Over a few years it became possible to build a digital global communications network on top of all this when it wasn't possible just a few years ago.

Almost no one, except for a few people, really saw the implications of that and how fast it would happen. And that's where I'm coming from.

That's what I think. Now, I don't know if a bunch of satellites crashed; maybe the Internet would have taken longer. If there was an economic crash, maybe it would have taken a little longer. So we can't be sure of the exact timelines, but I think people are getting fooled by the exponential and not realizing how fast it might be.

But so many folks in the AI industry are talking about diminishing returns from scaling. Now, that really doesn't fit with the vision you just laid out. Are they wrong? Yeah, from what we've seen, I can only speak in terms of the models at Anthropic. But what I've seen in terms of the models at Anthropic.

If we look at coding, coding is one area where I think Anthropic models have advanced very quickly. Adoption has been very quick. We're not just a coding company. We're planning to expand to many areas. But if you look at coding, we release 3.5 Sonnet, a model we call 3.5 Sonnet V2, which let's call it 3.6 Sonnet now 3.7 Sonnet and then 4.0 Sonnet and 4.0 Opus.

That series of four or five models, each one got substantially better at coding than the last. If you want to look at benchmarks, you can look at swee bench growing from I think 18 months ago was at like 3% or something, growing all the way to 72 to 80% depending on how you measure it.

The real usage has grown exponentially as well. Where we're heading more and more towards autonomously, you can just use these models. I think the actual majority of code that's written at Anthropic is at this point written by, or at least with the involvement of one of the cloud models.

Various other companies have made similar statements. We see the progress as being very fast, and the exponential is continuing, and we don't see any diminishing returns. But there are some liabilities, it seems like with large language models, for instance, continual learning.

We had Duar Kesh on a couple of weeks ago. Here's how he put it: the lack of continual learning is a huge, huge problem. The LM baseline at many tests might be higher than an average human, but you're stuck with the abilities you get out of the box. So you just make the model, and that's it. It doesn't learn.

That seems like a glaring liability. What do you think about that? So first of all, I would say even if we never solved continual learning, I think that the potential for the LLMs to do incredibly, incredibly well, to affect things at the scale of the economy will be very high.

If I think of the field I used to be in, biology and medicine, let's say I had a very smart Nobel Prize winner. And I said, okay, you've discovered all these things, you have this incredibly smart mind, but you can't read new textbooks or absorb any new information.

I mean, that would be difficult but still, if you had 10 million of those, they're still going to make a lot of biology breakthroughs. They're going to be limited; there are some things humans can do that they can't. But even that, even if I said you never solve that, I think people are underestimating the impact.

But look, context windows are getting longer, and models actually do learn during the context window. So as I talk to the model during the context window, I have a conversation. It absorbs information.

The underlying weights of the model may not change. But just like I'm talking to you here and we're having a conversation and I listen to the things you say, the models are able to do that. From a machine learning perspective, there's no reason we can't make the context length 100 million words today, which is roughly what a human hears in their lifetime.

There's no reason that we can't do that. It's really inference support. Even that fills in many of the gaps, not all the gaps, but it fills in many of the gaps. There are several things around new types of reinforcement learning, learning, training.

Many years ago, we used to talk about inner loops and outer loops. The inner loop is like, I have some episode, and I learned some things in that episode. The outer loop is the agents learning over episodes.

I think maybe that inner loop, outer loop structure is a way to learn the continual learning. One thing we've learned in AI is whenever it feels like there's some fundamental obstacle, like two years ago, we thought there was this fundamental obstacle around reasoning.

It turned out just to be RL. You just train with RL and you let the model write some stuff down; you let the model write things down to try and figure out objective math problems without being too specific. I think we already might have some evidence to suggest that this is another of those problems that is not as difficult as it seems that will fall to scale, plus a slightly different way of thinking about things.

Do you think your obsession with scale might blind you to some of the new techniques like Demisis, saa Says to get to AGI, or you might call it super powerful AGI, whatever? We're developing new techniques every day.

Claude is very good at code, and we don't really talk externally that much about why Claude is so good at code. Why is it so good? Like I said, we don't talk externally about it. Every new version of Claude that we make has improvements to the architecture, improvements to the data that we put into it, improvements to the methods that we use to train it.

So we're developing new techniques all the time. New techniques are a part of every model that we build. That's why I've said these things about we're trying to optimize for talent density as much as possible. You need that talent density in order to invent the new techniques.

There's one thing that's been hanging over this conversation, which is that maybe Anthropic is the company with the right idea but the wrong resources. Because you look at what's happening with XAIAI and inside Meta, where Elon's built his massive cluster, Mark Zuckerberg is building this 5-gigawatt data center. They are putting so many resources towards scaling up.

Is it possible? Anthropic, obviously you have raised billions of dollars, but these are trillion-dollar companies. We've raised, I think at this point, a little short of $20 billion. It's not bad. So that's not nothing.

If you look at the size of the data centers that we're building with, for example, Amazon, I don't think our data center scaling is substantially smaller than that of any of the other companies in the space. In many cases, these things are limited by energy, they're limited by capitalization.

When people talk about these large amounts of money, they're talking about it over several years. When you hear some of these announcements, sometimes they're not funded yet. We've seen the size of the data centers that folks are building, and we're actually pretty confident that we will be within a rough range of the size of data centers that they've built.

You talked about talent density. What do you think about what Mark Zuckerberg is doing on the talent density front? Combining that with these massive data centers, it seems like he's going to be able to compete.

Yeah, this is actually very interesting because one thing we noticed is that relative to other companies, I think a lot fewer people from Anthropic have been caught by these offers. It's not for lack of trying. I've talked to plenty of people who got these offers at Anthropic and who just turned them down, who wouldn't even talk to Mark Zuckerberg, who said no, I'm staying at Anthropic.

Our general response to this was I posted something to the whole company Slack, where I said, look, we are not willing to compromise our compensation principles, our principles of fairness to respond individually to these offers.

The way things work at Anthropic is there's a series of levels. When a candidate comes in, they get assigned a level, and we don't negotiate that level because we think it's unfair. We want to have a systematic way. If Mark Zuckerberg throws a dart at a dartboard and hits your name, that doesn't mean that you should be paid 10 times more than the guy next to you who’s just as skilled, just as talented.

My view of the situation is that the only way you can really be hurt by this is if you allow it to destroy the culture of your company by panicking, by treating people unfairly in an attempt to defend the company. I think actually this was a unifying moment for the company where we didn't give in.

We refused to compromise our principles because we had the confidence that people are at Anthropic because they truly believe in the mission. I think that gets to how I see this. What they are doing is trying to buy something that cannot be bought, and that is alignment with the mission.

There are selection effects here. Are they getting the people who are most enthusiastic, who are most mission-aligned, who are most excited to? But they have talent and GPUs. You're not underestimating them. We'll see how it plays out. I am pretty bearish on what they're trying to do.

Let's talk a little bit about your business because a lot of people have been wondering, is the business of generative AI a real thing? I'm also curious. I have questions all the time. You talked about how much money you've raised—close to $20 billion. You've raised $3 billion from Google, $8 billion from Amazon, $3.5 billion from a new round led by Lightspeed, who I've spoken with.

What is your pitch? Because you are not part of a big tech company or out there on your own. Do you just bring the scaling laws and say, can I have some money? My view of this has always been that talent is the most important thing.

If you go back three years ago, we were in a position where we had raised mere hundreds of millions. OpenAI had already raised $13 billion from Microsoft and, of course, the large hyper-cap tech companies. We were sitting on $100 billion, $200 billion.

Basically, the pitch we made then is we know how to make these models better than others do. There may be a curve of scaling laws, but look, if we are in a position where we can do for $100 million what others can do for a billion, and we could do for $10 billion what they can do for $100 billion, then it's 10 times more capital efficient to invest in Anthropic than it is to invest in these other companies.

Would you rather be in a position where you can do anything for 10 times cheaper or where you start with a large pile of money if you can do things 10 times cheaper? The money is a temporary defect that you can remedy if you have this intrinsic ability to build things for the same price much better than anyone else, or as good as anyone else for a much lower price.

Investors aren't idiots, at least they aren't always idiots. It depends which one you go to. I'm not going to name any names, but they basically understand the concept of capital efficiency. We've been in a position three years ago where these differences were like 1000x, and now you're saying with $20 billion, can you compete with $100 billion?

And my answer is basically yes because of the talent density. I've said this before, but Anthropic is actually the fastest-growing software company in history at the scale that it's at. We grew from $0 to $100 million in 2023, $100 million to a billion in 2024, and this year we've grown from $1 billion to, I think I said this before, $4.5 billion.

So that 10x a year, every year, I suspect that will grow at that scale. Every year, I'm almost afraid to say it publicly because I'm like no, it couldn't possibly happen again. So I think the growth at that scale speaks for itself in terms of our ability to compete with the big players.

OK, CNBC says 60 to 75% of Anthropic sales are through the API, according to internal documents. Is that still accurate? I won't give exact numbers, but the majority does come through the API. Although we also do have a flourishing apps business, and I think more recently the max tier which power users use, as well as Claude Code, which coders use.

I think we have a thriving and fast-growing apps business. But yes, the majority comes through the API. You're making the most pure bet on this technology, like OpenAI might be betting on ChatGPT, and Google might be betting on the fact that no matter where the technology goes, it can integrate into Gmail and Calendar.

So why have you made this bet on the pure bet on the tech itself? Yeah, I mean, I would say I wouldn't quite put it that way. I think I would describe it more as we bet on business use cases of the model more so than we've bet on the API per se.

And it's just that the first business use cases of the model come through the API. So as you mentioned, OpenAI is very focused on the consumer side. Google is very focused on the existing products that Google has.

Our view is that if anything, the enterprise use of AI is going to be greater even than the consumer use of AI. I should say the business use because it's enterprise, it's startups, it's developers, and it's power users using the model for productivity.

I also think that being a company that's focused on the business use cases actually gives us better incentives to make the models better. A thought experiment that I think is worth running is suppose I have this model and it's as good as an undergrad at biochemistry.

And then I improve it, and it's as good as a Ph.D. student at biochemistry. If I go to a consumer, if I give them a chatbot and I say great news, I've improved the model from undergrad to graduate level in biochemistry, maybe, I don't know, 1% of consumers care about that at all.

99% are just going to be like, I don't understand it either way. But now suppose I go to Pfizer, and I say I've improved this from undergraduate biochemistry to graduate biochemistry. This is going to be the biggest deal in the world. They might pay 10 times more for something like that.

It might have 10 times more value to them. The general aim of making the models solve the problems of the world is to make them smarter and smarter, but also able to bring many positive applications—the things I wrote about in 'Machines of Loving Grace'—of solving the problems of biomedicine, solving the problems of geopolitics, solving the problems of economic development, as well as more prosaic things like finance or legal or productivity or insurance.

I think it gives a better incentive to develop the models as far as possible. And I think in many ways, it may even be a more positive business. So I would say we're making a bet on the business use of AI because it's most aligned with the exponential.

OK, then briefly, how did you decide to go with the coding use case? Yes. So originally, as happens with most things, we were trying to optimize for making the model better at a bunch of stuff. Coding particularly stood out in terms of how valuable it was.

I've worked with thousands of engineers, and there was a point about a year, a year and a half ago where one of the best I'd ever worked with said every previous coding model has been useless to me, and this one finally was able to do something I wasn't able to do.

And then after we released, it started getting quick adoption. This was around the time that a lot of the coding companies, like cursor, windsurf, GitHub, and augment code, started exploding in popularity. When we saw how popular it was, we doubled down on it. My view is that coding is particularly interesting because A, the adoption is fast and B, getting better at coding with the models actually helps you to develop the next model.

It has a number of advantages. Now you're selling your AI coding through Claude Code. But it's very interesting. The pricing model has been confounding to some. You can spend dollars a month and get the equivalent.

I spoke to one developer. They got the equivalent of $6,000 a month from your API. Ed Zitron has pointed out the more popular that your models get, the more money you're going to lose. If people are superusers of this technology, how does that make sense?

Actually, pricing schemes and rate limits are surprisingly complicated. Some of this is basically the result of when we release Claude Code and the Max tier, which we eventually tied together, not fully understanding the implications of the ways in which people could use the models and how much they were actually able to get.

So over the last few days, as of the time of this interview, we've adjusted that, particularly on the larger models like Opus. I think it's no longer possible to spend that much with a $100 subscription. It's possible more changes will come in the future.

But we're always going to have a distribution of users who use a lot and users who use some amount. It doesn't necessarily mean we're losing money; there are some users who get more, if you were to measure via API credits. Spend, getting a better deal on the consumer subscription than they would on the API products.

There are a lot of assumptions there, and I can tell you that some of them are wrong. We are not in fact losing money. But I guess there's another question about whether you can continue to serve these use cases and not raise prices.

Just to give you a couple of stats, there are some developers that are upset because using Anthropic's newer models and cursor's costing them more than ever. Has startups that I've spoken with say Anthropic is down a bunch because they can't get access to the GPUs.

At least that's what they imagine is happening. I was just with Amjad Massad at Replit in an interview that we're going to air next week, who said there was a period of time where the price per token to use these models was coming down, and it stopped coming down.

So is what's happening that these models are just so expensive for Anthropic to run that it's hitting a wall of its own? Again, I think you're making assumptions here. That's why I'm asking the CEO.

Yeah. The way I think about it is we think about the models in terms of how much value are they creating. So as the models get better and better, I think about how much value they create.

There's a separate question about how the value is distributed between those who make the model, those who make the chips, and those who make the underlying applications. Again, without being too specific, I think there are some assumptions in your question that are not necessarily correct.

Can you tell me which ones? I'll say this. I expect the price of providing a given level of intelligence to go down. I expect the price of providing the frontier of intelligence, which will provide increasing economic value, that might go up or it might go down. My guess is it probably stays about where it is.

But again, the value that's created goes way up. So two years from now, my guess is that we'll have models that cost the same order of magnitude that they cost today, except they'll be much more capable of doing work much more autonomously, much more broadly than they are capable of today.

One of the things that Amjad mentioned was he thinks that the bigger models are not as intensive to run or more intensive to run, given their size, because of the architecture and some of these techniques that we talked about, that they're lighting up only certain sections of the model.

So his idea, I hope I’m conveying this truthfully, is that Anthropic can run these models without too much bulk on the back end, but it's still keeping those prices where they are. I think the line that I draw there is maybe that to get to software margins, there were some reports that Anthropic is slightly below software gross margins.

You're going to have to charge a little bit more for these models. Yeah, again, I think larger models cost more to run than smaller models. I think the technique you're referring to is maybe a mixture of experts or something like that. So whether your models are a mixture of experts or not, mixture of experts is a way to run models more cheaply that have a given number of parameters.

It's a way to train models. But if you're not using that technique, then larger models that don't use that technique cost more to run than smaller models that don't use that technique. And if you're using that technique, then larger models that use that technique cost more to run than smaller models that are using that technique.

I think that's a distortion of the situation. Basically, I'm just guessing, and I'm trying to find out what the truth is from you. Yeah, look, in terms of the cost of the models, one thing you'd be surprised by is that people impute this thing to like, oh man, it's going to be really hard to get the margins from X percent to Y percent.

We make improvements all the time that make the models 50% more efficient than they are before. We are just at the beginning of optimizing inference.

Inference has improved a huge amount from where it was a couple of years ago to where it is now. That's why the prices are coming down. Then how long is it going to take to be profitable? I think the loss is going to be like $3 billion this year. That's what they say.

I would distinguish different things. There's the cost of running the model. For every dollar the model makes, it costs a certain amount that is actually already fairly profitable.

There are separate things. There's the cost of paying people and buildings that is actually not that large in the scheme of things. The big cost is the cost of training the next model.

I think this idea that the company is losing money and not being profitable is a little bit misleading, and you start to understand it better when you look at the scaling laws as a thought exercise. These numbers are not exact or even close for Anthropic, but let's imagine that in 2023 you train a model that costs $100 million.

In 2024 you deploy the 2023 model, and it makes $200 million in revenue, but you spend a billion dollars to train a new model in 2024. In 2025, the billion-dollar model makes $2 billion in revenue, and you spend $10 billion to train the next model.

So the company every year is unprofitable. It lost $800 million in 2024, and then in 2025 it lost $8 billion. This looks like a hugely unprofitable enterprise. But if instead I think in terms of whether each model is profitable, think of each model as a venture.

I invested $100 million in the model, and then I got $200 million out of the model the next year. So that model had 50% margins and made me $100 million. The next year the company invested a billion dollars and made $2 billion.

This general dynamic explains what is going on. At any time, if the models stopped getting better or if a company stopped investing in the next model, you would have probably a viable business with the existing models.

But everyone is investing in the next model. So eventually it'll get to some scale. But the fact that we're spending more, the fact that we're spending more to invest in the next model suggests that the scale of the business is going to be larger the next year than it was the year before.

Now, of course, what could happen is like the model stopped getting better, and there's this kind of one-time cost that's like a boondoggle, and we spend a bunch of money. But then the companies, the industry will return to this plateau, to this level of profitability, or the exponential can keep going.

I think that's a long-winded way to say I don't think it's the right way to think about things. Right, but what about open source? Because if you stopped, let's say you stopped investing in the models, and open source caught up, then people could swap in open source.

Now, I'd love to hear your perspective on this because one of the things people have talked to me about when it comes to the Anthropic business is that there is that risk eventually that open source gets good enough that you can take Anthropic out and put open source in.

People have. I think one of the things that's been true of this industry is that—and I saw it early in the history of AI—every community that AI has gone through has a set of heuristics about how things work. Back when I was in AI, back in 2014, there was an existing kind of AI and machine learning research community that thought about things in a certain way, and we were like this is just a fad, this is a new thing, this can't work, this can't scale.

Because of the exponential, all those things turned out to be false. Then a similar thing happened with like people deploying AI within companies to various applications.

Then there was the same thought in the startup ecosystem. I think now we're at the phase where the world's business leaders like the investors and the business, they have this whole lexicon of commoditization motes, which layer is the value going to accrue to and open source.

This idea that you can kind of see everything that's going on, that it has a significance, that it kind of undermines the fact that the idea that it undermines business. I actually find as someone who didn't come from that world at all, who never thought in terms of that lexicon, this is one of these situations where not knowing anything often leads you to make better predictions than the people who have their way of thinking about things from the last generation of tech.

This is all I think a long-winded way of saying I don't think open source works the same way in AI that it has worked in other areas. Primarily because with open source, you can see the source code of the model.

Here we can't see inside the model. It's often called open weights instead of open source to kind of distinguish that. Many of the benefits, which is that many people can work on it, that it's kind of additive, it doesn't quite work in the same way.

I've always seen it as a red herring when I see a new model come out. I don't care whether it's open source or not. If we talk about DeepSeek, I don't think it mattered that DeepSeek is open source. I ask, "Is it a good model? Is it better than us at the things that?"

That's the only thing that I care about. It actually doesn't matter either way because ultimately you have to host it on the cloud. The people who host it on a cloud do inference. These are big models. They're hard to do inference on.

Conversely, many of the things that you can do when you see the weights, we're increasingly offering on clouds where you can fine-tune the model. We're even looking at ways to investigate the activations of the model as part of an interpretability interface. We did some little things around steering last time, so I think it's the wrong access to think in those terms.

When I think about competition, I think about which models are good at the tasks that we do. I think open source is actually a red herring.

But if it's free and cheap to run it, it's not free. You have to run it on inference, and someone has to make it fast on inference.

All right, so I want to learn a little bit more about Dario, the person. Yes. So we have a little bit of time left.

So I have some questions for you about early life and then how you became who you are. Yes. What was it like growing up in San Francisco?

Yeah. The city when I first grew up here had not really gentrified that much. When I grew up, the tech boom hadn't happened yet. It happened as I was going through high school, and I actually had no interest in it.

It was totally boring to me. I was interested in being a scientist. I was interested in physics and math. The idea of writing some website actually had no interest to me whatsoever.

Or founding a company. Those weren't things that I was interested in at all. I was interested in discovering fundamental scientific truth, and I was interested in how can I do something that makes the world better.

That was more. I watched the tech boom happen around me, but I feel like there were all kinds of things I probably could have learned from it that would have been helpful now, but I just actually wasn't paying attention and had no interest in it, even though I was right at the center of it.

You're the son of a Jewish mother, Italian father. That is true. Where I'm from in Long Island, we call that a pizza bagel. A pizza bagel? I've never heard that term before.

What was your relationship with your parents like? Yeah, I was always pretty close with them. I feel like they gave me a sense of right and wrong and what was important in the world.

I feel like imbuing a strong sense of responsibility is maybe the thing that I remember most; they were always people who felt that sense of responsibility and wanted to make the world better.

I feel like that's one of the main things that I learned from them. It was always a very loving family, a very caring family. I was very close with my sister Daniela, who of course became my co-founder.

We decided very early that we wanted to work together in some capacity. I don't know if we imagined that it would happen at quite the scale that it has happened, but I think that was something we kind of decided early on.

The people that I've spoken with that have known you through the years have told me that your father's illness had a big impact on you. Can you share a little bit about that?

Yes, yes. He was ill for a long time and eventually died in 2006. That was actually one of the things that drove me to. I don't think we mentioned it yet in this interview, but before I went into AI, I went into biology.

I showed up at Princeton wanting to be a theoretical physicist. I did some work in cosmology for the first few months of my time there; that was around the time that my father died.

That did have an influence on me and kind of was one of the things that convinced me to go into biology to try and address human illnesses and biological problems. I started talking to some of the folks who worked on biophysics and computational neuroscience in the department I was in at Princeton.

That was what led to the switch to biology and computational neuroscience. After that, I eventually went into AI. The reason I went into AI was actually a continuation of that motivation, which is that as I spent many years in biology, I realized that the complexity of the underlying problems in biology felt like it was beyond human scale.

In order to understand it all, you needed hundreds, thousands of human researchers. They often had a hard time collaborating or sharing their knowledge. AI, which I was just starting to see discoveries in, felt to me like the only technology that could bridge that gap, that could bring us beyond human scale to fully understand and solve the problems of biology.

So, yeah, there is a through line there, right. I could have this wrong, but one thing I heard was that his illness was largely uncurable when he had it.

There have been advances that have been—can you share a little bit more? Yes, there are advances that have made it much more manageable today. Yes, yes, that is true.

Actually, only in the maybe three or four years after he died, the cure rate for the disease that he had went from 50% to roughly 95%. Yeah. It has to have felt so unjust to have your father taken away by something that could have been cured.

Of course. But it also tells you the urgency of solving the relevant problems. There was someone who worked on the cure to this disease that managed to cure it and save a bunch of people's lives, but could have saved even more people's lives if they had managed to find that cure a few years earlier than they did.

I think that's one of the tensions here, that I think AI has all of these benefits, and I want everyone to get those benefits as soon as possible. I probably understand better than almost anyone how urgent those benefits are.

And so I really understand the stakes when I speak out about AI having these risks, and I'm worried about these risks. I get very angry when people call me a doomer.

I got really angry when someone's like, this guy is a doomer. He wants to slow things down. You heard what I just said. My father died because of cures that could have happened a few years later.

I understand the benefit of this technology. When I sat down to write Machines of Loving Grace, I wrote out all the ways that billions of people's lives could be better with this technology.

Some of these people who on Twitter cheer for acceleration, I don't think they have a humanistic sense of the benefit of the technology. Their brains are just full of adrenaline, and they're like, they want to cheer for something.

They want to accelerate. I don't get the sense they care when these people call me a doomer. I think they just completely lack any moral credibility in doing that.

It really makes me lose respect for them. I've been wondering. This word impact has been. Because it's come up so often that those who have been around you have said you've been singularly obsessed with having impact.

In fact, I spoke with someone who knew you well, who said you wouldn't watch Game of Thrones because it wasn't tied to impact, that it was a waste of time and you wanted to be focused on impact.

Actually, that's not quite right. I wouldn't watch it because it was so negative. So people were playing such negative. It was like these people start off and they're partly the situation and partly because they're just horrible people.

They create the situation where at the end of it, everyone is worse off than everyone was before. I'm really excited about creating positive-sum situations. I recommend you watch it; it's a great show.

But I was just very reluctant and didn't watch it for a long time. Let's get back to the impact. Okay, let's get back to the impact.

So that's what impact is. Is effectively your career has been this quest to have that impact to be able to—tell me if I'm going too far—to prevent other people from being in similar situations.

I think that's a piece of it. I have looked at many attempts to help people, and some of them are more effective than others. I've always tried to ensure that there should be strategy behind it.

There should be brains behind trying to help people, which often means that there's a long path to it. It can run through a company and many activities that are technical and not immediately tied to the impact that you're trying to have.

But I’m always trying to bend the arc toward that. I think that's my picture of it. That's really why I got into this. I think similar to the reason to get into AI was that I saw the problems of biology as almost intractable without it, or at least too slow moving.

I think my reason to start a company was that I had worked at other companies; I just didn't feel like the way those companies were run was really oriented towards trying to have that impact.

There was a story around it that was often used for recruiting. But it became clear to me over the years that story was not sincere. I'm going to circle around a little bit because it's clear that you're referring to OpenAI here.

From what I understand, you had 50% of OpenAI's compute. I mean, you ran the GPT-3 project, so if anyone was going to be focused on impact and safety it wouldn't have been you.

Yes, there was a period during which that was true. That wasn't true the entire time. That was for example when we were scaling up GPT-3. When I was at OpenAI, I and a lot of my colleagues, including the people who eventually founded Anthropic— the Pandas.

That's the name you gave them. That isn't a name I gave them; the name they took. That isn't a name I ever used for my team. Okay, sorry, go ahead. That's a good clarification. Thank you.

Yeah, we were involved in scaling up these models. Actually, the original reason for building GPT-2 and GPT-3 was an outgrowth of the AI alignment work that we were doing where myself, Paul Christiano, and some of the Anthropic co-founders had invented this technique called RL from human feedback, which was designed to help steer models in a direction to follow human intent.

It was really a precursor to we were trying to scale up another method called scalable supervision, which I think is just starting to work many years later to help models follow more kind of scalable human intent. What we found is that even with the more primitive technique, RL from human feedback, it wasn't working with the small language models with GPT-1 that we applied it to.

The scaling up of GPT-2 and GPT-3 was done to study these techniques, in order to apply RL from human feedback at scale. This goes to the intertwining of the alignment of AI systems and the capability of AI systems. It's very hard to work on the safety of AI systems and the capability of AI systems separately.

It's very hard to work on one and not the other. I actually think the value and the way to inflect the field in a more positive way comes from organizational-level decisions, like when to release things, when to study things internally, what kind of work to do on systems.

That was one of the motivations for me and some of the others to be Anthropic founders, to kind of go off and do it our own way. But again, if you're driving the capabilities and you were the guy driving the cutting-edge models within OpenAI, if you left, you knew they were going to be a company that was still doing this stuff.

That's right. It seems like if you're driving the capabilities, you'd be the one in the driver's seat to help it be safe the way that you wanted.

Again, I will say if there's a decision on releasing a model, if there's a decision on the governance of the company, if there's a decision on how the personnel of the company works, how the company represents itself externally, the decisions that the company makes with respect to deployment, the claims it makes about how it operates with respect to society, many of those things are not things that you control just by training the model.

I think trust is really important. The leaders of a company have to be trustworthy people. They have to be people whose motivations are sincere. No matter how much you're driving the company forward technically, if you're working for someone whose motivations are not sincere, who's not an honest person, who does not truly want to make the world better, it's not going to work.

So then I'm sure you've heard the criticism from people like Jensen who say, well, Dario thinks he's the only one who can build this safely. And therefore, speaking of that word control wants to control the entire industry.

I've never said anything like that. That's an outrageous lie. That's the most outrageous lie I've ever heard, by the way. I'm sorry if I got Jensen's words wrong. No, no, no. The words were correct; he's saying, but are the words are outrageous.

In fact, I've said multiple times—and I think Anthropic's actions have shown it—that we're aiming for something we call a race to the top. I've said this on podcasts over the years, and I think Anthropic's actions have shown that with a race to the bottom, everyone is competing to get things out as fast as possible.

So I say when you have a race to the bottom, it doesn't matter who wins; everyone loses because you make the unsafe system that helps your adversary or causes economic problems or is unsafe from an alignment perspective. The way I think about the race to the top is that it doesn’t matter who wins; everyone wins.

So the way the race to the top works is you set an example for how the field works. You say, we're going to engage in this practice. So a key example of this is responsible scaling policies. We were the first to put out a responsible scaling policy.

We didn't say, everyone else should do this or you're bad guys. We didn't try to use it as an advantage. We put it out and then we encouraged everyone else to do it.

We discovered in the months afterward that there were people within the other companies who were trying to put out responsible scaling policies. But the fact that we had done it allowed, gave those people permission and enabled those people to make the argument to leadership, hey, Anthropic is doing this, so we should do it as well.

The same has been true of investing in interpretability. We release our interpretability research to everyone and allow other companies to copy it, even though we've seen that it sometimes has commercial advantages, same with things like constitutional AI, same with the measurement of the dangers of our system, dangerous capabilities evals.

So we're trying to set an example for the field, but there's an interplay where it helps to be a powerful commercial competitor. I've said nothing that resembles the idea that this company should be the only one to build the technology. I don't know how anyone could derive that from anything I've said. It's just an incredible and bad faith distortion.

All right, let's see if we can lightning round like one or two before I ask you the last one, which we'll have five minutes for. What happened with SBF? What happened with SBF? He was—go ahead. I couldn't tell you. What didn't you answer? I probably met the guy four or five times, so I have no great insight into the psychology of SBF or why he did things as stupid or immoral as he did.

I think the only thing I had ever seen ahead of time with SBF was a couple of people mentioned to me that he was hard to work with, that he was a bit of a move fast and break things guy. I was like, okay, there's plenty of people—welcome to Silicon Valley.

Yeah, welcome to Silicon Valley. So I remember saying, okay, I'm going to give this guy non-voting shares; I'm not going to put him on the board. He sounds like a bad person to deal with every day, but he's excited about AI, he's excited about AI safety, he's a bull on AI and he's interested in AI safety.

So seems like a sensible thing to do in retrospect. That move fast and break things turned out to be much, much, much more extreme and bad than I ever imagined.

Okay, let's end here. You found your impact; you're working the dream pretty much right now. I mean, think about all the ways that AI can be used for biology—just a start.

You also say that this is a dangerous technology, and I'm curious if your desire for impact could be pushing you to accelerate this technology while potentially devaluing the possibility that controlling it might not be feasible.

I think I have more than anyone else in the industry warned about the dangers of the technology. We just spent 10, 20 minutes talking about a large array of people who run trillion-dollar companies criticizing me for talking about the dangers of these technologies.

I have US government officials, I have people who run 4 trillion-dollar companies criticizing me for talking about the dangers of the technology, imputing all these bizarre motives that bear no relationship to anything I've ever said, not supported in anything I've ever done, and yet I'm going to continue to do it.

I actually think that as the revenues, as the economic business of AI ramps up—and it's ramping up exponentially—if I'm right, in a couple of years it'll be the biggest source of revenue in the world. It'll be the biggest industry in the world, and people who run companies already think it.

So we actually had this terrifying situation where hundreds of billions to trillions, I would say maybe $20 trillion of capital is on the side of accelerating AI as fast as possible. We have this company that's very valuable in absolute terms, but looks very small compared to that $60 trillion.

And I keep speaking up, even if it makes folks...there have been these articles. Some folks in the US government are upset at us, for example, for opposing the moratorium on AI regulation, for being in favor of export controls for chips on China, for talking about the economic impacts of AI.

Every time I do that, I get attacked by many of my peers. But you're still assuming that we can control it; that's what I'm pointing out. But I'm just telling you how much effort, how much persistence, how much despite everything that's stacked up, despite all the dangers, despite the risk that it has to the company of being willing to speak up, I'm willing to do it.

And that's why I'm saying that, look, if I thought that there was no way to control the technology, right? I thought, even if I thought this is just a gamble, right? Some people are like, oh, you think there's a 5 or 10% chance that AI could go wrong? You're just rolling the dice.

That's not the way I think about it. This is a multi-step game, right? You take one step, you build the next step of more powerful models, you have a more intensive testing regime. As we get closer and closer to the more powerful models, I'm speaking up more and more and I'm taking more and more drastic actions because I'm concerned that the risks of AI are getting closer and closer.

We're working to address them. We've made a certain amount of progress. But when I worry that the progress that we've made on the risks is not fully aligned with the speed of the technology, I speak up louder.

So you're asking why I am—you started this interview by saying what's gotten into you? Why are you talking about this? It's because the exponential is getting to the point that I worry that we may have a situation where our ability to handle the risk is not keeping up with the speed of the technology, and that's how I'm responding to it.

If I believed that there was no way to control the technology, which I see absolutely no evidence for that proposition, we've gotten better at controlling models with every model that we release.

All these things go wrong, but you really have to stress test the models pretty hard. That doesn't mean you can't have emergent bad behavior. I'd be very concerned if we got to much more powerful models with only the alignment techniques we have today.

Then I'd be out there saying everyone should stop building these things, even China should stop building these. I don't think they listen to me, which is one reason I think export controls is a better measure.

If we got a few years ahead in models and had only the alignment and steering techniques we had today, then I would definitely be advocating for us to slow down a lot. The reason I'm warning about the risks is so that we don't have to slow down, so that we can invest in safety techniques, and can continue the progress, continue the progress of the field.

It would be a huge economic effort even if one company was willing to slow down the technology. That doesn't stop all the other companies; that doesn't stop our geopolitical adversaries, to whom this is an existential fight, fight, fight for survival.

So very—there's very little latitude here. We're stuck between all the benefits of the technology, the race to accelerate it, and the fact that that is a multi-party race. And so I am doing the best thing I can do, which is to invest in safety technology to speed up the progress of safety.

I've written essays on the importance of interpretability, on how important various directions in safety are. We release all of our safety work openly because we think that's the thing that everyone needs to share.

If you have a better strategy for balancing the benefits, the inevitability of the technology and the risks that it faces, I am very open to hearing it because I go to sleep every night thinking about it because I have such an incredible understanding of the stakes in terms of the benefits, in terms of what it can do, the lives that it can save.

I've seen that personally. I also have seen the risks personally. We've already seen things go wrong with the models. We have an example of that with Grok.

People dismiss this, but they're not going to laugh anymore when the models are taking actions, when they're manufacturing, and when they're in charge of medical interventions. People can laugh at the risks when the models are just talking. But I think it's very serious.

I think what this situation demands is a very serious understanding of both the risks and the benefits. These are high-stakes. Decisions need to be made with seriousness.

Something that makes me very concerned is that on one hand, we have a cadre of people who are just doomers. People call me a doomer. I'm not. But there are doomers out there.

People who say they know there's no way to build this safely. I've looked at their arguments; they're a bunch of gobbledygook. The idea that these models have dangers associated with them, including dangers to humanity as a whole, that makes sense to me.

The idea that we can logically prove that there's no way to make them safe, that seems like nonsense to me. I think that is an intellectually and morally unserious way to respond to the situation we're in.

I also think it is intellectually and morally unserious for people who are sitting on $20 trillion of capital who all work together because their incentives are all in the same way. There are dollar signs in all of their eyes to sit there and say we shouldn't regulate this technology for 10 years.

Anyone who says that we should worry about the safety of these models is someone who just wants to control the technology themselves. That's an outrageous claim, and it's a morally unserious claim.

We've sat here and we've done every possible piece of research. We speak up when we believe it appropriate to do so. We've tried to back up when we make claims about the economic impact of AI.

We have an economic research council. We have an economic index that we use to track the model in real time, and we're giving grants for people to understand the economic impact, the economic impact of the technology.

I think for people who are far more financially invested in the success of the technology than I am to just breezily lob ad hominem attacks, I think that is just as intellectually and morally unserious as the doomer's position.

What we need here is we need more thoughtfulness. We need more honesty. We need more people willing to go against their interest, willing to not have breezy Twitter fights, hot takes.

We need people to actually invest in understanding the situation, actually do the work, actually put out the research, and actually add some light and insight to the situation that we're in. I am trying to do that.

I don't think I'm doing that perfectly, as no human can. I'm trying to do it as well as I can. It would be very helpful if there were others who would try to do the same thing.

Well, Dario, I said this off camera, but I want to make sure to say it on as we're wrapping up. I appreciate how much Anthropic publishes. We have learned a ton from the experiments, everything from teaming the models to vending machine Claude, which we didn't have a chance to speak about today, but I think the world is better off just to hear everything going on here.

And to that note, thank you for sitting down with me and spending so much time together. Thanks for having me.

Thanks, everybody, for listening and watching. And we'll see you next time on Big Technology Podcast.