TLDW logo

Anthropic CEO Dario Amodei: AI's Potential, OpenAI Rivalry, GenAI Business, Doomerism

By Alex Kantrowitz

Summary

## Key takeaways - **AI exponential growth is not slowing down**: Dario Amodei believes the exponential growth of AI capabilities, driven by increased compute, data, and training techniques, is continuing and sees no diminishing returns. He likens the current trajectory to the early days of the internet, where rapid advancements were not fully anticipated. [06:36], [09:41] - **Continual learning is a solvable problem**: While acknowledging the challenge of continual learning in LLMs, Amodei argues it's not a fundamental obstacle to AI's impact. He suggests that longer context windows and advancements in training techniques will fill many of the gaps, and that this problem may yield to scale and new thinking, similar to past perceived roadblocks. [12:15], [14:46] - **Talent density, not just capital, wins in AI**: Anthropic's CEO emphasizes that talent density is the key differentiator in the AI race, not just sheer capital. He believes Anthropic's ability to achieve significant growth and capital efficiency stems from its highly skilled team, allowing it to compete effectively even against trillion-dollar companies. [16:21], [21:35] - **Personal tragedy fuels AI urgency**: Dario Amodei's father died from a disease that was later made highly treatable. This personal experience, coupled with his background in biology and medicine, deeply informs his understanding of AI's potential to save lives and the urgency to develop beneficial applications while mitigating risks. [00:09], [45:35] - **Race to the top, not a race to control**: Amodei refutes claims that he wants to control the AI industry, stating Anthropic aims for a 'race to the top' by setting examples in responsible scaling, interpretability, and open research. He believes this approach benefits the entire field, rather than seeking a monopoly. [54:44], [55:35]

Topics Covered

  • AI's Rapid Advancement and Broad Economic Impact
  • AGI and Superintelligence are Meaningless Marketing Terms
  • Anthropic's Models Show Continuous Improvement in Coding
  • Why I'm not a 'doomer' on AI: Understanding the benefits and the stakes
  • AI Safety is a race we must win, not slow down, to avoid existential threats

Full Transcript

I get very angry when people call me a

doomer. When, you know, when when

someone's like, "This guy's a doomer. He

wants to slow things down." You you

heard what I just said, like, you know,

my my father died because of, you know,

cures that, you know, could have could

have happened a few years later. I

understand the benefit of this

technology. I'm sure you've heard the

criticism from people like Jensen who

say, "Well, Daario thinks he's the only

one who can build this safely and

therefore wants to control the entire

industry."

I've never said I've never said anything

like that. That's an outrageous lie.

That's the most outrageous lie I've ever

heard.

Anthropic CEO Dario Ammoday joins us to

talk about the path forward for

artificial intelligence. Whether

generative AI is a good business and to

fire back at those who call him a

doomer. And he's here with us in studio

at Anthropic headquarters in San

Francisco. Dario, it's great to see you

again. Welcome to the show.

Thank you for having me.

So, let's recap the past couple months

for you. Uh, you said AI could wipe out

half of entry-level white collar jobs.

You cut off Windsurf's access to

Anthropic's top tier models when you

learned that OpenAI was going to acquire

them. You asked the government for

export controls and annoyed Nvidia CEO

Jensen Wong. Uh, what's gotten into you?

Um, you know, I think I think Anthropic,

myself and Anthropic are always focused

on kind of trying to do and say the

things that we believe. Um uh and I

think as we've gotten uh more close to

AI systems that are more powerful um you

know I think I've wanted to say those

things um you know more uh more

forcefully more publicly to make the

point clearer. Um you know I've been

saying for many years that you know we

have these we can talk in detail about

them but you know we have these scaling

laws. AI systems are getting more

powerful. They're going from the level

of, you know, a a few years ago they

were barely coherent now, you know, a

couple years ago they were at the level

of a smart high school student. Now

we're getting to smart college student,

PhD, and they're starting to they're

starting to apply across the economy.

And so I think all the issues related to

AI ranging from kind of the national

security issues to the economic issues

um you know are are are starting to

become quite near to where to where we

um you know to where we're actually

going to face them. And so and so I

think as these problems have come closer

I've you know even though you know in

some form anthropic has been saying

these things for a while I think the

urgency of these things has gone up and

and and and and you know I I I want to

make sure that we uh you know I want to

make sure that we say what we believe

and that we warn the world about the

possible downsides even though you know

no one can say what's going to happen.

we're, you know, we're saying what we,

you know, what we think might happen,

what we think is likely to happen. You

know, we we back it up as as best we

can, although it's often, you know,

extrapolations about the future where

where no one where no one can be sure.

Um, but, you know, I think we we see

ourselves as kind of as kind of having a

duty to, you know, to kind of warn the

world about what's going to happen. And

that's not to say, you know, I think

there's an incredible number of like

positive applications of AI, right? I've

I've kind of continued to talk about

that. I read this I wrote this essay,

Machines of Loving Grace. Um I I I feel

in fact that I and anthropic have often

been able to do a better job of

articulating the benefits of AI than

some of the people who call themselves

optimists or accelerationists. Um so I

think we probably appreciate the

benefits more than more than anyone. Um,

but for exactly the same reason, because

we can have such a good world if we get

everything right, I feel obligated to

warn about the risks.

So all of this is coming from your

timeline. Basically, you it seems like

you have a shorter timeline than most

and so you were feeling a sense of

urgency to get out there because you

think that this is imminent.

Yes, I'm not sure. Um, you know, I think

it's very hard to predict particularly

on the societal side. So if you say you

know when are people going to deploy AI

or when are companies going to use you

know XX X dollars of spend of AI or you

know when will when will AI um you know

be be be used in these applications or

when will it drive these medical cures

that's kind of harder harder to say I

think the underlying technology is more

predictable but still uncertain still no

one knows but I think on the underlying

technology I've started to become more

confident. There isn't no uncertainty

about it. You know, I think the the

exponential that we're on could could

kind of still um you know, could totally

peter out. You know, I think there's

maybe uh I don't know 20 or 25% chance

that sometime in the next 2 years the

models just start getting stop getting

better for reasons we don't understand

or maybe reasons we do understand like

you know data or compute availability

and then everything I'm saying just just

seems seems totally silly and everyone

makes fun of me for all the warnings

I've made and and you know I'm just I'm

just totally fine with that given given

the distribution that that that given

the distribution that I see

and so I should say that this is part

our conversation as part of a profile

I'm writing about you. I've spoken with

more than two dozen people who've worked

with you, who know you, who've competed

with you, and I'm going to link that in

the show notes if anybody wants to read

it. It's free to read. Uh, but one of

the themes that has uh come through

across everybody I've spoken with is

that you have about the shortest

timeline of any of the major lab leaders

and you just referenced it uh just just

now. So, why do you have such a short

timeline and why should we believe in

yours? Yeah, it it really depends what

you mean by timeline. Um, so one thing

and you know I've I've I've been

consistent on this over the years is you

know there are these terms in the AI

world like AGI and super intelligence.

Like you'll hear leaders of companies

say we've achieved AGI, we're moving on

to super intelligence or like it's

really exciting that someone stopped

working on AGI and started working on

super intelligence. So I think these

terms are totally meaningless. I don't

know what AGI is. I don't know what

super intelligence is. It it sounds like

a it sounds like a marketing term. Yeah,

it sounds like, you know, something

something designed to activate people's

people's dopamine. So, you'll see in

public I never use those terms and uh I

um you know, I'm I'm actually, you know,

careful to criticize the use of those

terms. Um but I think I think despite

that, I am I am indeed one of the most

bullish about about AI capabilities

improving very fast. The thing I think

is real that I've said over and over

again is the exponential. The idea that

every few months we get an AI model that

is better than the AI model we got

before. And that we we get that by

investing more compute in AI models,

more data, more new types of training

models. Initially, this was done by

what's called pre-training, which is

when you just feed a bunch of data from

the internet into the model. Now we have

a second stage that's reinforcement

learning or test time compute or

reasoning or whatever you want to call

it. I I think of it as a second stage

that involves reinforcement learning.

Now both of those things are scaling up

together as we've seen with our models

and as we've seen with models from other

from other companies and I don't see

anything blocking that the further

scaling of that. There's some stuff

about you know how do we broaden the

tasks on the RL side of it. we've seen

uh more progress on say math and code

where where the models are you know

getting pretty close to like a high

professional level and less on more

subjective tasks but I think that is

very much a temporary obstacle um uh so

when I look at it I see this exponential

and I say look people aren't very good

at making sense of exponentials right

like you know if something is doubling

every six months then uh you know two

years before it happens it looks like

it's only 1/16th of the way there and

and and so we are sitting here in the

middle of of 2025

um and the models are really starting to

explode in in in terms of the economy

right if you look at the capabilities of

the model they're starting to saturate

all the benchmarks if you look at

revenue and you know anthropics revenue

every year has grown 10x um uh every

year we're we're kind of conservative

and we say you know it can't grow 10x

this time you know I I I you know I

never I never assume anything and and

actually always am very conservative in

saying ah I think it's going to slow

down on the business side but we went

from zero to 100 million in 2023 we went

from 100 million to a billion in 2024

and you know this year in this first

half of the year we've gone from 1

billion to you know I think as of as of

speaking today it's it's well above four

it might be 4.5 um uh and so if you

think about it you know suppose that

exponential continued for two years I'm

not saying it will but suppose it

continued for two years you know you're

you're like well into the hundred

billions. I'm not saying that'll happen.

I'm saying the situation is that when

you're on an exponential, you can really

get fooled by it. Two years away from

when the exponential goes totally crazy.

It it, you know, it looks like it's just

starting to be a thing. Um, and so

that's the fundamental dynamic. You

know, we saw that with the internet in

we saw that with the internet in the

'9s, right? where it was like you know

networking speeds and the underlying

speed of the computers were getting fast

and over a few years it became possible

to have to basically build a digital

global communications network on top of

all this when it wasn't possible just a

few years ago and and almost no one

except for a few people really saw the

implications of that and how fast it

would happen and so that's that's where

I'm coming from that that's what I think

now I don't know like if a bunch of

satellites crashed maybe the internet

would have taken longer. If there was an

economic crash, maybe it would have

taken a little longer. So, we can't be

sure of the exact timelines, but I think

people are getting fooled by the

exponential and and not realizing how

fast it it might be. How fast I think it

probably will, although I'm not sure.

But so many folks in the AI industry are

talking about diminishing returns from

scaling. Now, that really doesn't fit

with the vision you just laid out. Are

they wrong?

Yeah. uh I I from what we've seen I can

only speak in terms of the models at

enthropic um but what I think seen in

terms of the models at enthropic if we

look at you know let's take coding

coding is one area where you know I

think anthropic models have advanced

very quickly adoption has been very

quick we're not just a coding company

we're planning to expand to many areas

but if you if you look at if you look at

coding um you know every you know we

release 3.5 sonnet a model we call 3.5

sonnet V2 um uh which I you know let's

call it 3.6 Sonnet now um 3.7 sonnet uh

and then 4.0 sonnet and 4.0 obus and you

know that series of four or five models

each one got substantially better at

coding than than the last. If you want

to look at benchmarks, you can look at,

you know, SweetBench growing from uh,

you know, I think 18 months ago was at

like 3% or something um, growing all the

way to, you know, 72 to 80% depending on

how you how you measure it and and the

real usage has grown grown exponentially

as well where we're heading more and

more towards autonomously you can just

use these models. I think the actual

majority of code um uh uh that's written

written at Enthropic uh is you know at

this at this point uh written by or at

least with the involvement of one you

know one of the clawed models um and

various uh various other companies have

have said you know have said you know

similar similar statements to uh similar

statements to that. So we see the

progress as being very fast and the

exponential is continuing and and we

don't see any diminishing returns.

But there are some liabilities it seems

like with large language models. For

instance, continual learning. We had

Dark Kesh on a couple weeks ago. Here's

how he put it and he wrote about it in

his Substack. The lack of continual

learning is a huge huge problems. The LM

baseline at many tasks might be higher

than an average human, but you're stuck

with the abilities you get out of the

box. So you just make the model and

that's it. It doesn't learn. that seems

like a glaring liability. What do you

think about that?

So, first of all, I would say even if we

never solved continual learning, um even

if we never solve continual learning and

memory, um I think that the potential

for the LLMs to do, you know, incredibly

uh incredibly well to, you know, affect

things at the scale of the economy will

be very high. Right? If I think of the

field I used to be in biology and

medicine like you know let's say I had a

very smart Nobel prize winner and you

know I I I said okay you know you're you

know you've you you've discovered all

these things you have this incredibly

smart mind but you know you can't you

can't you know you can't like read new

textbooks or absorb any new information.

I mean that would be difficult but like

still if you had like 10 million of

those like they're they're still going

to make a lot of biology breakthroughs.

Like they're going to be limited.

they're going to be able to do some

things humans can't and there are some

things humans can do that they can't.

But but but even that even if we impose

that as a ceiling like man that's pretty

damned impressive and transformative and

even if I said you never solve that like

I think you know I think people are

underestimating the impact but but look

context windows are getting longer and

models actually do learn during the

context window right so so as as I you

know talk to the model during the

context window I have a conversation it

absorbs information the underlying

weights of the model may not change but

but you know just just like I'm talking

to you here and we're having a

conversation and I listen to the things

you say and I I you know I think and and

I like respond to them. The models are

able to do that and and from a from an

from a machine learning perspective,

from an AI perspective, there's no

reason we can't make the context length

a 100red million words today, right?

Which is roughly what a human hears what

what a human hears in their lifetime. Um

uh there's no reason that we can't uh do

that. It's it's really inference

support. Um uh and so again even that

fills in many of the gaps not all the

gaps but it fills in many of the gaps

and then there are a number of things

like learning and memory that do allow

us to update the weights. So um you know

there there there are a number of things

around you know types of reinforcement

learning learning training you know you

know we used to many years ago talk

about inner loops and outer loops right

the inner loop is like I have some

episode and I learn some things in that

episode and I'm trying to optimize for

the lifetime of that episode and and

kind of the outer loop is is is the

agents learning over episodes. Um and so

I think maybe that inner loop outer loop

structure is a way to learn the

continual learning. One thing we learned

in AI is whenever it feels like there's

some fundamental obstacle like two years

ago we thought there was this

fundamental obstacle around uh around

reasoning turned out just to be just to

be RL you just train with RL and you let

the model write some stuff down you know

you you let the model write things down

to try and figure out objective math

problems um without being too specific

um you know I think and we already have

maybe some some you know some some

evidence to suggest that this is another

of those problems that is is not as

difficult as it seems that will fall to

scale plus a slightly different way of

thinking about things. Do you think your

obsession with scale might blind you to

some of the new techniques like Deis

Sabis says, you know, to get to AGI or

you might call it super powerful AGI,

whatever human level intelligence is

what we're all talking about. We might

need a couple new techniques for that to

happen. If you're

developing new techniques, we're

developing new techniques every day.

Okay.

Um uh you know, Claude is very good at

code and we you know, we don't really

talk externally that much about why why

Claude is so good at code.

Why is it so good at code? Um it

like I said we don't talk externally

about it.

Um uh uh uh so you know every new

version of claude that we make has you

know improvements to the architecture,

improvements to the data that we put

into it, improvements to the methods

that we use to train it. So we're

developing new techniques all the time.

Um new techniques are a part of every

you know model that we build. And you

know that's why we you know I've said

these things about like you know we're

trying to optimize for like talent

density as much as possible like you

need that talent density in order in

order to invent the new techniques. You

know, there's one thing that's been

hanging over this conversation, uh,

which is that maybe Anthropic is the

company with the right idea, uh, but the

wrong resources. Because you look at

what's happening with XAI and, um, and

inside Meta where, uh, Elon's built his

massive cluster, Mark Zuckerberg is

building this 5 gawatt data center and

they are putting so much resources

towards scaling up. Um, is it possible?

I mean anthropic obviously you have

raised billions of dollars but these are

trillion dollar companies.

Yeah. So we've raised I think at this

point a little short of $20 billion.

It's not bad.

So that's that's not that's not nothing.

And I would also say if you look at the

size of the data centers that we're

building with for example Amazon. Um I I

don't think our data center scaling is

substantially smaller than that of any

of the other companies in the space. Um

you know in many cases these things are

limited by energy. they're limited by

capitalization. Um, you know, when when

when when when people talk about, you

know, these large amounts of money,

they're they're talking about it over

over over several years, right? And when

you hear some of some of these

announcements, sometimes they're not

funded yet. They're, you know, we've

we've seen the size of the of the of the

data centers that folks are building and

we're we're actually we're actually

pretty confident that, you know, we're

we will be within we will be within a

rough range of the size of data centers

they build. You talked about talent

density. What do you think about what

Mark Zuckerberg is doing on the talent

density front? I mean, combining that

with these massive data centers, it

seems like he's going to be able to

compete.

Yeah. So, uh this is this is actually

very interesting because um you know,

one thing we noticed is that relative to

other uh companies um you know, I think

I think I think uh very you know, a lot

fewer people from Enthropic uh have been

have been caught by these. And it's not

for lack of trying. I've talked to

plenty of people uh you know who who got

these offers at Enthropic and and who

just turned them down. um who wouldn't

even talk to Mark Zuckerberg who said um

you know uh uh no I'm I'm I'm staying at

Anthropic and and our our general

response to this was you know I posted

something to the to the whole company

Slack where where where I said look um

you know we are not willing to

compromise our you know our compensation

principles our principles of fairness to

respond individually to these offers.

The way things work at Enthropic is

there's a series of levels. One

candidate comes in, they get assigned a

level, and we don't negotiate that

level. Um, uh, uh, because because we

think it's unfair. We want to have a

systematic way. If, you know, if Mark

Zuckerberg, you know, throws a dart at a

dart board and hits your name, that

doesn't mean that you should be paid 10

times more than the guy next to you

who's who's, you know, who's who's just

as skilled, who's just as talented. Um

uh and and and and my view of the

situation is that you know the only way

you can really be hurt by this is if you

allow it to destroy the culture of your

company by panicking by treating people

unfairly in an attempt to to defend the

company. Um and I think actually this

was a unifying moment for the company

where um you know we we we didn't give

in. we refused to compromise our

principles because we had the confidence

that people are enthropic because they

truly believe in the mission. Um uh and

and you know I think um I I that that

gets to kind of how I see this. I think

that what they are doing is trying to

buy something that cannot be bought. Um

and and that is alignment with the

mission. Um uh and you know I I you know

I you know I think there are selection

effects here like I you know I you know

are they getting the people who are most

enthusiastic who are most missional

aligned who are most excited to

but they have talent and GPUs you're not

underestimating them.

I I

we'll see how it we'll see how it plays

out. Um I am pretty bearish on what

they're trying to do.

So let's talk a little bit about your

business because a lot of people have

been wondering is the business of

generative AI a real thing? And I'm also

curious. I have questions all the time.

You talked about how much money you've

raised, close to 20 billion. Um, you've

raised three billion, three billion from

Google, 8 billion from Amazon, three and

a half billion uh from a new round led

by Lightseed who I've spoken with. Um,

what what is your pitch? Because you are

not part of like a big tech company.

You're out there on your own. Do you

just bring the scaling laws and say,

"Can I have some money?" So my my view

of this has my view of this has always

been um that uh talent is the most

important thing. Um so you know if you

go back three years ago um you know we

were in a position where we had raised

mere hundreds of hundreds of millions.

Open AAI had already raised, you know,

13 billion from from Microsoft. And of

course, you know, the the large hyper

cap tech companies were sitting on 100

billion, $200 billion. And and basically

the pitch we made then is we know how to

make these models better than others do,

right? There may be a curve. There may

be a curve of scaling laws. But but

look, if we are in a position where we

can do for a hundred million what others

can do for a billion and we could do for

10 billion what they can do for a

hundred billion then it's 10 times more

capital efficient to invest in entropic

than it is to invest in in these other

companies. Would you rather be in a

position where you can do anything for

10 times cheaper or where you start with

a large pile of money? If you can do

things 10 times cheaper, the the you

know the the the the the money is a

temporary defect that that that that you

can remedy. If you have this intrinsic

ability to build things for the same

price, much better than anyone else or

as good as anyone else for much lower

price. You know, investors aren't aren't

idiots or at least they aren't always

idiots. Um

depends which one you go to.

I'm not going to name any names. um uh

uh uh but um uh uh you know they they

basically understand the concept of

capital efficiency. Um and so we've been

in a position you know 3 years ago where

you know these differences were like a

thousandx and now you're saying you know

with $20 billion can you compete can you

compete with hundred billion um and and

and my answer is basically yes because

of the talent density. You know, I've

said this before, but uh you know, the

the the Anthropic is actually the

fastest growing software company in

history at the scale that it's at. Um so

we grew from zero to 100 million in

2023, 100 million to billion in 2024.

And this year we've we've grown from 1

billion to I think I said this before

4.5. So that that 10x a year I mean you

know every every you know every year I

like suspect that we'll grow at that

scale and and every year I'm almost

afraid to say it publicly because I'm

like no it couldn't it couldn't possibly

happen again. So you know I think I

think the growth at that scale like kind

of speaks for itself in terms of our

ability to compete with the big players.

Okay. So CNBC says 60 to 75% of

anthropic sales are through the API.

That was according to internal

documents. Is that still accurate? Um I

won't give exact numbers but the

majority does come through the API

although we also do have a flourishing

apps business and you know I think more

recently um the you know max tier which

power users use as well as claude code

which which coders use. So you know I

think we have a we have a thriving and

fast growing apps business but yes the

majority comes through the API. So

you're making the most pure bet on this

technology like you know OpenAI might be

betting on chat GPT and Google might be

betting on the fact that no matter where

the technology goes it can you know

integrate into Gmail and calendar. So

why have you made this bet on this the

pure bet on the tech itself?

Yeah I mean I would say I wouldn't quite

put it that way. I think we've I would

describe it more as we've bet on

business use cases of the model um more

so than we've bet on the API per se. And

it's just that the first business use

cases of the model come through the API.

So you know as you mentioned OpenAI is

very focused on the consumer side.

Google is very focused on kind of the

existing products that Google has. Our

view is that if anything the enterprise

use of AI is going to be greater even

than the consumer use of AI or I should

say the business use because it's

enterprise, it's startups, it's

developers and it's kind of you know

power users using the model model for

productivity. Um I I also think that uh

being a company that's focused on the

business use cases actually gives us

better incentives to make the models

better. Um a a thought a thought

experiment that I think is worth running

is you know suppose I have this model

and it's uh it's um you know it's it's

as good as an undergrad at biochemistry.

Um and then I improve it and it's as

good as a PhD student at at

biochemistry. If I go to a consumer,

right, if I give them the chatbot and I

say, "Great news. I've improved the

model from, you know, undergrad to

graduate level in in biochemistry, um,

you know, maybe I don't know 1% of

consumers care about that at all, right?

99% are just going to be like, I don't

understand it either way." Um, but now

suppose I go to Fizer and and I say, you

know, I've improved this from

undergraduate biochemistry to to

graduate biochemistry. Like, this is

going to be the biggest deal in the

world, right? They might pay 10 times

more for something like that. it might

have 10 times more value to them. And so

the general aim of making the models

solve the problems of the world to make

them smarter and smarter but also able

to to bring many of the positive

applications right the things I wrote

about in like machines of loving grace

of like solving the problems of biio

medicine solving the problems of

geopolitics solving the problems of you

know of economic development um you know

as well as more prosaic things like

finance or legal or productivity or

insurance. Um I think it gives a better

incentive to kind of develop the models

um as as far as uh as as as as far as

possible and I think in many ways it's

like it may even be a more a more a more

positive business. So I would say we're

making a bet on the business use of AI

because it's most aligned with kind of

the exponential.

Okay. Then briefly, how did you decide

to go with the coding use case?

Yes. So um you know uh originally as

happens with most things we're trying to

optimize for making the model better at

a bunch of stuff and you know coding

particularly stood out in in terms of

how how valuable it was. You know I've

worked with thousands of engineers and

there was a point about a year year and

a half ago where one of the best I'd

ever worked with said um uh you know

every previous coding model has been

useless to me and and this one finally

was able to do something I wasn't able

to do. And then after we released it, it

started getting quick adoption. This was

around the time that you know a lot of

the coding companies like Cursor,

Windsurf, GitHub, Augment Code started

started exploding in in popularity and

then when we saw how popular it was, we

kind of doubled doubled down on it. Um

my view is that coding is particularly

interesting because a the adoption is

fast um and and b getting better at

coding with the models actually helps

you to develop the next model. Um, so it

has it has a number of uh you know you

know I would say advantages

and now you're selling uh your AI coding

through claude code. Um but it's very

interesting the pricing model um has

been confounding to some. You can spend

$200 a month and get the equivalent I

spoke to one developer they got the

equivalent of $6,000 a month uh from

your API. Um, Ed Zitetron has pointed

out the more popular that your models

get, the more money you're going to lose

if people are super users of this

technology. So, how does that make

sense? So, um, uh, uh, so actually

pricing schemes and rate limits are

surprisingly complicated. Um so so some

of this is basically the result of when

we released our um uh when we when we

released claude code and the max tier

which we eventually tied together

actually not fully understanding the

implications of you know the ways in

which people could use the models and

how much they were actually able to get.

So over the last few days, as of the

time of this uh as of the time of this

interview, we've adjusted that

particularly on the larger models like

Opus, I think it's no longer possible to

spend that much um uh uh with a with a

with a $200 uh uh subscription. And you

know, it's possible more changes will

more changes will come in the future,

but we're always going to have a

distribution of users who use a lot and

and and users who lose who use some

amount. And it it doesn't necessarily

mean we're losing money that there are

some some users who get more um uh you

know who if you were to measure via API

credits spend you know get get a better

deal on the consumer on on the consumer

subscription than they would on the API

products. Right? There's there's a lot

of assumptions there. Um and I can tell

you that that some that some of them

that some of them are wrong. Um uh we

are not in fact uh losing money. But I

guess there's another question about

whether you can continue to serve these

use cases um and not raise prices. So uh

just to give you a couple stats, there

are some developers that are upset

because uh using Anthropic's newer

models in cursor is costing them more

than it ever has. U startups that I've

spoken with say Anthropic is down a much

down a bunch because they can't get

access to the GPUs. At least that's what

they um what they imagine is happening.

And I was just with Amjad Masad at

Replet um in an interview that we're

going to air next week who said there

was a period of time where the price per

token price per to use these models was

coming down and it stopped coming down.

So is it the is what's happening that

these models are just so expensive for

anthropic to run that it's hitting it's

a wall of its own

again I think you're I think you're

making assumptions here.

That's why I'm asking the CEO. Uh yeah

um uh you know uh you know the the the

way I think about it is we think about

the models in terms of how much value

are they creating right um so as the

models get better and better I think

about how much value they create and

there's a separate question about how

the value is distributed between those

who make the model those who make the

chips and and you know those who make

those who make the the the um the the

the the underlying applications. So

again, without being too specific, like

I think there are some assumptions in

your question that are not necessarily

correct. Um uh you know, I

tell me which ones

I I so I'll say this. I do expect the um

I I I expect the price of providing a

given level of intelligence to go down.

Um I expect the price of providing the

frontier of intelligence which will

which will provide kind of increasing

economic value that might go up or it

might go down. My guess is it probably

stays about where it is. But again the

value that's created you know goes way

up. So two two years from now my guess

is that we'll have models that cost of

the same order of magnitude that they

cost today except they'll be much more

capable of doing work much more

autonomously much more broadly than they

are capable of today. One of the things

that Amjad mentioned was he thinks that

the bigger models are not as intensive

to run or more intensive to run given

their size because of the architecture

and some of these techniques that we

talked about that they're lighting up

only certain sections of the model. So

his idea I'm hopefully conveying this

truthfully is that um Anthropic can run

these models without too much bulk on

the back end but is still keeping those

prices uh where they are. And I think

the the line that I'm going to draw

there is maybe that um to get to

software margins, there were some

reports that Anthropic is slightly below

software gross margins. You're going to

have to charge a little bit more for

these models.

So, um yeah, again, um I think larger

models cost more to run than smaller

models. Um uh uh you know, I think the

technique you're referring to is maybe

mixture of experts or something like

that. So whether your models are mixture

of experts or not, like mixture of

experts is is like a way to run models

more cheaply that have a given number of

parameters. It's a way to train models.

But if you're not using that technique,

then larger models that don't use that

technique cost more to run than smaller

models that don't use that technique.

And if you're using that technique,

larger models that use that technique

cost, you know, more to run than than

smaller models that that that are using

that technique. So I think I think

that's sort of a distortion of the I

think that's sort of a distortion of the

situation. Um

basically I'm I'm just guessing and I'm

trying to find out what the truth is

from you. So

yeah look so I um you know in in in in

terms of like the cost of the models

like one thing you'd be surprised by

people you know people kind of impute

this thing to like oh man it's going to

be really hard to get the margins from

like x% to y%. We make improvements all

the time that make the models like 50%

more efficient than they are before. We

are just the beginning of optimizing

inference. Um uh inference has improved

a huge amount where from from where it

was a couple a couple years ago to where

it is now. That's why the prices are

coming down.

And then how long is it going to take to

be profitable? Because I think the loss

is going to be like three billion this

year. That's what they would distinguish

different things. Okay.

Um there's the cost of running there's

the cost of running the model, right? So

so for every dollar the model makes it

costs a certain amount. Um that is

actually already fairly profitable. Um

there are separate things. There's you

know the cost of paying people um and

like buildings that is actually not that

that large in the scheme of things. The

big cost is the cost of training the

next model. Um, and I think this idea of

like the companies losing money and not

being profitable is it's a little bit

misleading. Um, and you start to

understand it better when you when you

look at the scaling laws. So, as a

thought exercise, these numbers are not

exact or even close for entropic. Let's

imagine that in 2023 you train a model

that costs $und00 million. Um, and then

in 2024, you deploy the 2023 model and

it makes $200 million in revenue, but

you spend a billion dollars to train um,

you know, to train a new model in 2024.

So, and then, you know, and then in

2025, the billion dollar model makes two

billion in revenue and you spend 10

billion to train the next model. Um, so

the company every year is unprofitable.

It lost 800 million in 2024 and then 20

2025 it lost$8 billion. Um so you know

this looks like a hugely unprofitable

enterprise but if instead I think in

terms of is each model profitable right

think of each model as a venture. Um I

invested 100 million in the model and

then I got then I got then I got 200

million out of the model the next year.

So that model had 50% margins and and

you know and like and like made me

hundred $100 million the next year. um

you know the the the company invested a

billion dollars and made and made $2

billion in in or sorry the next model

the company invested a billion dollar so

every model is profitable but the

company is unprofitable every year I'm

not I'm not this is this is this is a a

styliz I'm not like claiming these

numbers for anthropic or claiming these

facts for but this general philanthropic

this general dynamic is is this this

general dynamic is in general terms the

explanation for what is going on. And so

at you know you know at at any time if

the models stopped getting better or if

a company stopped investing in the next

model um you know you would you know you

would have probably a viable business

with the existing models but everyone is

investing in the next model and so

eventually it'll get it'll get to some

scale. the but the fact that we're

spending more to to to to this fact that

we're spending more to invest in in the

next model suggests that the the scale

of the business is going to be larger

the next year than it was than it was

the year before. Now, of course, what

could happen is like the models stop

getting better and there's this kind of

one-time cost that's like a boondoggle

and we spend we spend a bunch of money,

but then the the you know the the

companies the industry will kind of

return to this you know to this plateau

to this level of profitability or the

exponential can keep going. Um, so I

think I think that's a long-winded way

to say I don't think it's really the

right way to think about things,

right? But what about open source?

Because if you stopped, let's say you

stopped investing in the models and open

source caught up, then people could swap

in open source. Now, I'd love to hear

your perspective on this because one of

the things people have talked to me

about when it comes to the anthropic

business is there is that risk

eventually that open source gets good

enough that you can take anthropic out

and put open source in.

Yeah. So, you know, people have I you

know, I think one of the things that's

been true of this industry is that and

you know, I saw it early in I saw it

early in the history of AI. Every

community that that AI has gone through,

it has this set of heruristics about how

things work. like back when I was in you

know AI back in 2014 there was an

existing kind of AI and machine learning

research community that like thought

about things in a certain way and we're

like this is just a fad this is a new

thing this can't work this can't scale

and then because of the exponential all

those things turned out to be false then

a similar thing happened with kind of

like people deploying AI within

companies to various applications then

there was the same thought in the

startup ecosystem and I think now we're

at the phase days where kind of the

world's business leaders like the

investors and the business they have

this whole lexicon of commoditization

um you know modes which layer is the

value going to going to which layer is

the value going to acrew to and open

source is this idea that you can kind of

see everything that's going on you know

that that it has a significance that it

kind of undermines

um uh uh the the fact that it you know

the idea that it undermines business and

I actually find as someone who didn't

come from that world at all, who never

thought in terms of that lexicon. This

is one of these situations where not

knowing anything often leads you to make

better predictions than kind of the

people who have their way of thinking

about things from the last generation of

tech. Um, and I, you know, this is all,

I think, a long-winded way of saying I I

don't think open source works the same

way in in AI that it has worked in other

areas. primarily because with open

source you can you can see the you know

you can see the source code of the model

here we can't see inside the model um

you know it's often called open weights

instead of open source to kind of

distinguish that but a lot of the

benefits which is that many people can

work on it that it's kind of additive it

doesn't quite work in the same way um so

you know I've I've actually always seen

it as a red herring when I see it when I

see a new model come out I don't care

whether it's open source or not like if

we talk about deepseeek I don't think it

mattered that Deep Seek is open source.

I think I ask is it a good model? Is it

better than us at at you know the things

that that's the only thing that I care

about it. It actually it actually

doesn't doesn't matter either way. Um

because ultimately you have to you have

to host it on the cloud. The people who

host it on the cloud do inference. These

are big models. They're hard to do

inference on. And conversely, many of

the things that you can do when you see

the weights um uh uh you know, we're

increasingly offering on clouds where

you can fine-tune the model. You can you

know um you know we're even looking at

at ways to you know to to kind of you

know investigate the activations of the

model as part of like an

interpretability interface. We did some

little things around steering last time.

Um so I think it's the wrong axis to

think in terms of when I think about

competition I think about like which

models are good at the task that we do.

Um I think open source is actually a red

herring.

But if it's free and cheap to run

it's not free. You have to you have to

you have to you have to run it on

inference and someone someone has to

make it fast on inference.

All right. So I want to learn a little

bit more about Daario the person. Yes.

So we have a little bit of time left. Um

so I have some questions for you about

early life and then how you became who

you are. Yes.

So, what was it like growing up in San

Francisco?

Yeah. Um I, you know, the city when I

first grew up here had not really had

not really gentrified uh uh that much.

You know, when I grew up, the tech boom

hadn't hadn't happened, uh hadn't

happened yet. Um you know, it happened

as as I was going through high school.

And actually, I had no interest in it.

Um it was totally it was totally boring

to me. Um you know, I was interested in

being like a scientist. I was interested

in physics and math and you know the

idea of like you know you know like

writing some website actually had no

interest to me to me whatsoever or like

founding a company like those weren't

things that I was uh that I was

interested in at all. Um you know I was

interested in discovering fundamental

scientific truth and I was interested in

like you know how can I how can I do

something that like makes the world

better. Um uh so so you know that was

that was kind of more and you know I

watched the tech boom happen around me

but I I feel like you know there was all

kinds of things I probably could have

learned from it that would have been

helpful now but I just actually wasn't

paying attention and had no interest in

it even though I was like right at the

center of it.

So you were the son of a Jewish mother

Italian father that is true

from where I'm from in Long Island. We

call that a pizza bagel.

A pizza bagel. I've never I've never

heard that term before.

So what was your relationship with your

parents like? Yeah, I mean, you know, I

was I was always uh I was always I was

always pretty close with them. You know,

I feel like they gave me a sense of, you

know, of of kind of right and wrong and

what was important in the world. I feel

like, you know, kind of imbuing a strong

sense of responsibility is is maybe the

thing that I remember most. you know,

they were always people who felt that

sense of responsibility and, you know,

wanted to wanted to make the world um uh

wanted to make the world better. And and

I feel like, you know, that's one of the

one of the main things that I that I

learned from them. You know, it was

always a very a very um a very loving

family, a very caring family. I was very

close with my sister Daniela, who of

course became my became my became my

co-founder. And you know, I think we

decided very early that we wanted to

work together in some in some capacity.

I don't know if we imagined that it

would happen, you know, at quite the

scale that that that that it has

happened. Um uh but it um you know, I

think I think it really um you know,

that was that was something we kind of

decided early that we wanted to do.

The people that I've spoken with that

have known you through the years have

told me that your father's illness had a

big impact on you. Can you share a

little bit about that?

Um yes. Yes, he was. Um yeah, you know,

he was ill for a long time. Um and uh

eventually uh uh died in uh eventually

died in uh in in 2006. Um uh so that you

know that was actually one of the things

that drove me to you know I I don't

think we mentioned it yet in this

interview but before um you know before

I went into AI um you know I went into

biology. So, you know, I'd gone to uh I

I you know, I'd shown up at I'd shown up

at Princeton um uh wanting to be a

theoretical physicist and you know, I

did some did some work in in cosmology

for the first few few months of my time

there. um you know and and you know that

was that was around the time that that

my father died and you know that did

have an influence on me and kind of was

one of the things that convinced me um

you know to to go into biology you know

to try and address um uh uh uh you know

human illnesses and biological problems

and so I started talking to some of the

folks who worked on biohysics and

computational neuroscience in the

department that I was at at Princeton

and that was what led to the switch to

biology and computation neuroscience and

then you know of course after that I

eventually I eventually went into AI and

the reason I went into AI was actually a

continuation of that motivation which is

that um you know as I spent many years

in biology I realized that the

complexity of the

underlying problems in biology felt like

it was beyond human scale. you know, in

order to understand it all, you needed

hundreds, thousands of, you know, human

researchers and, you know, they often

had a hard time collaborating or sharing

their, you know, combining their their

knowledge. And AI, which was, I was just

starting to see the discoveries in it,

felt to me like the only technology that

could kind of bridge that gap, could

bring us beyond human scale to, you

know, to to fully understand and solve

the problems of biology. So, yeah, there

is a through line there,

right? And I could have this wrong. Uh

but one thing I heard was that his

illness was um largely unccurable when

he had it. And there have been advances

that have been Can you share a little

bit more? Yes. There are advances that

have made it much more manageable today.

Yes. Yes. That is uh that is uh that is

that that that is true. actually

actually only um uh uh only in the um uh

maybe 3 or four years after he died the

the cure rate for the disease that he

had went from uh went from 50% to uh uh

uh to roughly 95%.

Yeah. I mean it has to have felt so

unjust to have your father taken away by

something that could have been cured.

It it of course of course um but it also

tells you of the the urgency of solving

the relevant problems, right? that um

you know that that that you know there

there was someone who worked on the cure

to to this disease that you know managed

to cure it and save a bunch of people's

lives but you know could have could have

um saved even more people's lives if if

you know they they had managed to to

find that that to find that cure you

know a few years earlier than they did.

Um, and I think that's that's one of the

tensions here, right? That, you know,

um, I think AI has all of these

benefits. Um, and, you know, I want

everyone to get those benefits as soon

as possible. You know, I probably

understand, you know, better than almost

anyone how urgent those benefits are.

Um, and so I really understand the

stakes. You know, when I speak out about

AI has these risks and I'm worried about

these risks, I get very angry when

people call me a doomer. I got really

angry when, you know, when when

someone's like, "This guy's a doomer. He

wants to slow things down." You you

heard what I just said, like, you know,

my my father died because of, you know,

cures that, you know, could have could

have happened a few years later. I

understand the benefit of this

technology. When I sat down to to to

write machines of loving grace, you

know, I wrote out all the ways that

billions of people's lives could be

better with this technology. Some of

these people, some of these people who

on Twitter, you know, cheer for

acceleration. I don't think they have a

humanistic sense of the benefit of the

technology. Their their brain's just

full of adrenaline and and they're like

they want to cheer for something. They

want to accelerate. I don't get the

sense they care. Um and so when these

people call me a doomer, I think I think

they just completely completely lack any

moral credibility in doing that. Um uh

you know, it really makes me lose

respect for them. And I've been

wondering what this this word impact has

uh been because it's come up so often

that those who have been around you have

said you've been singularly obsessed

with having impact. In fact, I spoke

with someone who knew you well who said

you wouldn't watch Game of Thrones

because it wasn't tied to impact that it

was a waste of time and you wanted to be

focused on impact.

Actually, that's not quite right. I

wouldn't watch it because it was so

negative sum people were playing playing

such negative. It was like these people

start off and they're partly the

situation and partly because they're

they're just horrible people. They like

create this situation where at the end

of it everyone is like worse off than

everyone was before. Um I'm I'm really

I'm really excited about like creating

positive some situations. Um

I recommend you watch it. It's a great

it's a great show. But I hear I hear

some parts of it I was just very

reluctant and didn't watch it for a long

time.

Let's get back to the impact.

Let's get back to the impact. So that's

what impact is is effectively your

career has been this I this quest to

have that impact to be able tell me if

I'm going too far to prevent other

people from being in similar situations.

I you know I I I think I you know I

think I think that's a piece of it. I

mean you know I I have looked at you

know many uh you know many attempts to

help people and you know some of them

are more effective than others. Um, and

you know, I think I think I've always

tried to, you know, there should be

strategy behind it. There should be

brains behind um, you know, trying to

trying to help people. Um, you know,

which often means that there's a long

path to it, right? It can run through a

company and, you know, many activities

that are technical and not immediately

tied to the the kind of impact impact

that you're trying to have. But, you

know, the the the the arc is I'm always

trying to bend the arc towards that. I

think I think that's my that's my

picture of it. That's that's really why

I that's really why I got into this,

right? You know, um I think you know,

similar to the reason to get into AI was

that um you know, I I saw the problems

of biology as as almost intractable

without it or at least too slow moving.

Um, you know, I think my reason to start

a company was that I had worked at other

companies and I I I I just didn't feel

like the way those companies were were

run was was really oriented towards, you

know, trying trying trying to have that

impact. There was a story around it that

was often used for recruiting, but it

became clear to me over the years that

story was not sincere.

I'm going to circle around a little bit

because it's clear that you're referring

to OpenAI here. Um,

from what I understand, you had 50% of

OpenAI's compute. I mean, you ran the

GPT3 project. So, if anyone was going to

be focused on impact and safety,

wouldn't have been you.

Uh, yes. I I was, you know, there was a

period during which uh during which that

was uh that was true. That wasn't true

the entire time. That was, for example,

when we were scaling up GPT3. Um, yeah.

So, you know, I I when I was at OpenAI,

I and a lot of my colleagues, including

the people who, you know, eventually um

eventually founded Anthropic,

the pandas,

um the pandas. Um

that's the name you gave them.

I that that isn't a name I gave them.

The name they took.

Uh that isn't a name they took. Um

that's the name other people called

them.

Uh I I maybe it's a name other people

called them. Uh that's not a name I ever

used for my team.

Okay. Sorry. Go ahead.

Uh

that's good clarification. Thank you. Um

uh so uh yeah um you know we were

involved in scaling up these models

actually the original reason for

building GPT2 and GPT3 it was an

outgrowth of the kind of AI alignment

work that we were doing right where

myself and Paul Cristiano and some of

the anthropic co-founders had invented

this technique called RL from human

feedback um uh and that was designed to

help steer models in um uh you know in a

direction to follow human intent. It was

actually a precursor to you know we were

trying to scale up another method called

scalable supervision which I think is

just starting to to to work many years

later to help models follow more kind of

scalable uh uh uh human intent. But what

we found is even with the more primitive

technique RL from human feedback it

wasn't working with the small language

models with you know GPT1 that we

applied it to and that had had been

built by other people at OpenAI. So the

scaling up of GPT2 and GPT3 was done in

order to kind of study these techniques

in order to apply RL from human feedback

at scale. Um uh you know this goes to

one thing which is that I think in this

field the alignment of AI systems and

the capability of AI systems is

intertwined in this way that always ends

up being kind of more tied and more

intertwined than we think. Um actually

what this made me realize is that it's

very hard to work on the safety of AI

systems and the um capability of AI

systems separately. It's very hard to

work on one and not the other. I

actually think the value and the um way

to inflect the field in a more positive

way comes from organizational level

decisions. when to release things, when

to study things internally, what kind of

work to do on systems. Um, and that was

one of the things that kind of

motivated, you know, me and some of the

other, you know, to be entropic founders

to kind of go off and do it our own way.

But again like if you were driving if

you think language uh if you think

capabilities and safety are interlin and

you were the guy driving the cutting

edge models within open AI you know you

if you left you knew they were going to

be a company that was still doing this

stuff. That's right.

It seems like if you're driving the

capabilities you'd be the one in the

driver's seat to help it be safe the way

that you wanted to. Again, I will say um

you know if there's a decision on

releasing a model, if there's a decision

on the governance of the company, if

there's a decision on you know how the

personnel of the company works, um you

know, how the company represents itself

externally, the decisions that the

company makes with respect to

deployment, the claims it makes about

how it operates with respect to society.

um you know many of those things are not

things that you control just by just by

uh training the model and you know I

think I think trust is really important

I think the leaders of the company of a

company they have to be trustworthy

people they have to be people whose

motivations are sincere no matter how

much you're driving the the forward the

company technically if you're working

for someone whose motivations are not

sincere who's not an honest person who

does not truly want to make the world

better it's not going to work you're

just contributing to something bad.

So, and I'm I'm sure you've heard the

criticism from people like Jensen who

say, "Well, Daario thinks he's the only

one who can build this safely and

therefore, speaking of that word

control, wants to control the entire

industry."

I've never said I've never said anything

like that. That's an outrageous lie.

That's the most outrageous lie I've ever

heard. Um,

by the way, I'm sorry if I got Jensen's

words wrong, but

No, no, no. The words were correct.

Okay. the but but but but the words are

the words are outrageous. In fact,

I've said multiple times and I think

Anthropic's actions have shown it

that um

you know, we're aiming for something we

call a race to the top. Um you know,

I've said this on podcasts over the

years and I think anthropics actions

have shown it where you know with a race

to the bottom, right, everyone is

competing to like, you know, get things

out as fast as possible. And so I say

when you have a race to the bottom, it

doesn't matter who wins, everyone loses,

right? Because you make the unsafe

system that you know helps your

adversary or causes economic problems or

uh you know is is unsafe from an

alignment perspective. The way I think

about the race to the top is that um it

doesn't matter it doesn't matter who it

doesn't matter who wins, everyone wins,

right? So the way the race to the top

works is you set an example for how the

field works. to say um uh uh you know

we're going to engage in this practice.

So a key example of this is responsible

scaling policies. We were the first to

put out a responsible scaling policy and

you know we didn't say everyone else

should do this or you're bad guys. We

didn't you know we didn't you know kind

of try to use it at his advantage. We

put it out and then we encouraged

everyone else to do it. Um and many and

and then we discovered in the months

after that that you know there were

people within the other companies who

were trying to put out responsible

scaling policies but the fact that we

had done it allowed you know gave those

people permission right kind of kind of

enabled those people to um you know to

make the argument to leadership hey

anthropic is doing this so we should do

it as well. The same has been true of

investing in interpretability. we

release our interpretability research to

everyone um uh and allow other companies

to copy it even though we've seen that

it sometimes has commercial advantages

same with things like constitutional AI

same with the the the measure the the

the you know the measurement of the

dangers of our system dangerous

capabilities evals so we're trying to

set an example for the field but there's

an interplay where it helps to be a

powerful commercial competitor I've said

nothing that that any that anywhere ware

near resembles the idea that this

company should be the only one to build

the technology. I don't know how anyone

could ever derive that from anything

that I've said. Um it's it's it's it's

just Yeah. Yeah. It's just it's just

it's just a it's just an incredible and

bad faith distortion.

All right, let's see if we can

lightening around like one or two before

I ask you the last one, which we'll have

five minutes for. Um

what happened with SPF? Like

what happened with SPF?

I mean, he was one of Go ahead.

I couldn't tell you. I couldn't was what

was the what didn't you answer?

I I probably met the guy four or five

times.

Um uh so I have no great insight um into

the you know what in you know into the

psychology of SPF or or you know why why

he did things as stupid or immoral as as

as as as

he did. I think the only uh the only you

know the the only thing I had ever seen

ahead of time with uh SPF was uh you

know a couple people mentioned to me

that he was like hard to work with that

you know he was he was like a bit of a

move fast and break things guy.

Um and I was like okay you know there's

like plenty of people

Welcome to Silicon Valley.

Yeah. Like welcome welcome to Silicon

Valley. Um and so I remember saying okay

I'm going to give this guy non- voting

chairs. I'm not going to put him on the

board. He sounds like a bit, you know,

he sounds like a bad person to deal with

every day. Um, but, you know, he's

excited about AI. He's excited about AI

safety. He's, he's, you know, he's a

bull on on on AI and he's interested in

AI safety. So, you know, seems like a

seems like a sensible seems like a

sensible thing to do. you know in in uh

in in you know in like in in in in

retrospect um you know that that you

know move fast and and and break things

you know was turned out to be much much

much more extreme and bad than than you

know than than than I ever imagined.

Okay. So let's end here. So you found

your impact I mean you're you're working

the dream pretty much right now. I mean

think about all the ways that uh AI can

be used for biology uh just a start. You

also say that this is a dangerous

technology and I'm curious if your

desire for impact um could be pushing

you to accelerate this technology um

while you know potentially devaluing the

possibility that it could that

controlling it might not be feasible. So

you know I think I have more than anyone

else in the industry warned about the

dangers of the technology. Right? We

just spent 10 20 minutes talking about,

you know, the the frightening the, you

know, the large array of, you know,

people who run, you know, trillion

dollar companies criticizing me for, you

know, for talking about the the the

dangers of these technologies, right?

You know, I have US government

officials. I have people who run $4

trillion companies criticizing me for

talking about the dangers of the

technology, right? imputing all these

bizarre motives that bear no

relationship to, you know, to anything

I've ever said, not supported in

anything I've ever done. And yet, I'm

going to continue to do it. Um, I

actually think that, you know, as the

revenues, as the economic business of AI

ramps up, and it's ramping up

exponentially, you know, if if I'm

right, in a couple years, it'll be the

biggest source of revenue in the world,

right? It'll be the biggest industry in

the world. And people who run companies

already think it. So we actually have

this terrifying situation where uh you

know hundreds of billions to trillions

to I would say maybe 20 trillion of

capitals on the side of accelerate AI as

fast as possible. we have this, you

know, company that's very valuable in

absolute terms, but you know, looks very

small compared to that, right? 60 60 $60

60 billion. And I keep speaking up even

if, you know, it makes folks in, you

know, there have been these articles f

you know, some folks in the US

government are upset at us, for example,

for opposing the moratorium on AI

regulation, for being in favor of export

controls for chips on China, for talking

about the economic impacts of AI. Every

time I do that, I get attacked by many

of my peers.

Right. But you're still assuming that we

can control it. That's what I'm pointing

out.

But I'm I'm just I'm just telling you

how much how much effort, how much

persistence, how much despite everything

that stacked up, despite all the

dangers, despite the risk that it has to

the company of being willing to speak

up, I'm willing to do it. And and and

that's that's what that's that's why I'm

saying that look if if I thought that

there was no way to control the

technology, right? If I thought even

even if I thought this is just a gamble,

right? Some people are like, "Oh, you

think there's a five or 10% chance that

AI could go wrong, you're just rolling

the dice." That's not the way I think

about it. This is a multi-step game,

right? You take one step, you build the

next step of most powerful models, you

have a more intensive testing regime. As

we get closer and closer to the more

powerful models, I'm speaking up more

and more and I'm taking more and more

drastic actions because I'm concerned

that the risks of AI are getting closer

and closer. We're working to address

them. we've made a certain amount of

progress, but when I worry that the

progress that we've made on the risks

does not you know is not fully aligned

with the um uh you know is not going as

fast as we need to go for the speed of

the technology then I speed up then I

then I speak up louder. Um, and so you

know, you're asking why am I why am I

talk, you know, what, you know, you

started this interview by saying what's

gotten into you? Why are you talking

about this? It's because the exponential

is getting to the point that that I

worry that we may have a situation that

our ability to handle the risk is not

keeping up with the speed of the

technology. And that's how I'm

responding to it. If I believe that

there was no way to control a

technology, which I I I I I see

absolutely no evidence for that

proposition, we've gotten better at

controlling models with every model with

every model that we release, right? All

these things go wrong, but like you

really you really have to stress test

the models pretty hard. That doesn't

mean you can't have emergent bad

behavior. And I think, you know, if we

got to much more powerful models with

only the alignment techniques we have

now, then I'd be very concerned. then

I'd be out there saying everyone should

stop building these things. Even China

should stop building these. I don't

think they'd listen to me, which is one

reason I think export controls is a

better is is is is a better measure. But

if if we got a few years ahead in models

and had only the alignment and steering

techniques we had today, then you know,

I would definitely be advocating for us

to to, you know, to to to to slow down a

lot. The reason I'm warning about the

risk is so that we don't have to slow

down. So that we can invest in safety

techniques and can continue the progress

continue the progress of the field. It

would be a huge economic effort even if

one company was willing to slow down the

technology. You know that doesn't stop

all the other companies that doesn't

stop our geopolitical adversaries to

whom this is a existential fight fight

for survival. So, you know, there

there's there's there's very little, you

know, there's very little latitude here,

right? We're stuck between all the

benefits of the technology, the race to

the race to accelerate it and the fact

that that is a multi-party race. And so,

I am doing the best thing I can do,

which is to invest in safety technology

to speed to to speed up the progress of

safety. I've written essays on the

importance of interpretability on how

important various directions in uh in in

in safety are. We release all of our

safety work openly because we think

that's the thing that's a public good.

That's the thing that everyone that that

that that that everyone needs to share.

So if you have if you have a better

strategy for balancing the benefits, the

inevitability of the technology and the

risks that it face, I am very open to

hear it because I go to sleep every

night thinking about it because I have

such an incredible understanding of the

stakes in terms of in terms of the

benefits in terms of you know what it

can do, the lives that it can save. I've

seen that personally. I also have seen

the risks personally. We've already seen

things go wrong with the models. You

know, we have an example of that with

Grock. And you know, people dismiss

this, but they're not going to laugh

anymore when the models are taking

actions, when they're manufacturing, and

when they're in charge of, you know,

medical medical interventions, right?

People can laugh at the at the risks

when the models are just talking. But I

think it's very serious. And so I think

what this situation demands is a very

serious understanding of both the risks

and the benefits. These are highstakes

decisions. They need to be made with

they they they they need to be made with

a seriousness. And and I think something

that makes me very concerned is that on

one hand we have a cadre of people who

are just doomers. People call me a

doomer. I'm not. But there are doomers

out there. People who say they know

there's no way to build this safely. You

know, I I've I've looked at their

arguments. They're a bunch of

gobbledegook. The idea that these models

have dangers associated with them,

including dangers to humanity as a

whole, that makes sense to me. The idea

that we can kind of logically prove that

there's no way to make them safe, that

seems like nonsense to me. So, so I

think that is an intellectually and

morally unserious way to respond to the

situation we're in. I also think it is

intellectually and morally unserious for

people who are sitting on 20 trillion

dollars of capital who all work together

because their incentives are all in the

same way. There are dollar signs in all

of their eyes um to sit there and say we

shouldn't regulate this technology for

10 years. Anyone who says that we should

worry about the safety of these models

is someone who just wants to control the

technology themselves. That's an

outrageous claim and it's a morally

unserious claim. We've sat here and

we've done every possible piece of

research. We speak up when we believe

it's appropriate to do so. We've tried

to back up, you know, when we make

claims about the economic impact of AI.

We have an economic research council. We

have an we have a we have a economic

index that we use to track the model in

real time. And we're giving grants for

people to understand the economic the

economic impact the the economic impact

of the technology. I think for people

who are far more financially invested in

the success of the technology than than

than than I am to just you know breezily

lob add ad homonym attacks you know I

think that is just as intellectually and

morally unserious as the doomer's

position um I think what we need here is

we need more thoughtfulness we need more

honesty we need more people willing to

willing to go against their interest

willing to not have you know breezy

Twitter fights, uh, hot takes. We need

people to actually invest in

understanding the situation, actually do

the work, actually put out the research,

and and actually add some light and some

insight to to the situation that we're

in. I am trying to do that. I don't

think I'm doing that perfectly as no

human can. I'm trying to do it as well

as I can. It would be very helpful if

there were others who would try to do

the same thing. Well, Dario, I said this

off camera, but I want to make sure to

say it on as we're wrapping up. Um, I

appreciate how much Anthropic publishes.

We have learned a ton from the

experiments, everything from red teaming

the models to vending machine Claude,

which we didn't have a chance to speak

about today. Um, but I think the world

is better off just to hear everything

going on here. And, and to that note,

thank you for sitting down with me and

spending so much time together.

Thanks for having me.

Thanks everybody for listening and

watching. and we'll see you next time on

Big Technology Podcast.

Loading...

Loading video analysis...