TLDW logo

Amjad Masad & Adam D’Angelo: How Far Are We From AGI?

By a16z

Summary

## Key takeaways - **AGI is 5 years away: Automating remote work**: Adam D'Angelo believes that within five years, AI will be able to automate a large portion of jobs, satisfying many current critiques of AI, even if not fully reaching AGI. [01:55], [02:29] - **LLMs are not AGI; they have clear limitations**: Amjad Masad argues that LLMs are a different kind of intelligence than human intelligence, with clear limitations that are being papered over, and they are not on the path to true AGI. [06:00], [06:43] - **AI creates a 'missing middle' in jobs**: AI may automate entry-level jobs but not expert roles, leading to a situation where new people aren't hired because agents are more efficient, creating a 'weird equilibrium' and a 'missing middle' in the job market. [15:35], [16:26] - **Solo entrepreneurs enabled by AI**: The technology is vastly increasing what a single person can do, enabling a massive number of solo entrepreneurs to bring ideas into existence that previously required teams and funding. [00:25], [28:51] - **The era of agents is here**: Amjad Masad states that we are entering the decade of agents, where AI can handle the entire development lifecycle, including coding, infrastructure, testing, and debugging, significantly boosting developer productivity. [45:04], [49:13] - **Underhyped 'vibe coding' potential**: Amjad Masad believes 'vibe coding' is unbelievably high potential, opening up software creation to the mainstream by making tools capable of what once required large teams of engineers. [52:56], [54:11]

Topics Covered

  • Is Brute Force AI Enough, or Do We Need True Intelligence?
  • Do LLMs Create an Expert Bottleneck, Eliminating New Talent?
  • Will AI Massively Empower Solo Entrepreneurs Globally?
  • Is AI a Disruptor or a Supercharger for Incumbents?
  • Is Undocumented Human Knowledge AI's Next Big Bottleneck?

Full Transcript

Nothing seems fundamentally so hard that

it couldn't be solved by the smartest

people in the world working incredibly

hard for the next five years.

>> Humanity went through the agricultural

revolution and the industrial

revolution. We're going through another

revolution. We will not be able to call

it something. It's like future people

will call it something. But we are going

through something. The number of solo

entrepreneurs that this technology is

going to enable. It's vastly increased

what a single person can do.

>> For the first time, opportunity is

massively available for everyone. Just

the ability for more people to be able

to become entrepreneurs is Yeah,

>> it's massive.

>> Adam, welcome to the podcast.

>> Thank you. Yeah, thanks for having us.

>> So, a lot of people have been throwing

cold water over LLMs lately. It's been

some general bearishness. People talking

about the limitations of of LLMs, why

they won't get us to AGI. Well, maybe uh

what we thought was just a couple years

away is now maybe 10 years away. Adam,

you seem a bit more optimistic. Why

don't you share your broad general

overview?

>> Yeah, I mean I I actually honestly I

don't know what people are talking

about. I think I think if you look a

year ago, the world was very different.

And so just judging on how much progress

we've made in the last year with things

like reasoning models, um things like

the improvement in code generation

ability, um the improvements in video

gen, it seems like things are going

faster than ever. And so I I don't

really understand where the the kind of

bearishness is coming from. Well, I

think there's some sense that we hoped

that they would be able to um replace

all of tasks or all all jobs. And maybe

there's some sense that it's like middle

to middle but not end to end. And maybe,

you know, labor won't be automated in

the same way that we we thought it would

on the same timeline.

>> Yeah. I mean, I I don't know what the

previous timelines people were were

thinking were, but you know, I think I

think if you if you go 5 years out from

now, we're in a very different world. I

think I think a lot of what's holding

back the models these days is not

actually intelligence. It's getting the

right context into the model so that it

can be able to to use its intelligence.

Um, and then there's some things like

computer use that are still not quite

there, but I I think we'll almost

definitely get there in the next year or

two. And when you have that, I I think

we're going to be able to automate a

large portion of what people do. I don't

think I don't know if I would call that

AGI, but I I think it's going to satisfy

a a lot of the critiques that people are

making right now. I I think they won't

be valid in in a year or two.

>> What is your definition of AGI?

>> I don't know. Everyone everyone thinks

it's something different. I think I mean

you know one one definition I I I I kind

of like is um if you say that you have a

remote worker a human any job that could

be done by someone whose job can be done

remotely

um that that's AGI you know you can you

can then say does have to be better than

the best person in the world at every

single job some people call that ASI um

does have to be better than teams of

people

you can you can argue with those

different definitions. But I I think

once we get to be better than a typical

remote worker at the job they're doing,

we're living in a a very different

world. And I think that's that's

effectively what people that that's a

very useful anchor point for for these

definitions.

>> So in summary, you're not sensing the

same limitations of LM that other people

are. You think there's a lot more room

that LMS can can go from here? We don't

need like a brand new architecture or

other breakthrough.

>> I don't think so. I mean I I think there

are certain things like memory and

learning like continuous learning that

are not very easy with the current

architectures. I think even those you

can sort of fake and maybe we're going

to be able to to get them to work well

enough. Um but we we just don't seem to

be hitting any kind of of limits. The

the progress in reasoning models is

incredible. And I think the progress in

in pre-training is is also going pretty

quickly. Maybe not as quickly as people

had expected, but certainly fast enough

that you can expect a lot of progress

over over the next few years.

>> Amad, what's your what's your reaction

hearing all this? Yeah, I I I think I've

been pretty consistent and consistently

right perhaps

>> dare I say

>> consistent with yourself or consistent

with what I'm saying

>> with with um with myself and with I

think how things are unfolding that uh

you know I started being a bit of a more

public doubter of of things around uh

the time when the AI safety discussion

was uh was reaching its height back in

maybe 22 23. Um, and I I thought it was

important for us to be realistic about

the progress. Um, because, you know,

otherwise we're going to scare

politicians. We're going to scare

everyone. You know, uh, DC will descend

on Silicon Valley. We they'll shut

everything down. So my criticism of the

idea of like AGI 2027, you know, that

paper that I think it's called

Alexander, someone else wrote

>> uh and then um and the situational

awareness and all this uh hype papers

that are not really science, they're

just vibe. Here's what I think will

happen. Uh you know, the whole economy

will get automated. You know, jobs are

uh are going to disappear. all of that

stuff is that again is just I think um

it's unrealistic. It is not following

the kind of progress that we're seeing

and it is uh going to lead to just bad

policy. So my view is LMS are amazing

amazing machines. Uh I don't think they

are exactly human uh intelligence

equivalent. You can still trick LMS with

things like they might have solved the

strawberry one, but you can still, you

know, uh trick it with like single

sentence questions like how many Rs are

in this sentence. I think I think I

tweeted about it the other day, which

was like three out of the four four

models couldn't didn't get it even. Um

and then GP5 with high thinking had to

think for like 15 seconds in order to

get a question like that. So uh LMS are

I think a different kind of intelligence

than uh what humans are uh and also

uh they have they have clear limitations

and we're papering over the limitations

and we're kind of working around them in

all sorts of ways whether it's in the

LLM itself and the training data or uh

and the infrastructure around and

everything that we're we're doing to

make them work. Um but that that makes

me less optimistic that we're we've

we've cracked intelligence. And I think

once we truly crack intelligence

um it'll feel a lot more scalable and

that you can uh and that the the idea

behind the lesson will actually be true

and that you can just pour more um more

power, more resources, more compute into

them and they'll they'll just scale more

naturally. I think right now uh there's

a lot of manual work going into making

these models better. In the pre in the

true pre-training scaling era, you know,

GPT2, 3, 3.5, maybe up to four, um it it

felt like you you can just uh put more

internet data in there and just it just

got better. uh whereas now it feels like

there's a lot of labeling work

happening. There's a lot of contracting

work happening. A lot of these uh

contrived RL environments are getting

created in order to make uh LLMs good at

coding and becoming coding agents and

they're going to go do that. I think the

news from OpenAI that they're going to

do that for for investment banking. And

so I uh try to coin this term I call

functional AGI which is the idea that

you can automate a lot of aspects of a

lot of jobs by just going in and like

collecting as much data and creating

these RL environments. It's going to

take enormous effort and money and data

and all of that in order to do and I

think we're yeah I I I agree with Adam

that you know things are going to get

better uh 100% over the next 3 months 6

months cloud 4.5 was a huge jump uh I

don't think it's appreciated how much of

a jump it was over over four there's

really really amazing things about cloud

4.5 so there is progress we're going to

continue to see progress I don't think

LM as they can understand are on on the

way to AGI. And my definition for AGI is

I think the old school RL definition,

which is um a machine that can go into

any environment and learn efficiently in

the same way that a human could go into

uh you can put a put a human into a a a

pool game and you know within 2 hours

they can like shoot pool and be able to

do it. Uh right now there's no way for

us to have machines learn skills like

that on the fly. You know everything

requires enormous amount of data and

compute and time and effort and and and

uh and more importantly it requires

human expertise which is the non bitter

lesson uh idea which is you know uh

human expertise is not scalable and we

are relying today we are in a human

expertise regime. Yeah, I mean I I think

that

humans are certainly better at learning

a new skill from a limited amount of

data in a new environment than the

current models are. I think that on the

other hand, human intelligence is the

product of evolution which used a

massive amount of effective computation.

And so this is a different this is a

different kind of intelligence. And so

because it didn't have this this massive

equivalent of evolution, it just has

pre-training for for that which is not

as good. You then need more data to

learn everything, every new skill. But I

guess I think in terms of like the

functional consequence. So like if if

you're like when when will the world

when will the job landscape change? When

will the e economic growth hit? I think

that's going to be more a function of

when we can produce something that is as

good as human intelligence. Even if it

takes a lot more compute, a lot more

energy, a lot more training data, we

could just put in all that energy and

still get to software that's as good as

the average person at doing a typical

job.

>> So, I don't disagree with that and and

that's it is it feels like we're in a

brute force type of regime, but but

maybe that's fine. And

>> yeah.

>> Yeah.

>> So, where's the disagreement then, I

guess? So, there's agreement on that.

Where is the deer?

>> I I don't think that we'll get to the

singularity or I don't think that I

don't think we're going to get to the

next level of human civilization

uh until we um we we we crack the true

nature of intelligence

like until we understand and have

algorithms that are actually uh not

brute force and and you think those will

take a long time to come? Uh I I'm sort

of agnostic on on that. It just does it

does feel like the LMS

uh in a way are distracting from that

because um all the talent is going there

um and therefore there's less talent

that are trying to do basic research on

on intelligence.

>> Mhm. Yeah. At the same time a huge

portion of talent is going into AI

research that used to previously

wouldn't have gone into AI at all.

>> Mhm. And so you have this this massive

industry, massive funding, you know,

funding compute but also funding human

employees. And that is

I guess I nothing seems fundamentally

so hard that it couldn't be solved by

the smartest people in the world working

incredibly hard for the next 5 years on

it. But but basic research is is

different, right? like trying to

um like trying to get into the

fundamentals and as opposed to like

there's a lot of industry research like

how do we make these things more useful

uh in order to generate profit and um so

I I think that's that's different and

often I mean Thomas [ __ ] this

philosopher of science talks a lot about

how these research programs end up you

know becoming like a bubble and like

sucking all the attention and ideas and

like think think about physics and how

there are like these industry of a

string theory and like it pulls

everything in and there's sort of a plug

black hole of progress and you know

>> Yeah. Yeah. No, and I think I think one

of his things was like you got to wait

until the current people retire even

have a chance at changing the paradigm.

>> He's very pessimistic about paradigms.

But I I guess I feel like the current

paradigm, this is maybe our disagree, I

think the current paradigm is pretty

good and I think we're nowhere near the

sort of like diminishing returns of

continuing to push on it.

>> Mhm.

>> And I bet Yeah, I guess I would just bet

that you can keep doing different

innovations within the paradigm to to

get there.

>> So let let's say we continue to brute

force it. um we're able to automate a

bunch of labor. Do you estimate that GDP

is is something you know four or five

percent a year or are we going up to 10%

plus or what does it do to the economy?

>> I think it it depends a lot on exactly

where we get to and what what AGI means.

But so so let's say you have let's say

you have

LLM that with with an amount of energy

that costs

$1 an hour,

they could do a job of any

human. Let's just just just take that as

a as a theoretical point you could get

to. I think you're going to get to much

more than four to 5% GDP growth in that

world. I think the issue is you may not

get there. So it may be that the LMS

that can do everything a human can do

actually cost more than humans do

currently or they can do kind of like

80% of what humans can do and then

there's this other 20%. Um and and I I

think I do think at some point you get

to also

like I I don't see a reason why we don't

eventually get there. That may take

five, 10, 15 years. But I think until

you get there, we're going to get

bottlenecked on the things that the LM

still can't do or the, you know,

building enough power plants to to

supply the energy or

other bottlenecks in in the supply

chain.

One thing I worry about uh is uh the

delotterious effect of LMS in the

economy in that say LM's uh you know

effectively automate uh the entry level

job but not but but but the but not the

expert's job right so um let's take uh

you know QA Q quality assurance um And

uh it it's it's so good, but uh there's

still all these longtail event uh you

know events that it doesn't handle. And

so you have a lot of uh really good QA

people now like managing like hundreds

of agents and you effectively increase

productivity a lot. Uh but they're not

hiring new people because the agents are

better than new people. Uh and and and

that that feels like a weird equilibrium

to be in, right? And I don't think that

many people are thinking about it.

>> Yeah. Yeah. For sure. Yeah. No, I I I

think that's, you know, I think it's

happening with um CS majors graduating

from college, there's just not as many

jobs as there used to be. And

>> and um LLMs are a little more

substitutable for what they previously

would have done. And I'm I'm sure that's

contributing to it. And then it means

that you're going to have fewer people

going up that ramp that, you know,

companies paid a lot of money to to

employ them and and and train them. Um

and so I I think it's a real problem. I

think it's going to I'm guessing you'll

probably see some kind of like that

problem also creates a economic

incentive to solve the problem. So

>> it may be that there's like more

opportunities for companies that can

train people or maybe use of AI to to

teach people these things. Um but for

sure that's that's an issue right now.

Another related problem is that uh since

we're dependent on uh expert data in

order to train the LLMs and the LM start

to substitute

um those workers but but but you know at

some point there's no more experts

because they're all out of jobs and and

and and they're equivalent to the LLMs.

If the LMS is truly dependent on on

labeling data, expert RL environments,

then how would they improve beyond that?

I think that's something question for an

economist to really sit down and think

about is like once you get the first

tick of automation, I mean there there

are some challenges there. And so how do

you go how do you go how do you go to

the next part? Yeah, I mean I I think it

a lot of it is going to depend on how

good of RL environments can be

>> created. So, you know, in the one

extreme you have something like Alph Go

where just a perfect environment and you

can just blast past expert level. Um,

but I think a lot of jobs have limited

data that anyone can can train from. And

so I think it'll be interesting to see

how how easy is it for

research efforts to to overcome that

that bottleneck. If you had to make a

guess on what job category is going to

be introduced or explode in in the

future um you know some people say it's

like the you know everyone's an

influencer you know or in some sort of

caring um field or um you know

everyone's employed by the government

and some sort of bureaucrat thing or um

you know maybe training the AI in in

some way uh you know as as more and more

things start to get automated you know

what is your your guess as to what

more and more people start to

you know, doing art and poetry is

>> at some point you have everything

automated and then I think people will

do art and poetry and you know there's a

data point that the people playing chess

is up since computers got better at

human than than humans at chess. So I

don't think that's

a bad world if people are all just kind

of free to to pursue their their

hobbies. uh as long as you have some

kind of you know way to distribute

wealth so that so people can afford to

to live. Um but I you know in the near

that that's a while away and in the near

term

>> well like 10 15 years out

>> I I don't know how much but yeah in the

in the I'll put it in the at least 10

years range. Um, I I think in the near

term the job categories that are going

to explode, the jobs that can really

leverage AI and so so people who are

good at using AI to to accomplish their

jobs, especially to accomplish things

that the AI couldn't have done by

itself, there's just there's just

massive demand for for that.

>> I don't think we're going to get to a

point where you automate every every

job. Uh, definitely not in the current

paradigm. I would uh I would doubt it

happening.

I I I'm not certain it would ever

happen, but definitely not in the

current paradigm. Now, here's why I

think because a lot of jobs is about

servicing other humans. You need to be

fundamentally human in order to you need

to be actually human in order to

understand what other people want, you

know, and so you need to have the human

experience. So unless we're going to uh

create uh human humans, unless the m

unless AI is actually embodied in the

human experience, then humans will

always be the generators of ideas in the

economy. Adam, respond to Andre's point

around the human part because you

created one of the most, you know, the

best wisdom of the crowds, you know, uh

platforms in in the universe. Um and now

you've gone, you know, all all in with

Po. Um what are your thoughts on you

know to what extent will we be relying

on um humans versus will we be trusting

AIs to you know be our therapists be our

you know caretakers in other ways.

Humans have a lot of knowledge

collectively and you know even like one

individual person who's an expert and

has lived a whole life and had a whole

career and seen a lot of things they

they often know a lot of things that are

not written down anywhere

>> tacet knowledge

>> and um you call it tested knowledge but

also also what they're capable of

writing down if you did ask them a

question I think there's still an

important role for for people to play in

the

by sharing their knowledge, especially

when they have knowledge that that just

wasn't otherwise in an LLM's training

set. Um, you know, whether they will be

able to make a full-time living doing

that, I I don't know. But if that

becomes a bottleneck, then then for sure

that's going to mean that all the sort

of like economic pressure goes goes to

that. I don't in terms of the like you

know you have to be human to know what

humans want. I don't know about that. So

like as an example I think I think

recommener systems

the system that ranks your Facebook or

Instagram or Kora feed those recommener

systems are already superhuman at

predicting what you're going to be

interested in in reading. Like if if if

I gave you a task that was like make me

a feed that I'm going to read, like

there there's just no way. No matter how

much you knew about me, there's no way

you could compete with these algorithms

that just have so much data about

everything I've ever clicked on,

everything everyone else has ever

clicked on, what all the similarities

are between all those those different

data sets. And so I don't know, you

know, it's true that as a human you can

kind of like simulate being a human and

that makes it easier for you to like

test out ideas. And I'm sure that

composers and artists are this is an

important part of their their process

for doing work is they

>> or chefs or Yeah.

>> Yeah. They they produce something and

you know a chef will cook something and

they taste it and it's important that

they can taste it

>> but I don't know you know they they just

they have very little data compared to

what AI can be trained on. So So I I

don't know how that's going to shake

out.

>> That's a that's a good point. I mean

ultimately what recommended systems uh

are they're like aggregating all the

different tastes and then sort of

finding where you sit in the sort of

multi-dimensional taste vector space and

like getting you the best content there.

So I guess there's some of that. I think

that's more narrow than we think like

like yes it it's true in recommener

systems but I'm not entirely sure it's

true of of of everything. Um but so I I

think the best prediction for

where the world is headed and this is

not a

endorsement or necessarily like this is

where I think the world's headed because

I think part of it is

uh will be slightly in uh instable

unstable system but I think the

sovereign individual continues to be I

think a really good set of predictions

for the future although it's not a

scientific book or not. It's a very pyic

book and um but but the idea is uh you

know in the late 80s early 90s

um are they economists? I'm not sure. I

think they're economists or political

science majors uh two people out of the

UK um wrote this book about trying to

predict what happens uh when computer

technology matures, right? They're like,

you know, humanity went through the

agricultural revolution and the

industrial revolution. We're going

through another revolution. Uh clearly,

uh information revolution, now we call

it intelligence revolution, whatever. I

think we will not be able to call it

something. It's a future people will

call it something, but we are going

through something. And so they're trying

to predict, okay, what happens from

here? And what they arrive at is that

the um ultimately you're going to have

large swaths of people that are

potentially unemployed or economically

not um contributing, but you're going to

have the entrepreneur the entrepreneur

capitalists going to be

so highly leveraged because they can

spin up these companies with AI agents

very quickly. Oh, because they have this

because they're very generative. They

have interesting ideas. They're human.

They've uh they have interesting ideas

about what other people want. They can

create these companies very quickly in

these products and services and they can

organize the economy in certain ways.

And the politics will change because uh

to you know today's politics is based on

um

every human being uh economically

productive. Uh but when you have only uh

when you have massive automation and

then you have a few entrepreneurs and

very intelligent generative people are

actually uh able to be productive then

the political structures also change. Um

uh and so they talk about how the you

know nation state sort of subsides and

instead you go back to uh to an era

where um states are like competing over

people over wealthy people and like they

you know uh as a sovereign individual

you can like uh negotiate your tax rate

with your favorite state and so it

starts to sound like biology a little

bit and I don't think it is far from

where I where it might be headed. Now

again, it's it's not a sort of a value

judgment or or desire. Uh but but I do

think it's worth thinking about when

when people are not the the

you know unit of economic productivity,

things have to change, including culture

and and politics.

>> Yeah. I I think there's a question with

that book and in some of this

conversation more broadly of like when

does a technology reward the uh you know

the defender versus this the sort of

aggregator or something or like the um

when does it incentivize more

decentralization versus centraliz like

uh remember Peter Tiel had this quip a

decade ago of like you know crypto is

libertarian is more decentralizing AI is

you know communist or more centralizing

and it it um it's not obvious to me that

that that that's entirely accurate. um

on on either side AI does seem to

empower a bunch of individuals as you

were saying and then also you know

crypto turns out is like fintech or it's

like stable you know uh it does empower

sort of uh you know in nation states

we're talking about doing the sort of

like you know the the China thing that

they were going to do so yeah I think

there's an open question as to

you know which technology leads to who

does it empower more the edges or the

the center and I think if it empowers

the edges it seems like the sovereign

individual is is and and maybe there's a

barbell uh where it's like both

basically the big the incumbents just

get much much much much bigger and

there's like these edges but anyways

that's

>> I'm I'm very excited for the um the

number of solo entrepreneurs that

>> this technology is going to enable. I

think it's it's just greatly it's vastly

increased what what a single person can

do and there's so many ideas that just

never got explored because it's a lot of

work to get a team of people together

and maybe raise the funding for it and

get the right kind of people with all

the different skills you need. Um and

now that one person can can bring these

things into existence, I I think I think

we're going to see a lot of really

amazing stuff. Yeah, I get these tweets

all the time about people who like quit

their jobs because they started making

so much money. You're using tools like

like rapid and um it's it's really

exciting. I think uh if for the first

time opportunity is massively available

for for everyone

>> uh and I think that that is to me the

most exciting thing about this

technology other than all the other

stuff that we're talking about just the

ability for more people to be able to

become entrepreneurs is yeah it's

massive

>> that that trend is obviously going to

happen as we look out of the next decade

or two do you think that AI is more

likely to be sustaining or disruptive in

the Christian sense to ask it another

Okay. Do you think that most of the

value capture is going to come from

companies that were scaled pre OpenAI

starting? Um uh so is so replet still

counts as the the latter and so does

court to some degree or or do um do you

think most of the value is going to be

captured by companies that started you

know after let's say 2015 2016? So

there's a related question which is how

much of the value is going to go to the

hyperscalers versus everyone else and I

think on that one we are I actually

think we're in a pretty good balance

where there's enough competition among

the hyperscalers that the

um there's enough competition that as an

application level company you have

choice and you have alternatives and the

the prices are coming down incredibly

quickly. Um, but there's also not so

much competition that the hyperscalers

and the you know labs like Anthropic and

OpenAI, there's not so much competition

that they are unable to raise money and

make these long-term investments. And so

I actually think we're in a in a pretty

good balance and and we're going to have

a lot of a lot of new companies and a

lot of growth among the the

hyperscalers.

I think that's that's about right. So

the terminology of sustaining versus

disruptive comes from uh uh the

innovator's dilemma. Uh and uh it's it's

this idea that uh whenever there's a new

technology trend, it sort of there's

this idea of a power curve. It starts as

a toy almost or something that doesn't

really work or captures the lower end of

the market. But as it sort of evolves,

uh it goes up the power curve and

eventually disrupts even the incumbents.

So originally the encompass don't pay

attention to it uh because it looks like

a toy and then eventually disrupts

everything and eats the entire uh sort

of market. Uh and so that that was true

of PCs. You know, when PCs came along,

the big main mainframe manufacturers did

not uh uh pay attention to it and and

initially it was like yeah, it's for

it's, you know, for kids or whatever. Uh

but we we have to run these large

computers or data centers or whatever,

but now even data centers are running on

PCs and so on. Um and and so PCs were

just a hugely disruptive uh force. Uh

but there are technologies that come

along and really benefit the incumbents

and really don't really benefit the uh

the uh new players, the startups. Uh I

think Adam's right. It's uh it's both.

Um and maybe for the first time it's

kind of both like a a huge technology

trend cuz the internet was hugely

disruptive. Um but but this time uh it

feels like it is an obvious supercharge

for the incumbents for the hyperscalers

for the large uh internet companies but

it also enables uh new business models

that uh that is perhaps counterposition

against the uh the existing existing

ones. Al although the the you know I

think what happened is everyone read

that book and everyone learned how to

not be disrupted. Uh for example Chad

GPT was fundamentally counterposition

against Google because uh Google had a

business that that was actually working.

Uh Chad GPT was seen as this uh

technology that hallucinates a lot and

creates a lot of bad information and

Google wanted to be trusted and so

Google had chatb internally. they didn't

release Gemini until like two years

after Chachup and Chachup had sort of

already won the like at least brand

recognition. Um and and so there there

was in a way open AI came out as a

disruptive technology uh but but now

Google realizes it's a disruptive

technology and kind of responds to it.

At the same time it was always obvious

that AI is going to benefit Google at

minimum. It's uh you know overview uh

search overview has gotten a lot better.

um all its uh you know workspace suite

is is getting a lot better with Gemini.

Uh their mobile phones, everything gets

better. So it's it seems like it's it's

both. Yeah, I I really agree. Like

everyone read the book and and that

changes what the theory even means

because you have

>> you've like all the all the public

market investors have read that book and

they

>> now are going to punish companies for

not adapting and reward them for

adapting even if it means they have to

make long-term investments. I think, you

know, all the the management leadership

of the companies have have read the book

and they're on top of their game. I

think also just like the people running

these companies are

in I I guess I would say smarter I think

than like the the companies from the

generation that that book was sort of

built on. and they're they're on at the

top of their game and they are a lot of

them are founder controlled and so they

can make it's it's easier for them to

sort of take a hit and and make these

these investments. So that's I actually,

you know, I think if if you had an

environment more like we had in say like

the '90s,

I think this would actually be more

disruptive than than the the current

hyper hyper competitive uh

>> world that we're in now.

One mistake that we as a firm have

reflected on over the past few years,

though of course I haven't been here for

more than just a few months, is this

idea of we've that we've passed on

companies because we they weren't going

to be the market leader or the or the

category winner. And thus we thought,

oh, you know, learning the lessons from

from web 2, you have to invest in the in

the category winner. That's where things

are going to consolidate. Value is going

to acrew over time. And um it seems so

you why do the the next foundation model

company if the first one already has a

has a head start. Um but it seems like

the market has gotten so much bigger

that in foundation models but also in

applications there's just multiple

winners and they're kind of you know

fragmenting you know and taking parts of

the market that are all venture scale.

I'm curious if this is a durable

phenomenon or but um it that seems just

one difference than than the web two era

is just more winners um across more

categories.

>> I think network effects are playing much

less of a role now than they did in the

web 2 era also and that that makes it

easier for competitors to get started.

There's still a scale advantage

>> because you know if you have more users

you can get more data. If you have more

users, you can raise more capital. But

that advantage is not it doesn't make it

absolutely impossible for a competitor

of smaller scale. It makes it hard, but

it's there there's definitely like room

for more winners than than there was

before. I I think another difference is

that people are seeing the value um so

strongly that they're willing to pay um

early on in maybe a way that they the

question with web two companies was how

are they going to make money you know

you were Facebook super early obviously

you know Google etc was like oh how are

they going to monetize and you know the

companies here are monetizing from from

the get-go you know your guys' companies

included

>> yeah yeah and the I I think with the

earlier generation of companies

the monetization kind of depended on

scale.

>> Like you couldn't build a good ad

business until you got to millions, tens

of millions of users. And now with

subscriptions, you can just charge right

away, I think, especially thanks to

things like Stripe that are making it

easier. Um, and so that that that's also

made it a lot more friendly to to new

entrance. There's there's also uh

questions of geopolitics like you know

it seems clear that we're not uh in this

um globalized era and perhaps it's going

to get much worse and so investing in

the foundation in the open AI of of

Europe might be a good idea and like

similarly China being an entirely

different different world and so there's

um sort of a geo aspect of it that

interesting

>> all of a sudden our geopolitics you know

nerdiness is helpful is is useful. Um,

Adam, you were talking about sort of

human knowledge. Did you see yourself

with Po kind of disrupting yourself in a

sense or or talk about the the the bet

that you you made with with PO and the

sort of evolution there?

>> You know, I I think we saw Po more as

just an additional opportunity than than

as disruption to to Kora. Um the the way

we got to it was we in early 2022 we

started experimenting with using GBD3 to

generate answers for Kora and and we

compared them to the the human answers

and sort of realized that they weren't

as good but what was really unique was

that you could instantly get an answer

to anything you wanted to ask about and

we realized it didn't need to be in

public. It actually was your preference

would be to to have it be in private and

so we felt like there was just a new

opportunity here to to let people chat

with with AI and in private.

>> Yeah. And it seemed like you were also

making a bet on how the different

players were going to that there was

going to be

>> Yeah. Yeah. So it was also a bet on

diversity of of model companies which

took a while to play out. But I think

now we're we're getting to the point

where there's there's a lot of models.

There's a lot of companies especially

when you go across modalities. You think

about image models, video models, audio

models. Um especially like the reasoning

research models are are sort of

diverging. Agents are starting to be

their own source of diversity. Um, so,

so we're lucky to to now be getting into

this world where there's there's sort of

enough diversity for a a general

interface aggregator to to make sense.

Um, but yeah, it was it was a bet early

on. We kind of

>> it's surprising actually that um even uh

not particularly technical consumers

actually do use multiple AIS. Uh like I

didn't expect that like you know people

only used Google. they never like looked

at Google and then Yahoo or like very

few people do it. But now you talk to

just average people and they'll say,

"Yeah, I use CHP most of the time, but

Gemini is much better at like these

types of questions." And it's like, "Oh,

interesting. The sophistication of

consumers have gone."

>> And even people saying that they have

different personalities and they, you

know, you know, sort of resonate with

Claude more, you know, or whatever. the

um I want to return back to this point

you said earlier Adam about you're kind

of talking about like dark matter about

how we're going to you know brute force

there's a lot of knowledge that people

have that's you know sort of not um sort

of categorized yet and it's not just

task of knowledge it's actually

knowledge that you could you know ask

them about and they could describe it

how you know because one question people

have with LMS is like how much we've

already trained the whole internet how

much more knowledge is there um and so

is it like 10x is it like a thousand

like what is sort of the what is kind of

intuitive sense of if we do brute force

it and build this whole you know machine

that gets all the knowledge out of

humans onto sort of you know a data set

that we can then you know implement

how do we think about the upside from

there

>> you know I think it's very hard to

quantify but

there's a massive industry developing

around getting human knowledge into for

the form where AI can use it so this is

things like scale AI I Surge

Merkore, but there there's a massive

>> long tale of other companies just

getting started. And

as you have, you know, as intelligence

gets cheaper and cheaper and more and

more powerful, the bottleneck, I think,

is increasingly going to be on the data

and what do you need to create that

intelligence? And so that's going to

cause this that's going to cause more

and more of this to happen. It might be

that people can make more and more money

by training AI. It might be that more

and more of these companies get started.

Um or it might be it might be that

there's there's other forms of it. But I

I think I think it's going to be sort of

like the economy is going to naturally

value whatever the AI can't do.

And

>> what is the framework for what it can't

like? what has meant a model for what it

can't do?

I don't, you know, you could you could

ask a an AI researcher, they they might

have a a better answer, but to me,

there's just information that's not in

the training set. And that is something

that's inherently going to be, you know,

going to be something AI can't do. There

will be, you know, the AI will get very

smart. It can do a lot of reasoning. It

could prove every math theorem at some

point. If it starts from, you know, some

axioms that you that you give it, but if

it doesn't know

how did this particular company solve

this problem 20 years ago, if that

wasn't in the training set, then only a

human who who knows that is going to be

able to answer that question. And so

over time, how do you see Kora um

interfacing with or like how are you

running these in parallel? How how do

you think about this?

>> Yeah, so I mean Kora, our focus is on

human knowledge and and letting people

share their knowledge and um that

knowledge may be helpful for you know

it's it's it's helpful for other humans

and it's it's also helpful for AI to to

learn from. um we have relationships

with some of the AI labs um and we're

going to sort of play the role core will

play the role that it is meant to play

in this ecosystem which is a as a a

source of of human knowledge. Um at the

same time AI is making core a lot

better. we've been able to make

uh major improvements in moderation

quality and in uh in ranking answers and

in uh just just improving the product

experience. So uh so it's gotten a lot

better by applying AI to it.

>> Yeah. And and talk talk about your

future as well. Obviously you know you

had this business for for a long time

you know focused on developers. Because

at one point you're targeting you know u

nonprofit. No

>> exactly the edtech market I believe you

did two or three million in revenue

reported and then you know recently

techrunch I know it's outdated but I

think it reported something like 150

million. I know it's since you've had

this incredible growth as as you've

shifted the the business model um and

and the customer segment. How do you

think about the the future of replet?

>> Um I think Kpathy uh recently said that

it's going to be the decade of agents.

Uh and I think that's absolutely right.

It's um uh as opposed to like prior

modalities of AI like when uh AI first

came to coding it was autocomplete with

co-pilot then it became sort of chat

with chat

then I think cursor innovated on this

composer modality which is like editing

like large chunks of uh files but that's

it. I think replet what Replet innovated

is is is is the agent um and the idea of

like not only editing code, provisioning

infrastructure like databases, doing

migrations,

um you know connecting to the cloud,

deploying uh having the entire debug

loop like executing the code, running

tests, um and so just like the entire

development life cycle loop happening

inside an agent and that's going to take

a long time to mature. So we're agent in

beta came September 2024 and it was the

first of its kind that did this both

code and infrastructure but it was you

know fairly janky didn't work very well

and then agent v1 around December

um it took another

um uh generation of models so you go

from claw 3.5 to 3.7 3.7 was the first

model uh that uh really knew how to use

a computer, a virtual machine. So,

unsurprisingly, it was the first also

computer use model. Um, and these things

have been moving together. Uh, and so

with every generation of models, we see

we find new capabilities. And, um, you

know, um, Agent V2 improved on autonomy

a lot. Agent V1 could run for like 2

minutes. Agent V2, uh, uh, ran for 20

minutes. Agent 3, we advertised it as

running for 200 minutes. just felt like

it should be symmetrical, but like it's

actually runs kind of indefinitely. Like

we've had users running it for 28 plus

hours. Wow.

>> Um, and the main idea there was that if

we put a verify on the loop. I remember

reading Deepseek uh a paper from Nvidia

about how they um used DeepSeek to write

CUDA kernels and they were able to run

Deepseek for like 20 minutes if they put

a verifier in the loop like being able

to run tests or something like that. And

I thought oh okay so what kind of

verifier can we put in the loop?

Obviously, you can put unit tests, but

unit test doesn't really capture whether

the app is working or not. So, we

started kind of digging into computer

use and whether computer use was going

to be able to test apps. Computer use is

very expensive and um it's actually kind

of still very buggy and like Adam talked

about that's going to be uh a big area

of improvement that'll unlock a lot of

applications. But we ended up building

our own framework with like bunch of

hacks and some some AI research and

repless computer use I think testing

models. I think one of the best. Um and

uh and once we put that into the loop

then you can put replet in high

autonomy. So we have an autonomy scale.

Uh uh you can you can you can choose

your autonomy level and then it just

writes the code goes and tests the

applications. If there's a bug it reads

the error log and like writes the code

again and and can go for for for hours.

And we've seen people build amazing

things by letting it run for for a long

time.

Now, that needs to continue to get

better. That needs to um to get cheaper

and faster. Uh so, it's not necessarily

a point of pride to run for a lot

longer. Like, it should be as fast as

possible. So, we're working on that. Um

a agent for there's a bunch of ideas

that are going to be uh coming out.

Agent 4, but one of the big things is

you shouldn't be just like waiting for

that one feature that you requested. you

should be able to work uh on a lot of

different features. So the idea of like

parallel agents is very interesting to

us. So you know you ask for a login page

but you could also ask for a stripe uh

checkout and and then you ask for an

admin dashboard. The AI should be able

to figure out how to paralyze all these

different tasks or some tasks are not

paralyzable but should also be able to

do merge across the code. So being able

to do collaboration across AI agents um

is very important and that way the

productivity of a single developer goes

up by a lot. right now even when you're

using clot code or cursor and others

that there isn't a lot of parallelism

going on but I think the next uh boost

in productivity is going to come from

sitting in front of programming

environment like replet and being able

to manage uh tens of agents maybe at

some point hundreds but you know at

least you know five 6 7 8 9 10 agents uh

all different all you know working in

different parts of your your product. I

also think that um UI and UX uh could

could use a lot of work in terms of um

right now um you're trying to translate

your ideas uh into this like textual

representation. I'm just like like a

PRD, right? The what product managers

do, right? Just product descriptions.

But product descriptions don't it's

really hard and you see it in a lot of

tech companies. it's really hard to

align on the exact features because it's

l language is fuzzy. And so I think

there's a there's a world in which

you're interacting with AI in a more

multimodal fashion. So open up uh like a

whiteboard and being able to draw and

like diagram with AI and and and really

work with it like you work with a human.

Uh and then um then the next stage of

that uh having uh like better memory

better memory inside the project but

also across project and perhaps having

different instantiations of replet agent

that uh you know that this this agent is

really good at like um Python data

science because um you know it has all

the information and skills and memories

of about my company what it's done in

the past. So I'll have a data analysis

like sort of rapid agent and I'll have

like a front-end replet agent and they

have memory over multiple projects and

over time and over interactions and

maybe they sit in your Slack like a like

a worker and you can like talk to them.

So again like I can I can keep going for

another 15 minutes about a road map that

could span like 3 to four to 5 years

perhaps. and but but this this agent

this agent phase that we're in is just

there's so much work to do and it's it's

it's going to be a lot of fun.

>> Yeah, it's a I was talking to one of our

mutual friends, one of the co-founders

of one of these uh you know big

productivity companies and he leads a

lot of their R&D and he's like man uh

during the week these days I'm not even

talking to humans anymore as much. I'm

just like it's just you know using all

all these agents to to build. So it's

living in the future to some degree is

already in the present.

>> There's something interesting about that

and that are people talking to each

other less at at companies

>> and is that a bad thing? Um

>> so it's a you know I think uh I I I'm

starting to think more about these

second order effects of of things like

that. um uh you know will it make it

awkward for like again the new grads I

feel so bad for them like uh you know if

if people are not sharing as much

knowledge between each other or it's

like

>> it's not culturally easy to go ask for

help because like you should be able to

use AI agents

>> uh there's something there's some

cultural forces that I think need to be

reckoned with.

>> Yeah, I think a lot of tough cultural

forces for zoomers these days. Yes. Um

let's gearing towards closing here. Um

obviously you guys are you know focused

on running your companies but to stay

current on the AI ecosystem. You you

guys also make angel investments as

well. Um where are you guys most uh most

excited? Um you we haven't talked about

robotics. Are you guys bullish on on

robotics in the in the near term or any

emerging categories or use cases or

spaces that you're looking to make more

investments in or you have made some? I

just think vibe coding generally is just

unbelievably

>> like high potential. Um just the idea

that all the you know this

>> you think underhyped even still

>> I think so I I I think

>> you know just opening up the potential

of software to the mainstream of you

know every everyone. I think that and

yeah and actually I think one reason I

think it's underhyped is that the tools

are still very far from what you can do

as a professional software engineer and

if you imagine that they're going to get

there and I think there's no reason why

they wouldn't might it'll take a few

years but um then it's like everyone in

the world is going to be able to

create any things that would have taken

a team of 100 professional software

engineers that's just going to massive

open up opportunities for for everyone.

So I think Replet is like a great

example of this, but I think it's also

going to that there will be cases other

than just like building applications

that that this also creates. By the way,

just on that note, if you were going to

Stanford or Harvard, you know, today

2025, just entering, would you major

again in computer science or just focus

on building something or

>> I think I would. I mean I I I

went to college starting in 2002 and it

was right after the dotcom bubble had

burst and there was a lot of pessimism

and I remember my um my roommate his

parents had told him like don't study

computer science even though that was

that was something he really liked. Um

and I just kind of did it because I I

liked it. And

I think that

I think that it's definitely like the

job market is worse than it was a few

years ago.

At the same time, I think having these

skills to understand the sort of

fundamentals of what's possible with

algorithms and data structures, I think

that actually really helps you in in

managing agents when when you're using

them. Um, and I I I'm guessing that it

will continue to be a valuable skill in

the future. I also think the other

question is like what else are you going

to study? And and every single thing you

could imagine, there's an argument for

why it's going to be automated.

>> So, I think you might as well study what

you enjoy and and and I think this is as

good as as anything.

>> Yeah. I um I think there's a lot to to

get excited by. One thing is maybe kind

of random, but like I get really fired

up to see like mad science experiments

like the uh Deepseek OCR that came out

the other day. Did you Did you see it?

It's It's wild where um correct me if

I'm wrong cuz I only looked at it

briefly, but basically you can um get a

lot more economical with a context

window if you like have a screenshot of

the text instead of the [ __ ] text.

>> Yeah, I'm not I'm not the right person

to be

>> correcting you on. than that. But like

it's there's there's definitely some

some really interesting things. Yeah, I

saw another thing on hacker news the

other day where um you know uh text

diffusion uh where someone made a text

diffusion model by instead of doing go

saying dnoising he would take like a

single BERT instance and like try to you

know mask different words and uh and

just predict like these different tokens

and um and so we have a lot of

components like I don't think people

think a lot about that you know we have

now the you know base pre-trained

models. We have the all these RL

reasoning models. We have the uh you

know encoder decoder models. We have

diffusion models. We have there's all

these different things like just like

you know you mix them in different ways.

>> Yeah.

>> Uh I feel like there isn't a lot of

that. I mean it' be great. It'd be great

if like a new research company just like

comes out and is like not trying to like

compete with OpenI and things like that

but instead uh is just trying to like

discover how to put these different

components together in order to create a

new flavor of these models.

>> Yeah. In crypto they talk about

composability and like mixing primitives

together and and AI maybe there needs to

be more exploation.

>> There's less playing around I found like

there is like I remember in the like

>> web 2.0

era when we were like playing around

with JavaScript what browsers could do

and what web workers could do whatever

there was a lot of like really

interesting weird experiments I mean

replet was born out of that the original

version of replet in open source pre pre

the company which my interest was like

can you compile C to JavaScript right

that was like one of the interesting

things that became WM by the time it was

uh mcriptton and it was like such a such

a nasty hack and um but I think there's

so much I think We're in an era of

Silicon Valley where it's like very uh

very getrich driven and that makes me a

little sad and that's partly why I moved

the company out of SF. I feel like the

culture in SF has has gotten maybe to

maybe I I I wasn't there but like during

the com era a lot of people talked about

how it's sort of like get rich fast or

the crypto thing. So I feel like there

needs to be a lot more tinkering and I

would love to see more of that and more

companies getting funded that are trying

to just do something a little more novel

even if it doesn't mean like it

fundamentally new new model.

>> Last question. Um Amad you've uh been

into consciousness for a long time. Are

are you bullish that we will um via some

of this AI work or just some you

scientific progress elsewhere make some

progress in understand in in uh you know

getting across this this hard problem or

you know something happened recently uh

which is interesting um uh cloud 4.5

uh seemed to have to become more aware

of its context length. So as it gets

closer to the end of the context, it

starts be becoming more economical with

tokens, it also it looks like its

awareness when it's being redteamed or

in a test environment like jumped

significantly. And so there's something

happening there that's quite

interesting. Now I think uh in terms of

you know the the question of of

consciousness it is still fundamentally

not a scientific question and there is a

sort of uh we've given up on trying to

make it scientific but I think it I

think this is also

uh the problem that I talked about with

all the energy going into LMS um uh no

one is trying to really think about the

true nature of intelligence, true nature

of uh consciousness. Um, and there's a

lot of really core core questions. Like

one of my favorite one is uh the uh

Roger Penrose um Emperor's New Mind

where he wrote a book about how everyone

in the sort of philosophy of mind space

uh and perhaps the larger scientific

ecosystem start thinking about the brain

in terms of a computer. And in that book

he tried to show that it fundamentally

is impossible for the brain to be uh a

computer because uh humans uh are able

to do things that touring machines

cannot do or Turing machines like

fundamentally get get stuck on such as

um uh you know just uh basic logic um

puzzles

uh that we're able to kind of detect,

but like there's no way to encode that

in a in a in a cheering machine. For

example, like this statement is false.

You know, those like old logic puzzles.

Um and uh anyways it's like a

complicated argument but uh if you read

that book or or many others uh there's

like a core strain of arguments in the

theory of mind about how uh computers uh

are fundamentally different from from

human intelligence and uh and so yeah I

I haven't really I've been very busy so

I haven't really updated my thinking too

much about that But

but I think there's there's a there's a

there's a huge field of study there that

is not being studied.

>> If you were a freshman uh entering

college today, would you study

philosophy?

>> I would do that. I would definitely

study philosophy of mind. I would

probably go into neuroscience. Uh cuz I

think those are the core questions that

are kind of become very very important

as AI kind of continues to see more of

jobs and economy and things like that.

>> That's a great place to wrap. I'm John.

Adam, thanks for coming on the podcast.

>> Thank you.

>> Thank you.

[Music]

Loading...

Loading video analysis...