Andrew Ng: Building Faster with AI
By Y Combinator
Summary
## Key takeaways - **Execution speed is key for startup success**: A strong predictor of a startup's odds of success is execution speed, and new AI technology is enabling startups to move much faster. (00:47, 01:00) - **Application layer holds biggest AI opportunities**: Despite hype on lower tech layers, the biggest opportunities for startups are at the application layer, as applications generate revenue that supports the underlying technologies. (01:38, 02:00) - **Agentic AI enables complex iterative workflows**: Agentic workflows allow AI to perform tasks iteratively, such as outlining, researching, drafting, and revising, leading to a much better work product than a single linear output. (03:18, 03:43) - **Concrete ideas drive faster development**: Vague ideas receive praise but are hard to build, while concrete ideas, specified in enough detail for an engineer to build, enable speed and faster validation or falsification. (04:52, 05:23) - **Rapid prototyping is 10x faster with AI**: AI coding assistance makes building quick prototypes at least 10 times faster by reducing integration needs, lowering requirements for reliability and security during initial testing. (10:44, 11:46) - **Learning to code empowers everyone with AI**: As AI tools make coding easier, more people should learn to code, as understanding how to instruct computers, even through AI, is a critical future skill. (15:21, 16:44)
Topics Covered
- Vague ideas get praise, but concrete ideas get built.
- AI makes code disposable and decisions reversible.
- Why everyone, including your CFO, should learn to code.
- Product management, not engineering, is the new bottleneck.
- AI doomerism is a self-serving promotional narrative.
Full Transcript
It's really great to see all of you.
What I want to do today since this is
build as startup school is share with
you some lessons I've learned about
building startups at AI fund. AI funds a
venture studio and we build an average
of about one startup per month. And
because we co-founded startups, we're in
there writing code, talking about
customers, design on features, detering
pricing. And so we've done a lot of reps
of not just watching others build
startups, but actually being in the
weeds, building startups with
entrepreneurs. And what I want to do
today is um share with you some of the
lessons I've learned building startups,
especially around this changing AI
technology and what it enables. And
it'll be focused on the theme of speed.
So it turns out that for those of you
that want to build a startup, I think a
strong predictor for startup's odds of
success is execution speed. And I
actually have a lot of respect for the
entrepreneurs and executives that can
just do things really quickly and new AI
technology is enabling startups to go
much faster. So what I hope to do is
share with you some of those best
practices which are frankly changing
every two to three months still to let
you get that speed that hopefully lets
you have a higher odds of success.
Before diving to speed, you know, a lot
of people ask me, hey Andrew, where are
the opportunities for startup? So this
is what I think of as the AI stack where
at the lowest level are the
semiconductor companies then the clouds
are hyperscalers built on top of that. A
lot of the AI foundation model companies
built on top of that. And even though a
lot of the PR excitement and hype has
been on these uh technology layers, it
turns out that almost by definition, the
biggest opportunities have to be at the
application layer because we actually
need the applications to generate even
more revenue so that they can afford to
pay the foundation cloud and
semiconductor technology layers. So for
whatever reason media and social media
tends not to talk about the application
layer as much but for those of you think
you're building startups almost by
definition the biggest opportunities
have to be there although of course the
opportunities at all layers of the
stack. One of the things that's changed
a lot over the last year um and in terms
of AI tech trends I if you ask me what's
the most important tech trend in AI I
would say is the rise of agentic AI and
about a year and a half ago when I
started to go around and give talks to
try to convince people that AI agents
might be a thing I did not realize that
around last summer a bunch of marketers
would get a hold of this term and use it
as a sticker and slap it on everything
in site which made it almost lose some
of meaning but I want to share with you
from a technical perspective why I think
agentic AI is exciting and important and
also opens up a lot more startup
opportunities. So, it turns out that the
way a lot of us use LMS is to prompt it
to have it gener output. And the way we
have an LM output something is as if
you're going to a human or in this case
an AI and asking it to please type out
an essay for you by writing from the
first word to the last word all in one
go without ever using backspace. And
humans, we don't do our best writing,
being forced to type in this linear
order. And it turns out neither does AI.
But despite the difficulty of being
forced to write in this linear way, um
our LMS do surprisingly well. With
agentic workflows, we can go to AI
system and ask it to please first write
an essay outline, then do some webs
research if it needs to and fetch uh
some web pages to put in their own
context, then write a first draft, then
read the first draft and critique it and
revise it and so on. And so we end up
with this iterative workflow where your
model does some thinking and some
research does some revision goes back to
do more thinking and by going around
this loop many times. Uh it is slower
but it delivers a much better work
product. So for a lot of t lot of
projects AI fund has worked on
everything from um pulling out complex
compliance documents to uh medical
diagnosis to reasoning about complex
legal documents. We found that these
agentic workflows are really a huge
difference between it working versus not
working. But a lot of the work that
needs to be done, a lot of the valuable
businesses to be built still will be
taking workflows existing or new
workflows and figuring out how to
implement them into these size of
agentic workflows. So just to update the
picture for the AI stack, um what has
emerged over the last year is a new
agentic orchestration layer that helps
application builders orchestrate or
coordinate a lot of calls to the
technology layers underneath. And the
good news is uh the orchestration layer
has made it even easier to build
applications. But I think the basic
conclusion that the application layer
has to be the most valuable layer of the
stack still holds true with a bias or
focus on the application layer. Let me
now dive into some of the best practices
I've learned for how startups can move
faster.
It turns out that um at AI fun we only
focus on working on concrete ideas. So
to me a concrete idea a concrete product
idea is one that's specified in enough
detail that an engineer can go and build
it. So for example if you say let's use
AI to optimize healthcare assets you
know that's actually not a concrete
idea. It's too vague. If you tell me
it's a very software to use AI to
optimize healthcare assets different
engineers would do totally different
things and because it's not concrete you
can't build it quickly and you don't
have speed. In contrast, if you had a
concrete idea like let's write software
to let hospitals, let patients book MR
machine slots online to optimize usage.
I don't know if this is a good or a bad
concrete idea. Actually business is
already, you know, doing this. But it is
concrete and that means engineers can
build it quickly. If it's a good idea,
you find out it's not a good idea, you
will find out. But having concrete ideas
buys you speed. Or someone to say, let's
use AI for email personal productivity.
Too many interpretations of that. That's
not concrete. But if someone says could
you build an app Gmail integrate the
automation that use let's use the right
prompts right filter entire emails that
is concrete I could you know I could go
build that this afternoon. So
concretness buys you speed and the
deceptive thing for a lot of
entrepreneurs is the vague ideas tend to
get a lot of kudos. If you go and tell
all your friends we should use AI to
optimize the use of healthcare assets.
Everyone will say that's a great idea.
But it's actually not a great idea at
least in the sense of being something
you can build. Uh it turns out when
you're vague, you're almost always
right. Uh but when you're concrete, you
may be right or wrong. Either way is
fine. We can discover that much more
fast, which is what's important for SA.
In terms of executing on concrete ideas,
I I I find that at AI fun, I ask my team
to focus on concrete ideas because um a
concrete idea gives clear direction and
the team can run really fast to build it
and either validate it, prove it out or
falsify and conclude it doesn't work.
Either way is fine. So we can do that
quickly and it turns out that finding
good concrete ideas usually requires
someone could be you could be a subject
matter expert thinking about a problem
for a long time. Uh so for example
actually before before you know starting
Corsera um I spent years right thinking
about online education talking to users
holding my own intuitions about what
would make a good edtech platform and
then after that long process I think YC
sometimes calls it wondering the idea
maze but after thinking about it for a
long time you find that the guts of
people that have thought about this for
a long time can be very good about
rapidly making decisions as in after
you've thought about this talked to
customers and so on for a long time if
you ask this expert, should I build this
feature or that feature? You know, the
gut, which is an instantaneous decision,
uh, can be actually a surprisingly good
proxy. It can be surprisingly good
mechanism for making decisions. And I
know I work on AI, you might think I'll
say, oh, we need data. And of course, I
love data, but it turns out getting data
for a lot of startups is actually slow
mechanism for making decisions. And a
subject matter expert with a good gut is
often a much better mechanism for making
a speedy decision. And then one other
thing for many successful startups at
any moment in time you're pursuing one
very clear hypothesis they building out
and trying to sell China Valley of
Hospice. Um and a startup doesn't have
resources to hedge and try 10 things at
the same time. So pick one go for it and
if data tells you to lose faith in that
idea that's actually totally fine. Just
pivot on a dime to pursue a totally
different concrete idea. So that's what
often feels like an AI fund. We're
pursuing one thing doggedly with
determination until the world tells us
we were wrong then change and pursue a
totally different thing with equal
determination and equal doggedness. And
one other pattern I've seen, if every
piece of new data causes you to pivot,
it probably means you're starting off
from two weaker a base of knowledge,
right? If every time you talk to a
customer, you totally change your mind,
probably means you don't know enough
about that sector yet to have a really
high quality concrete idea and uh
finding someone that's thought about a
subject for longer may get you on to
better path in order to go faster. The
other thing I often think about is the
built feedback loop which is rapidly
changing when it comes to how we build
with AI coding assistance. So when
you're building a lot of applications,
one of the biggest risks is custom
acceptance, right? A lot of startups
struggle not because we can't build
whatever we want to build, but because
we build something and it turns out
nobody cares. And so for a lot of the
way uh you know I build startups
especially applications less so deep
tech less so technology startups but
definitely application startups is we
often build software so this is an
engineuring toss and then we will get
feedback from users and this is a
product management toss and then we'll
go back you know then based on the user
feedback we'll tweak our views on what
to build go back to write more software
and we go around this loop many many
times iterate toward product market fit
and it turns out that with AI coding
assistance which Andre talked about as
well um rapid engineering is becoming
possible in a way that just was not
posing much more feasible. So the speed
of engineing is going up rapidly and the
cost of engineing is also going down
rapidly. This changes the mechanisms by
which we drive startups around this
loop. When I think about the software
that I do, I maybe put into two major
buckets. Sometimes I'll build quick and
dirty prototypes to test an idea. You
say build a new customer service
chatbot. Let's build AI to process legal
documents whatever build a quick and
dirty prototype to see if we think it
works. The other type of software where
I do is write maintain production
software maintain legacy software but
these massive production ready code
bases depending on which analysts report
you trust. It's been hard to find very
rigorous data on this. You know when
writing production quality code maybe
we're 30 to 50% faster with AI systems.
Hard to find a rigorous number. maybe
he's plausible to be but in terms of
building quick and dirty prototypes
we're not 50% faster I think we're
easily 10 times faster maybe much more
than 10 times faster and there are a few
reasons for this uh when you're building
standalone prototypes there's less
integration with legacy software
infrastructure legacy data needed um
also the requirements reliability even
scalability even security are much lower
and I know I'm not supposed to tell
people to write insecure code right
feels like the wrong thing to say, but I
routinely go to my team and say, "Go
ahead, write insecure code." Because if
this software is only going to run on
your laptop and you don't plan to
maliciously hack your own laptop, it's
fine to have insecure code, right? But
of course, after it seems to be working,
please do make it secure before you ship
it to someone else. And you know, like a
leaking PI, leaking sense data that that
is, you know, very damaging. So before
you ship it, make it secure and
scalable, but they're just testing it.
It's fine. And so I find increasingly
startups will systematically pursue
innovations by building 20 prototypes to
see what works, right? Uh because I I I
know that there's some angst in AI. A
lot of proof of concepts don't make to
production. But I think by driving the
cost of a proof of concept low enough,
it's actually fine if lots of proof of
concepts don't see the light of day. And
I know that um the mantra move fast and
break things got a bad rep because you
know it broke things. And some teams
took away from this that you should not
move fast, but I think that's a mistake.
I tend to tell my teams to move fast and
be responsible. And I think they
actually lots of ways to move really
quickly while still being responsible.
And in terms of the AI assistance uh
coding landscape, I think was it three
four years ago code autocomplete right
popularized by GitHub copilot and then
there was a cursor windserve generation
of AI enabled ids which great use winds
and cursor quite a lot um and then
starting I don't know six seven months
ago uh there started to be this new
generation of highly agentic coding
assistants uh including that she's using
o3 a lot for coding um cloud code is
fant Fantastic. Since quad 4 release,
it's become and and ask me again in a
few months, I may use something
different. But the tools are evolving
really rapidly, but I think uh cloud
codeex this is a new generation of
highly agentic coding assistance that is
making developer productivity keep on
growing. And the interesting thing is if
you're even half a generation or one
generation behind actually makes a big
difference compared to if you're on top
of the latest tools and I find my team
is taking really different approaches to
software engineing now compared to even
three or six months ago. One surprising
thing is we we're used to thinking of
code as this really valuable artifact
because it's so hard to create but
because the cost of software engine is
going down code is much less of a
valuable artifact as it used to. So I'm
on teams where you know we've completely
rebuilt a codebase three times the last
month right because it's not that hard
anymore to just completely rebuild a
codebase pick a new data schema is fine
because the cost of doing that has
plummeted some of you may have heard of
Jeff Bezos's terminology of a two-way
door versus a one-way door. A two-way
door is decision that you can make. If
you change your mind, come back out, you
know, reverse it relatively cheaply.
Whereas a one-way door is you make a
decision and you change your mind is
very costly or very difficult to
reverse. So choosing the software
architecture of your tech stack used to
be a one-way door. Once you built on top
of a certain tech stack, you set a
database schema, really hard to change
it. So that used to be a one-way door. I
don't want to say it's totally a two-way
door, but I find that um my team will
more often build on a certain tech stack
a week later, change your mind, let's
throw the code base away and redo it
from scratch on a new tech stack. I I
don't want to overhype it. We don't do
that all the time. There are still costs
to redoing that. But I find my team is
often rethinking what is a one-way door
and what's now a two-way door because
the cost of software engineering is so
much lower now. And maybe going a little
bit beyond software engineering, I I I
feel like this actually a good time to
empower everyone to build of AI. Uh over
the last year, a bunch of people advised
others not to learn to code on the
grounds of AI were automated. I think
we'll look back on this as some of the
worst career advice ever given because
as better tools make software engineing
easier, more people should do it, not
fewer. So when many decades ago the
world moved from punch calls to keyboard
and terminal that made coding easier.
When we moved from assemblies high level
languages like cobalt um there actually
people arguing back then that now we
have cobalt we don't need programmers
anymore like people actually wrote
papers to that effect but of course that
was wrong and programming languages made
it easier to code and more people learn
to code text IDs ID is the AI coding
assistant um and as coding becomes
easier more people should learn to code.
I have a controversial opinion which is
uh I think actually it's time for
everyone of every job role to learn to
code and in fact on my team you know my
CFO my head of talent my recruiters my
front desk uh uh person all of them know
how to code and I actually see all of
them performing better at all of their
job functions because they can code and
I think um I'm probably a little bit
ahead of the curve probably most
businesses are not there yet but in the
future I think we empower everyone to
code a lot of people can be more
productive I want to share with you One
lesson I learned as well on on why we
should have people learn to do this
which is um when I was teaching
generative VI for everyone on Corsera,
we needed to generate background art
like this uh using midjourney and you
know one of my team members uh new art
history and so he could prompt
midjourney with the genre the palette
the artistic inspiration had a very good
control over the images he generated. So
we end up using all of Tommy's generated
images. Whereas in contrast, I don't
know art history. And so when I prompt,
you know, image generation, I could
write, please make pretty pictures of
robots for me, right? And and I could
never have the control that my
collaborators could. And so I couldn't
generate as good images as he could. And
I think with computers, one of the most
important skills of the future is the
ability to tell a computer exactly what
you want. So they'll do it for you. And
will be people that have that deeper
understanding of computers that will be
able to command a computer to get the
outcome you want. And learning to code,
not not that you need to write the code
yourself. Steer AI to code for you seems
like it will remain the best way to do
that for a long time. with software
engineering becoming much faster. The
other interesting dynamic I'm seeing is
that the product management work getting
user feedback deciding what features to
build that is increasingly the
bottleneck and so I'm seeing very
interesting dynamics in multiple teams
over the last year a lot more of my
teams have started to complain that
their bottlenecks on product engineering
and design because the engineers have
gotten so much faster some interesting
trends I'm seeing three four five years
ago Silicon Valley used to have these
slightly suspicious rules of thumb but
nonetheless rules of thumb will have 100
p.m. to four engineers or 1 PM to seven
engineers was this like PM product
manager to engineering ratio right which
should take with a grain of salt but it
was typical of a 1 PM to six seven
engineers and with engineers becoming
much faster I don't see product
management work designing what to build
becoming faster at the same speed
engineers I'm seeing this ratio shift so
literally yesterday one of my teams came
to me and for the first time when we're
planning headcom for a project this team
proposed to me not at 1:00 PM to four
engineers but to have 1 PM to 0.5
engineers. So the team actually proposed
to me I still know no this is a good
idea for the first time in my life that
I saw you know managers proposed to me
having twice as many PMs as engineers
was a very interesting dynamic. I I
still don't know if this proposal I
heard yes is a good idea but I think
it's a sign of where the world is going
and I find as PMs that can code or
engineers with some product instincts
often end up doing better. The other
thing that I found important for startup
found for startup leaders is because
engineing is going so fast. If you have
good tactics for getting rapid feedback
to shape your perspective what to build
faster that helps you get faster as
well. So um I'm going to go through a
portfolio of tactics for you know
getting product feedback to keep shaping
what you will decide to build. And we're
going to go through a list of the faster
maybe less accurate the slower more
accurate tactics. So the fastest tactic
for getting feedback is look at the
product yourself and just go by your
gut. And if you're a subject matter
expert, this is actually surprisingly
good you if you know what you're doing.
A little bit slower is go ask three
friends or teammates to get feedback to
play with your product and get feedback.
Um little bit slower is ask three to 10
strangers you know for feedback. Um, it
turns out for when I built products, one
of the most important skills I think I
learned was how to sit in the coffee
shop, how to sit in a when it's
traveling, when I travel, I often sit in
the hotel lobby. It turns out learn to
spot places of high foot traffic and
very respectfully, you know, grab
strangers and ask them for feedback on
whatever I'm building. This used to be
easier when I was less known. When when
people recognize you, it's a little bit
more awkward. I found that um I've
actually sat with teams the hotel lobby
very high foot for traffic and you know
very respectfully ask strangers hey
we're building this thing do you mind
taking a look oh and I actually learned
in a coffee shop there a lot of people
working a lot of people don't want to be
working so we give them excuse to be
distracted they're very happy to do that
too but I've actually kind of made tons
of product decisions in a hotel lobby or
a coffee shop with collaborators just
just just like that send prototypes to
100 testers you if you have access to
logic group of users and prototype to
more users and these are these get to be
slow and slower tactics and I know
Silicon Valley you know we like to talk
about AB testing of course I do a ton of
AB testing but contrary to what many
people think AB testing is now one of
the slowest tactics in my menu because
it's just slow to ship it yeah it depend
on how many users you have right so and
then uh the other thing is um as you use
anything but the first tactic some teams
will look at the data they make a
decision but the missing piece is When I
AB test something, um, I don't just use
result of AB test to pick product A or
product B. My team will often sit down
and look carefully at the data to hone
our instincts to speed up to improve the
rate. I wish we're able to use the first
tactic to make high quality decision.
Often sit down and think, gee, I
thought, you know, this product name
will work better than her product name.
Clearly, my mental model the users
wrong. to really sit down and think to
update our mental model using all of
that data to improve the quality of our
guts on how to make product decisions
faster. That turns out to be really
important. All right, so talked about um
concrete ideas, speed up engineering,
speed up product feedback. This is one
last thing I want to touch on which is
I've seen that understanding AI actually
makes you go faster. Um and and here's
why. As a AI person, maybe I'm biased to
be pro AI, but I want to share you why.
So it turns out that when it comes to
mature technology like mobile, you know,
many people have had smartphones for a
long time. We kind of know what a mobile
app can do, right? So many people
including nontechnical people have good
instincts about what a mobile app can
do. If you look at mature job roles like
sales, marketing, HR, legal, they're all
really important and all really
difficult. But you know, there are
enough marketers that that have done
marketing for long enough and the
marketing tactics haven't changed that
much in the last year. So there are a
lot of people that are really good in
marketing and it's really important
really hard but that knowledge is
relatively diffused because you know the
knowledge of how to do HR like it hasn't
changed dramatically you know in the
last six months but AI is emerging
technology and so the knowledge of how
to do AI really well is not widespread
and so teams that actually get it that
understand AI do have a advantage over
teams that don't whereas if you need if
you have an HR problem you can find
someone you know that knows how to do it
well probably but If an AI problem,
knowing how to actually do that could
put you ahead of other companies. So
things like what accuracy can you get
for a customer service chatbot? You
know, should you prom fine tune a
workflow? Um how do you get a voice out
to low latency? There a lot of these
decisions that if you make the right
technical decision, you can like solve
the problem in a couple days. They make
the wrong technical decision, you could
chase a blind alley for three months,
right? And and one one thing I've been
surprised by, it turns out if you have,
you know, two possible architecture
decisions, it's one bit of information.
It feels like if you don't know the
right answer, at most you're twice as
slow, right? One bit, you know, try
both. It feels like one bit of
information can at most buy you a 2x
speed up. And I think in some
theoretical sense that is true. But what
I see in practice, if you flip the wrong
bit, you're not twice as slow. You spend
like 10 times longer chasing a blind
alley. which is why I think going in to
have this right technical judgment, it
really makes startups go so much faster.
The other reason why I find staying on
top of AI really helpful for startups is
um over the last two years we have just
had a ton of wonderful genai tools or
genai building blocks right partial list
but prompting workflows evals guardrails
rack voice act async programming lots of
ETL embeddings fine-tuning graph DB how
to integ
models there's a long and wonderful list
of building blocks that can quickly
combine to build software that no one on
the planet could have built, you know,
even a year ago. And this creates a lot
of new opportunities for starters to
build new things. So when I learned
about these building blocks, this is
actually a picture that I have in mind.
If you own one building block, like you
have a basic white building block, yeah,
you can build some cool stuff. Maybe you
know how to prompt. So you have one
building block, you build some amazing
stuff. But if you get a second building
block like you also know how to build
chat bots. So you have a white Lego
brick and a black Lego brick, you can
build something more interesting. Um if
you acquire a blue building brick as
well, you can build something even more
interesting. Get few red building bras,
maybe a little yellow one, more
interesting, get more building bras, get
more building bras, and very rapidly the
number of things you comb combine them
to into grows kind of combinatorily or
grows exponentially. And so knowing all
these wonderful building blocks lets you
combine them in much richer combination.
One of the things that deep learn does
so I actually take a lot of deep learn
courses myself you know to because work
with great we work with I think like
pretty much all the leading AI companies
in the world and cert and and um and uh
try to hand out building blocks. Um but
when I look at the deep learning course
catalog this is actually what I see. And
whenever I take these courses to learn
these building blocks, I feel like I'm
getting new things that can combine to
form kind of combinatorally or
exponentially more software applications
that were not possible just one or two
years ago. So just to wrap up, this is
my last slide. I then want to take
questions if if y'all have any. I find
that there are many things that matter
for startup, not just speed. But when I
look at the startups that AI fund is
building, I find that the management
team's ability to execute at speed is
highly correlated with its odds of
success. And some things we've learned
to get you speed is, you know, work on
concrete ideas. Um uh it's got to be
good concrete ideas. I find that as a as
executive, I'm judged on the speed and
quality of my decisions. Both do matter,
but speed absolutely matters. rapid
entrying with AI coding assistance makes
you go much faster but that shifts the
bottleneck to getting user feedback on
the product decisions and so having a
portfolio of tactics to go get rapid
feedback and if you haven't learned to
go to coffee shop and talk to strangers
it it's not easy but then just just be
respectful right just be respectful of
people that's actually very valuable
skill for entrepreneurs to have I think
and I think also um staying on top of
the technology buys you speed all right
with that let me thank Thank you very
much.
[Applause]
Happy questions.
As AI advances, do you think it's more
important for humans to develop the
tools or learn how to use the tools
better? Like how can we p position
ourselves to remain essential in a world
where you know intelligence is becoming
democratized? I feel like AGI has been
overhyped and so for a long time
there'll be a lot of things that humans
can do that AI cannot and I think in the
future the people that are most powerful
are the people that can make computers
do exactly what you want it to do and so
I think staying on top of the tools some
of us will build tools sometimes but
there were a lot of other tools others
will build that we can just use but so
people that know how to use AI to get
computers to do what you want it to do
will be much more powerful, not worry
about people running out of things to
do, but um people that can use AI will
be much more powerful than people that
don't.
Hey, so well first of all uh thank you
so much. I have a huge respect for you
and I think that you are true
inspiration for a lot of us. My question
is about uh the future of compute. So as
we move towards uh more powerful more
powerful AI, where do you think that
comput is heading? I mean we see people
saying let's ship GPUs to space. Some
people talking about nuclear power data
centers. What do you think about it?
There's something I'm debating what I
wanted to say in response to the last
question about kind of AGI about maybe
I'll answer this and a little bit the
last question. So it turns out there's
one framework you can use for deciding
what's hype and what's not hype. I think
over the last two years there's been a
handful of companies that um hyped up
certain things for promotional PR
fundraising influence purposes. And
because AI was so new, um, handful of
companies got away with saying almost
anything without anyone fact-checking
them because the technology was not
understood. So, one of my mental filters
is there's certain hype narratives that
make these businesses look more powerful
that's been amplified. Um, and so, for
example, this idea that um, AI is so
powerful, we might accidentally lead to
human extinction. That's just
ridiculous. But it is a hype narrative
that made certain businesses look more
powerful and it got you know ramped up
and actually helped certain businesses
fundraising goals. AI is so powerful
soon no one will even have a job
anymore. Just not true, right? But again
that made these businesses look more
powerful got hyped up or we are so
powerful so when the hype narrative
we're so powerful that by training a new
model we will casually wipe out
thousands of startups. That's just not
true. Yes, Jasper ran into trouble.
Small number of companies got wiped out.
But it's not that easy to casually wipe
out thousands of startups. AI needs so
much electricity. Only nuclear power is
good enough for that. You know, that
wind solar stuff not that's just not
true. So, I think a lot of this um GPUs
in space, you know, I don't know. It's
like um go for it. I think we have a lot
of room to run still for terrestrial
GPUs. Uh yeah, but but I think uh uh
some of these hype narratives are have
been amplified that that I think uh are
a distortion of what what actually will
be done.
There's a lot of hype in um AI and how
and nobody's really certain about how
we're going to be building the future
with it. But what are some of the most
dangerous biases or overhyped narratives
that you've seen people talk about or
get uh poisoned by that they end up
running with that we should try to avoid
or be more aware of and allow us to have
a more realistic view as we are building
this future.
So I think the dangerous AI narrative
has been overhyped. Uh AI is a fantastic
tool, but like any other powerful tool
like electricity, lots of ways to use it
for beneficial purposes. Also some ways
to use it in harmful ways. I find myself
not using the term AI safety that much.
Um not because I think we we should
build dangerous things, but because I
think safety is not a function of
technology, it's a function of how we
apply it. So like electric motor you
know you can't the maker of electric
motor can't guarantee that no one will
ever use it from unsafe downstream toss
like use electric motor can be used to
build a Dallas machine electric vehicle
can be used to build a smart bomb but
the electric motor manufacturer can't
control how be used downstream. So
safety is not a function of the electric
motor as a function of how you apply it
and I think the same thing for AI. AI is
neither safe nor unsafe. It is how you
apply it that makes it safe or unsafe.
So instead of thinking about AI safety,
I often think about responsible AI
because it is how we use it responsibly
hopefully or irresponsibly that
determines whether or not what we build
with AI technology ends up being harmful
or beneficial. And I feel like sometimes
that the really weird corner cases that
get hyped up in the news. I think just
one or two days ago there was a Wall
Street Journal article about AI losing
control of AI or something. And I feel
like that article took uh corner case
experiments run in a lab and you know
sensationalized it in a way that I think
was really disproportionate relative to
the lab experiment that was being run
and unfortunately technology is hard
enough to understand that many people
don't know better and so these hype
narratives do keep on getting amplified
um and I feel like this has been used as
a weapon against open source software as
well right which is really unfortunate.
Thank you for your work. I think your
impact is remarkable. Uh my question is
um as aspiring founders, how should we
be thinking about business in the world
where anything can be disrupted in a
day? Whatever great mode, product or
feature you have can be replicated with
VIP code and competitors in basically
hours.
It turns out when you start a business,
there are a lot of things to worry
about. The number thing I worry about is
uh are you building a product that users
love? Um it turns out that when you
build a business there are lots of
things to think about the go to market
channel competitors technology mode all
that is important but if I were to have
a singular focus on one thing it is are
you building a product that users really
want until you solve that you know is
very difficult to build a valuable
business. After you solve that the other
questions do come to play. Uh, do you
have a channel to get to customers? What
is pricing long-term? What is your moat?
I find that moes tend to be overhyped.
Actually, I find that more businesses
tend to start off with a product and
then evolve eventually into a moat. But
consumer products rand is somewhat more
defensible. Um, and if you have a lot of
momentum, it becomes harder to catch
you. But enterprise products sometimes
if you have a uh maybe mo is more of a
consideration if they're channels that
are hard to get into enterprises. So I
think um sorry when when AI fund looks
at businesses we actually wind up doing
a fairly complex analysis of these
factors and writing a you know two to
six page narrative memo to analyze it
before we decide whether or not to
proceed it or not. And and I think um uh
all of these things are important, but I
feel like at this moment in time, the
number of opportunities, meaning the
amount of stuff that is possible that no
one's built yet in the world, seems much
greater than the number of people with
the skill to build them. So definitely
at the application layer, it feels like
there's a lot of white space for new
things you can build that no one else
seems to be working on. And I would say,
you know, focus on building a product
that people want, that people love. Um,
and then figure out the rest of it along
the way. Although this important figure
along the way.
Uh, hi professor. Uh, thanks for your
wonderful speech. Uh, I am a
Andagramress researcher from Stanford
and I think your uh, metaphor in your
speech is very interesting. You said the
uh, current AI tools are like bricks and
are be uh, and can be built upon
accumulation. However, so far it is
difficult to see the accumulative
functional expansion of the uh
integration of AI tools because they
often align on the stacking of functions
based on uh intent distribution and are
accompanied by dynamic problems of
tokens and time overhead. So um which is
uh which is different from static
engineering. So what do you think will
be the perspective of a possible agent 2
accumulation effect in the future? But
hey just just some quick remarks to that
right you mentioned agent uh OM token
cost my most common advice to developers
is to first approximation just don't
worry about how much tokens cost only a
small number of startups are lucky
enough to have users use so much of your
product that the cost of tokens becomes
a problem it can become a problem I've
definitely been on a bunch of teams
where the cost you know users like our
product and we started to look at our
right geni uh uh bills and it was
definitely climbing in a way that really
became a problem. But it's actually
really difficult to get to a point where
your token usage costs are a problem.
And for the teams I'm on where we were
lucky enough that users made our token
cause a problem, we often had engine
solutions to then bend the cursor and
bring it back down through prompting
fine-tuning USDs by optimize or
whatever. And then what I'm seeing is
that I'm seeing a lot of agentic
workflows that actually integrate a lot
of different steps. So for example, if
you build a customer service chatbot,
we'll often have to use prompting, maybe
optimize some of the results in DSPI,
build evals, build guard rails, maybe
the customer service chatbot needs rag
up part of the way to get information to
feedback to the user. So I actually do
see these things grow. But one tip for
many of you as well is I will often
architect my software to make switching
between different building block
providers relatively easy. So for
example, um have a lot of products that
build on top of OM, but sometimes you
point to a specific product and ask me
which OM are we using? I honestly don't
know because we built up evals and when
there's a new model that's released,
we'll quickly run evals to see if the
new model is better than the old one.
And then you'll just switch to the new
model if the new model does better on
evals. And so the model we use week by
week, you know, sometimes our engines
will change it without even bothering to
tell me because the eval show the new
model works better. So it turns out that
switching cost for foundation models is
relatively low and we often architect
our software. Oh, AI suite is open
sourcing that my friends and I worked on
to make switching easier. Um, switching
cost for the orchestration platforms is
a little bit harder. Uh but I find that
preserving that flexibility in your
choice of building blocks often let you
go faster even as you're building more
and more things on top of each other. Um
so hope that helps.
Thank you so much.
In the world of education in AI, there
are two paradigms mostly. So one is AI
can make teachers more productive. Uh
automating grading and automating
homeworks. But another school of thought
is that there'll be personal tutors for
every student. So every student can have
one tutor that gets feedback from an AI
and gets personal questions from them.
So how do you see these two paradigms
converge and how would education look
like in the next five years?
I think everyone feels like a change is
coming in edtech but I don't think the
disruption is here yet. I think a lot of
people are experimenting with different
things. So you know Corsera has Corsera
coach which actually works really well.
Um deep learn is more focused on
teaching AI also has some built-in chat
bots. Um a lot of teams experiment of
autograding. Oh there's an avatar with
me on the deep learn website you can
talk to if you want. Uh deep learn.ai.
And then I think for some things like
language learning with you know speak
Dolingo that has become clearer some of
the ways AI would transform it for the
broader educational landscape the exact
ways that AI would transform it I see a
lot of experimentation I think what key
learning which I've been doing some work
with is doing is is very promising for
K12 education but I think uh what I'm
seeing is frankly tons of
experimentation but the final end state
is still not clear. I do think education
will be hyperpersonalized. Uh but that
workflow is an avatar, is a text
chatbot, what's the workflow? I think um
I feel like the hype from a couple years
ago that with AGI soon and it will be
all so easy. That was hype. The reality
is work is complex, right? teachers,
students, people do really complex
workflows and for the next decade we'll
be looking at the work that needs to be
done and figuring out how to map it to
agentic workflows and education is one
of the sectors where this mapping is
still underway but it's not yet mature
enough to the point where the end state
is clear. So I think I think we should
all yeah just keep working on it.
All right. All right. Thank you so much
Andrew.
Thank you. Uh hey my question is I think
AI uh it has a lot of great potential
for good but there's also a lot of
potential for bad consequences as well
such as exacerbating economic inequality
and things like that and I think a lot
of our startups here while they'll be
doing a lot of great things will also be
you know just by virtue of their product
be contributing to some of those
negative consequences. So I was curious
how do you think you know us as AI
builders should kind of balance our uh
product building with also the potential
societal downsides of some AI products
and essentially how can we uh both move
fast and be responsible as you mentioned
in your talk
look in your heart and if fundamentally
what you're building if you don't think
it'll make people at large better off
don't do it right I I know it sounds
simple but actually really hard to do in
the moment but AI fund we've killed
multiple projects projects not on
financial grounds but on ethical grounds
where there are multiple projects we
looked at the economic case is very
solid but we said you know what we don't
want this exist in the world and we just
killed it on that basis so I hope more
people will do that and then I worry
about uh bring everyone with us one
thing I'm seeing is um people in all
sorts of job roles that are not
engineering are much more productive if
they know AI than if they don't and so
for example on my marketing team my
marketers they know how the code.
Frankly, they they were running circles
around the ones that don't. So, then
everyone learned to code and then they
got better. But I feel like um trying to
bring everyone with us to make sure
everyone is empowered to build with AI.
That'll be an important part of what all
of us do, I think.
Um I'm one of your big fans and thank
you for your online courses. Your
courses make the deep learning like uh
much more accessible to the world. And
my question is also about education. uh
as AI becomes more powerful and
widespread, there seems to be a growing
gap between what can actually do and
what people perceive it. So what do you
think about like is it important to
educate the general public about deep
learning stuff and not only like uh
educate those technical people and make
people understand more what really uh
what AI really do and how it works.
I think that knowledge will diffuse deep
learn AI we want to empower everyone to
build with AI. So we're working on it.
Many of us work on it. I'll just tell
you what I think is the main d I think
there are maybe two dangers. One is if
you don't bring people with us fast
enough, I hope we'll solve that. There's
one other danger which is um it turns
out that if you look at the mobile
ecosystem, mobile phones, it's actually
not that interesting. And one of the
reasons is there are two gatekeepers,
Android and iOS. And unless they let you
do certain things, you're not allowed to
try certain things on mobile. And I
think this, you know, hampers
innovators. These dangers of AI have
been used by certain businesses. They're
trying to shut down open source because
a number of businesses that love to be a
gatekeeper to large scale foundation
models. So I think hyping up dangers,
supposed false dangers of AI in order to
get regulators to pass laws like the
proposed SP 1047 in California, which
thank goodness we shut down, would have
put in place really burdensome regry
requirements that don't make anyone
safer, but would make it really
difficult for TS to release open source
and open weight software. So one of the
dangers to inequality as well is if
these regulatory you know awful regry
approaches and I've been in the room
where some of these businesses said
stuff to regulators that was just not
true. So I think that um some of these
arguments the danger is if these
regulatory proposals succeed and end up
siphoning regulations leaving us with a
small number of gatekeepers where
everyone needs the permission of a small
number of companies to fine-tune the
model prompt in a certain way that's
what will cipher innovation and prevent
the diffusion of this information to let
lots of startups you know build whatever
they want responsibly but the freedom to
innovate so I think so long as we um
prevent this line of attack on open
source open weight models from
succeeding and we we've made good
progress but the threat is still there
then I think eventually we'll get to the
diffusion of knowledge and we can
hopefully then bring everyone with us
but this fight to protect open source
we've been winning but the fight is
still on and we still have to keep up
that work to to protect open source
thank you all very much it's wonderful
to see my thank you
Loading video analysis...