TLDW logo

OpenAI’s CPO on how AI changes must-have skills, moats, coding, startup playbooks, more | Kevin Weil

By Lenny's Podcast

Summary

## Key takeaways - **AI models are the worst you'll ever use**: The AI models you are using today are the worst you will ever use for the rest of your life. Every couple of months, computers can do something they've never been able to do before, requiring a complete rethinking of what's possible. [00:02] - **Pace and evolving tech define OpenAI's work**: At OpenAI, the pace is significantly faster than previous companies due to the rapidly evolving technology. Unlike stable databases, AI capabilities change every two months, demanding constant adaptation and a different approach to product development. [16:25] - **Evals are critical for AI product development**: Writing effective evals, or tests for AI models, is becoming a core skill for product managers. These tests gauge a model's proficiency in areas like creative writing or scientific understanding, crucial for building products where model performance dictates success. [18:45] - **Chat is an enduring AI interface**: Chat remains a versatile and effective interface for interacting with AI due to its universality and flexibility. It mirrors human communication, allowing for open-ended, adaptable conversations that can accommodate varying levels of intelligence and complexity. [40:40] - **Vibe coding accelerates AI-driven development**: Vibe coding, using tools like Cursor and Windsurf, allows developers to rapidly prototype and explore ideas by collaborating closely with AI. This approach, where AI suggests code and edits in real-time, significantly speeds up the creation of proofs-of-concept and demos. [54:47] - **AI's rapid advancement is a steep exponential**: AI models are advancing at an unprecedented pace, becoming smarter, faster, cheaper, and safer with each iteration. This exponential improvement, far exceeding Moore's Law, suggests the future will be dramatically different from today. [01:14:43]

Topics Covered

  • Build for the model that will exist in two months.
  • Think of AI as a human to build better products.
  • Why chat is an amazing and durable AI interface.
  • The future of product is ensembles of specialized models.
  • Today's AI is the worst you will ever use.

Full Transcript

the AI models that you're using today is

the worst AI model you will ever use for

the rest of your life And when you

actually get that in your head it's kind

of wild Everywhere I've ever worked

before this you kind of know what

technology you're building on But that's

not true at all with AI Every 2 months

computers can do something they've never

been able to do before and you need to

completely think differently about what

you're doing You're chief product

officer of maybe the most important

company in the world right now I want to

chat about what it's just like to be

inside the center of the storm Our

general mindset is in 2 months there's

going to be a better model and it's

going to blow away whatever the current

set of limitations are And we say this

to developers too If you're building and

the product that you're building is kind

of right on the edge of the capabilities

of the models keep going cuz you're

doing something right give it another

couple months and the models are going

to be great and suddenly the product

that you have that just barely worked is

really going to sing Famously you led

this project at Facebook called Libra

Libra is probably the biggest

disappointment of my career It

fundamentally disappoints me that this

doesn't exist in the world today because

the world would be a better place if

we'd been able to ship that product We

tried to launch a new blockchain It was

a basket of currencies originally It was

integration into WhatsApp and Messenger

I would be able to send you 50 cents in

WhatsApp for free It should exist To be

honest the current administration is

super friendly to crypto Facebook's

reputation is in a very different place

Maybe they should go build it now

[Music]

Today my guest is Kevin Wheel Kevin is

chief product officer at Open AI which

is maybe the most important and most

impactful company in the world right now

being at the forefront of AI and AGI and

maybe someday super intelligence He was

previously head of product at Instagram

and Twitter He was co-creator of the

Libra cryptocurrency at Facebook which

we chat about He's also on the boards of

Planet and Strava and the Black Product

Managers Network and the Nature

Conservancy He's also just a really good

guy and he has so much wisdom to share

We chat about how Open AI operates

implications of AI and how we will all

work and build product which markets

within the AI ecosystem companies like

OpenAI won't likely go after and thus

are good places for startups to own Also

why learning the craft of writing evals

is quickly becoming a core skill for

product builders what skills will matter

most in an AI era and what he's teaching

his kids to focus on and so much more

This is a very special episode and I'm

so excited to bring it to you If you

enjoy this podcast don't forget to

subscribe and follow it in your favorite

podcasting app or YouTube If you become

an annual subscriber of my newsletter

you get a year free of Perplexity Pro

Linear Notion Superhum and Ranola Check

it out at lennisnewsletter.com and click

bundle With that I bring you Kevin Wheel

This episode is brought to you by EPO

EPO is a next generation AB testing and

feature management platform built by

alums of Airbnb and Snowflake for modern

growth teams Companies like Twitch Miro

ClickUp and DraftKings rely on EPO to

power their experiments Experimentation

is increasingly essential for driving

growth and for understanding the

performance of new features And EPO

helps you increase experimentation

velocity while unlocking rigorous deep

analysis in a way that no other

commercial tool does When I was at

Airbnb one of the things that I loved

most was our experimentation platform

where I could set up experiments easily

troubleshoot issues and analyze

performance all on my own EPO does all

that and more with advanced statistical

methods that can help you shave weeks

off experiment time and accessible UI

for diving deeper into performance and

out-of-the-box reporting that helps you

avoid annoying prolonged analytic cycles

EPO also makes it easy for you to share

experiment insights with your team

sparking new ideas for the AB testing

flywheel EPO powers experimentation

across every use case including product

growth machine learning monetization and

email marketing Check out EPO at get

epo.com/lenny and 10x your experiment

velocity That's get

epo.com/lenny This episode is brought to

you by Persona the adaptable identity

platform that helps businesses fight

fraud meet compliance requirements and

build trust While you're listening to

this right now how do you know that

you're really listening to me Lenny

these days it's easier than ever for

fraudsters to steal PII faces and

identities That's where Persona comes in

Persona helps leading companies like

LinkedIn Etsy and Twilio securely verify

individuals and businesses across the

world What sets Persona apart is its

configurability Every company has

different needs depending on its

industry use cases risk tolerance and

user demographics That's why Persona

offers flexible building blocks that

allow you to build tailored collection

and verification flows that maximize

conversion while minimizing risk Plus

Persona's orchestration tools automate

your identity process so that you can

fight rapidly shifting fraud and meet

new waves of regulation Whether you're a

startup or an enterprise business

Persona has a plan for you Learn more at

withpersona.com/lenny Again that's with

pso

na.com/lenny Kevin thank you so much for

being here and welcome to the podcast

Thank you so much for having me We've

been talking about doing this forever

and we made it happen We did it I can't

imagine how insane your life is So I

really appreciate you that you made time

for this And we're actually recording

this uh the week that you guys launched

your new image model which is a happy

coincidence Uh my entire social feed is

filled of with skiiblifications of

everyone's life and family photos and

everything So good job Yep Mine too My

wife Elizabeth sent me one of hers So

I'm I'm right there with you Uh let me

just ask did you guys expect this kind

of reaction feels like this is the most

viral thing that's happened in AI which

is a high bar Uh since I don't know chat

GPT launched Just like did you guys

expect it to go this well uh what does

it feel like internally you know there

have been a handful of times in my

career when you're working on a project

or a product internally and the internal

usage just explodes Uh this was true by

the way when we were building stories at

Instagram More than anything else in my

career we could feel it was going to

work because we were all using it

internally and we'd go away for a

weekend you know before it launched We

were all using it and we'd come back

after a weekend and we would know what

was going on and be like "Oh hey I saw

you were at that camping trip that how

was that?" you were like "Man this thing

really works." Imagen was definitely one

of those Uh so we've been playing with

it for I don't know couple months And um

uh when it first went live internally to

the company there was kind of a a a

little gallery where you could generate

your own You could also see what

everyone else was generating and it was

just like non-stop buzz So yeah we had a

sense that this was going to be a lot of

fun for people to play with That's a

really cool like that should be a

measure of just like uh confidence in

the something going well that you're

launching is internally everyone's going

crazy for it Yeah especially social

things because

um you have a very tight network as a

company socially So you know each other

and you're experts in your product

hopefully and so there's some sense in

which if you're doing something social

and it's not taking off internally you

might you might question what you're

doing Yeah Uh and by the way the Giblly

thing is that something you guys did or

how did that even start was that like an

intentional example i think it's just

the style people love and model is is

really capable at at emulating style or

understanding what you know it's very

good at instruction following That's

actually something that I think people

I'm starting to see people discover with

it but you can do very complex things

You can give it two images you know one

is your living room and the other is a

whole bunch of photos or memorabilia or

things you want and you say like "Tell

me how you would arrange these things."

Or you can say "I'd like you to show me

what this will look like if you put this

over here and this thing to the right of

that and this one to the left of this

but under that one." And the model

actually will understand all of that and

do it It's incredibly powerful So I'm

I'm I'm just excited about all the

different things people are going to

figure out Yeah All right Well good job

Good job team OpenAI Uh let's get

serious here and let's kind of zoom out

a little bit The way I see it is you're

chief product officer of maybe the most

important company in the world right now

Uh just not to set the bar too high but

you guys are ushering in AI AGI at some

point super intelligence at some point

No big deal Uh I've had I have more

questions for you than I've had for any

other guest Actually put out a call out

on Twitter and LinkedIn and my community

just like what would you want to ask

Kevin

i had 300 over 300 well-formed questions

and we're going to go through every

single one So let's just get started I'm

just joking I picked out the best and

there's a lot of stuff I'm really

curious about It's it's 1 p.m here It

doesn't get dark for a while So let's do

it Okay here we go Okay so first of all

I'm just going to take notes here Uh

when is AGI launching when is the sign

up road map i mean we just launched a

good image gen model Does that count

it's uh it's getting there It's getting

there There's this um there's this quote

I love which is AI is whatever hasn't

been done yet because once it's been

done when it kind of works then you call

it machine learning and once it's kind

of ubiquitous and it's everywhere then

it's just an algorithm Um so I I've

always loved that that we call things AI

when they still don't quite work and

then you know by the time it's like an

AI algorithm that's recommending you

follow you know oh that's just an

algorithm but this new thing like

self-driving cars that's

I think to some degree we're always

going to be there and the next thing is

always going to be AI and the current

thing that we you know use every day and

is just a part of our lives that's an

algorithm It's so interesting because

yeah like uh in the Bay Area you see

self-driving cars driving around and

it's so normal now when like four years

ago I don't know three years ago you

would have thought you would have seen

this and you'd be like "Holy what

is how we're in the future and now we're

just so taken for granted." It's I mean

there's something like that with

everything If I showed you when GPT3

launched right i wasn't at OpenAI then I

was just uh I was just a user but it was

mind-blowing And if I gave you GPT3 now

I just plugged that into chat GPT for

you and you started using it you'd be

like "What is this thing?" I like this

like mess Uh slop slop There's I had the

same experience when I when I first got

into a Whimo right your your very first

ride at least my very first ride my

first like 10 seconds in a Whimo it

starts driving and you're like "Oh my

god watch out for that bike." You're

you're holding on to whatever you can

And then like five minutes in you've

calmed down and you realize that you're

getting driven around the city without a

driver and it's working You're just like

"Oh my god I am living in the future

right now." And then like another 10

minutes you're bored You're doing email

on your phone answering Slack messages

and you know suddenly this miracle of

human invention is just an expected part

of your life from then on And I there is

really something in the way that we all

are adapting to AI that's kind of like

that these miraculous things happen and

computers can do something they've never

been able to do before and it blows our

mind collectively for like a week and

then we're like oh yeah like oh yeah now

now it's just machine learning on its

way to being an algorithm the craziest

thing about what you just shared

actually is like I don't know chatg

which is like now feels terrible uh 3.5

was like a couple years ago and uh

imagine what life will be like in a

couple years from now we're gonna get to

that where things are going what you

think is going to be the next big leap

but I want start with the beginning of

your journey at

OpenAI Uh so you worked at Twitter you

worked at Facebook you worked at Planet

Instagram Uh at some point you got

recruited to go and come work at OpenAI

I'm curious just what that story was

like of the recruiting process of

joining Open Open AAI as CPO Is there

any are there any fun stories there

uh if I'm remembering the timeline right

we communicated uh planet I was leaving

and I was planning to just go take some

time you know like I wasn't going to

stop working but um but I was also happy

to take the summer This is like maybe

April or something I was like "Cool I'm

going to have the summer with my kids

We're going to you know go up to Tahoe

or something and I'll actually get to

hang out rather than what I usually do

going up and down and all that." And

then you know Sam and I had known each

other lightly for a bunch of years And

he's he's always involved in so many

interesting things you know like

companies building fusion and and all

these things So he'd always been

somebody that I would like call

occasionally if I was starting to think

about my next thing Um because I like

working on big like tech forward sort of

you know next next wave kind of things

And um and so uh I called him and I

think Venode also helped put us in touch

again and and this time it wasn't like

oh you should go talk to like these guys

working on fusion it he said actually

you know we're thinking about something

you should come talk to us I was like

"Okay that sounds amazing Let's do it."

And it goes really fast really really

fast Like I met uh you know most of the

management team in a brief period of

time a few days and they were telling me

"Look we're going to we're basically

going to move as fast as we as we want

to move." And uh it kind of if every if

you talk to everyone everyone likes you

We're ready to go Uh Sam came over for

dinner uh and we had we had a great

evening together just like talking about

OpenAI and the future and getting to

know each other better And at the end I

was like I I I was going to go in the

next day for like a bigger round of

interviews And um Sam was saying you

know hey it's going really well We're

really excited And I said cool So how do

I think about tomorrow and he said oh

you'll be fine Don't worry about it And

if it goes well like we're basically

there And so I go in the next day meet a

bunch of people have a great time like I

really enjoyed everybody I met

with in any interview You can always

second guess yourself you know like oh I

shouldn't have said that thing or I that

thing I gave a bad answer on I wish I

could redo But I I came away feeling

like I think that went pretty

well

And ex I was expecting to hear like that

weekend basically because they'd sort of

set expectations soon as you know if

this goes well we're ready to go

and uh I didn't hear

anything And then it was like Monday

Tuesday Wednesday I still didn't hear

anything And uh I reached out to uh to

folks on the OpenAI side a couple of

times still nothing And I was like "Oh

my god I screwed it up." Like I don't

know where I screwed it up but I totally

screwed it up I can't believe it And I

was going back to Elizabeth my wife and

being like "What did I do like where

where do you think I you know getting

all crazy about it and um and then it's

still nothing and finally it was like it

was like 9 days later they finally got

back to me and it turned out you know

there was like a bunch of stuff

happening internally and this that and

the other thing and uh you know there's

just a million things happening and they

finally were like oh yeah that went well

let's do this and I was like oh okay

cool let's do

it but uh it was like nine days of agony

and they were just super busy on some

internal stuff and uh there I was like

fretting every single day and re re

goinging over every line of our

interview process It makes me think

about when you're like dating someone

and you texted them and then they just

you're not hearing anything back and all

like you assume something is wrong Yeah

totally They might just be busy Uh I I

give them a hard time about it still So

that's wild Uh I love I love that it

worked out Uh and I guess I guess the

lesson there is don't don't jump to

conclusions Yeah H have a little bit of

chill

Speaking of that I want to chat about

what it's just like to be inside the

center of the storm Again you worked at

uh a lot of let's say traditional

companies even though they're not that

traditional Twitter and Instagram and

Facebook and Planet and now you work at

OpenAI I'm curious what is most

different about how things work in your

day-to-day life at OpenAI i think it's

probably the pace Uh maybe it's two

things One is it's the pace The second

is you know everywhere I've ever worked

before

this You kind of know what technology

you're building on So you spend your

time thinking about what what problems

are you solving who are you building for

you know how are you going to make their

lives better how are you going to is

this a big enough problem that you're

going to be able to to change habits you

know do people care about this problem

being solved all those like good product

things But the stuff that you're

building on is like kind of fixed you

know you're talking about databases and

things and I bet the database you use

this year is probably 5% better than the

database you used 2 years ago but that's

not true at all with AI It's like every

two months computers can do something

they've never been able to do before and

you need to completely think differently

about what you're

doing There there's like something

fundamentally interesting about that

makes life fun here There's also

something you know we'll maybe like talk

about evals later but that it also

really in this world of um you know

everything we're used to with computers

is about giving a computer very defined

inputs You know if you look at Instagram

for example there are buttons that do

specific things and you know what they

do And then when you give a computer

defined inputs you get very defined

outputs You're confident that if you do

the same thing three times you're going

to get the same output three times LLMs

are completely different than that right

they're good at fuzzy subtle inputs the

all the nuances of human language and

communication they're pretty good at And

also they don't really give you the same

answer You you probably get spiritually

the same answer for the same question

but it's certainly not the same set of

words every time And so you're much more

it's fuzzier inputs and fuzzier outputs

And it when you're building products it

really matters whi whether you know

there's some use case that you're trying

to build

around If the model gets it right 60% of

the time you build a very different

product than if the model gets it right

95% of the time versus if the model gets

it right 99.5% of the time And so

there's also something you have to get

really into the weeds on your use case

and the evals and things like that in

order to understand the right kind of

product to build So that is just

fundamentally different You know if your

database works once it works every time

and that's not true in this world Let's

actually follow this thread on evals I

definitely wanted to talk about this So

we had this uh legendary panel uh at the

line friends summit It was you and Mike

Griger and Sir Guo uh moderating So fun

And uh the thing that I heard that kind

of stuck with people from that panel was

a comment you made where you said that

writing evals is going to become a core

skill for product managers Yeah And I

feel like that probably applies further

than just product managers A lot of

people know what eval are A lot of

people have no idea what I'm talking

about So could you just briefly explain

what is an eval and then just why do you

think this is going to be so important

for people building products in the

future yeah sure I I think the easiest

way to think about it is almost like a a

quiz for a model a test to to gauge how

much it how well it knows a certain set

of subject material or how how good it

is at responding to a certain set of

questions So in the same way you you

know you take a calculus class and then

you have calculus tests that see if

you're you've learned what you're

supposed to learn You have eval that

test how good is the model at at

creative writing how good is the model

at uh at you know graduate level science

how good is the model at competitive

coding Uh and so you have these set of

evals that basically you know perform as

benchmarks for how smart or capable the

model is is like a simple way to think

about it like unit tests for model yeah

unit tests tests in general for models

totally great great okay and then uh why

is this so important for people that

don't totally understand what the hell

is going on here with eval why is this

so so key to building AI products uh

well it gets back to what I was saying

you need to know whether your model is

going to there are certain things that

models will get right 99.95% of the time

and you can just be confident there are

things that they're going to be 95%

right on and things that are going to be

60% right on if the model's 60% right on

something you're going to need to build

your product totally differently And by

the way these things aren't static

either So a big part of eval is if you

know you're you're building for some use

case So let's say let's take our deep

research product which is one of my

favorite things that we've released

maybe ever Um right the idea is with

deep research for people who haven't

used it you can give chat GPT now a an

arbitrarily complex query like it's not

about returning you an answer from you

know a search query which we can also do

it's it's here's a thing that if you

were going to answer it yourself you'd

go off and do you know two hours of

reading on the web and then you might

need to read some papers and then you

would come back and start writing up

your thoughts and realize you had some

gaps in your thinking so you go out and

do more research search you might it

might take you a week to write some like

20page answer to this question You can

let chat GPT just like chug for you for

2530 minutes You know it's not the

immediate answers you're used to but it

might go work for 2530 minutes and do

work that would have taken you a week So

as we were building that product we were

designing eval

uh sort of at the same time as we were

thinking about how this product was

going to work And we were trying to go

through like hero use cases You know

here's a question you want to be able to

ask Here's an amazing answer for that

question and and then turning those into

eval and and then hill climbing on those

eval So it's not just that the model is

static and we hope it does okay on a

certain set of things You can teach the

model You can make this a continuous

learning process And so as we were

fine-tuning our model for deep research

to to be able to answer these things we

were able to test is it getting better

on these evals that we said were

important measures of how the product

was working And it's when you start

seeing that and you start seeing

performance on eval

you made a kind of a comment along these

same lines around Ebells that uh that AI

is almost like capped in how amazing it

can be by that how good we are at Ebells

Does that resonate any more thoughts

along those lines these I mean these

models are are they're intelligences and

intelligence is so fundamentally

multi-dimensional So you can talk about

a model being amazing at competitive

coding which may not be the same as that

model being great at front-end coding or

back-end coding or taking a whole bunch

of code that's written in Cobalt and

turning it into Python you know like and

that's just within the software

engineering world And so I I think

there's a sense in which you can think

of these models as incredibly smart very

like factually aware uh intelligence is

but still most of the world's data

knowledge process is is not public It's

behind the walls of companies or

governments or other things And same way

if you were going to join a company you

would spend your first two weeks

onboarding You'd be learning the company

specific processes is you get access to

company specific data It's you can teach

these the models are smart enough you

can teach them anything but they need to

have the the sort of the raw data uh to

to learn from And so there's a there's a

sense in which um yeah I think the

future is really going to be incredibly

smart broad base models that are

fine-tuned and and and um tailored with

company specific or use case specific

data so that they perform really well on

company specific or use case specific

things Um and you're going to measure

that with custom evals And so you know

what I what I was referring to is just

like these models are really smart You

need to still teach them things if the

data is not in their training set And

there's a huge amount of use cases that

are not going to be in their training

set because they're relevant to one

industry or one company I'm just going

to keep following the thread that you're

leading us down but I'm going to come

back because I have more questions

around some of these things So you you

came to a a space that I think a lot of

AI founders are thinking about is just

where is open AI not going to come

squash me in the future or one of the

other foundational models and so it's

unclear to a lot of people just like

should I build a startup in this space

or not is there any advice you have or

any guidance for where you think openi

or just foundational models in general

likely won't go and where you have an

opportunity to build a company well one

of my So this is something that E

Williams used to say um back at Twitter

that's always stuck with me which is no

matter no matter how big your company

gets no matter how like incredible the

people are there are way more smart

people outside your walls than there are

inside your walls And it's why we are so

focused on building a great API We have

3 million developers using our API

uh no matter how ambitious we are how

big we grow By the way we don't want to

grow super big there are going to be

there there are so many use cases places

in the world where AI can fundamentally

make our lives better We're not going to

have the people We're not going to have

the the you know the the knowhow to

build most of these things And I think

like I was saying the data is is

industry specific use case specific you

know behind certain company walls things

like that And there are immense

opportunities in every industry and

every vertical in the world to go build

AI based products that improve upon the

the state-of-the-art And there's just no

way we could ever do that ourselves We

don't want to We couldn't if we did want

to And we're really excited to power

that for 3 million plus developers and

way more in the future Coming back to

your earlier point about the the the

tech changing constantly and getting

faster not exactly knowing what you'll

have by the time you launch something in

terms of the power that the model Uh I

was I'm curious what allows you to ship

quickly and consistently and such great

stuff And it sounds like one answer is

bottoms up empowered teams versus a very

top down road map that's you know

planned out for a quarter What what are

some of those things that allow you to

ship such great stuff so often so

quickly yeah I mean we try and we try

and have a a sense of where we're trying

to go you know point ourselves in a

direction so that we have some rough

sense of alignment Um like thematically

uh I don't for a second and we do

quarterly road mapping you know we we

laid out sort of a year-long strategy I

don't for a second believe that what we

write down in these documents is what

we're going to actually ship you know 3

months from now let alone six or nine

But that's okay There's a um I think

it's like an Eisenhower quote Plans are

useless Planning is helpful Uh which I

totally subscribe to especially in this

world It's really valuable if you think

about quarterly road mapping for example

It's really valuable to have a moment

where you stop and go okay what did we

do what worked what went well what

didn't go well what did we learn and now

what do we think we're going to do next

and by the way everybody has some

dependencies You you know you need the

infrastructure team to do the following

things Partnership with research here

And so you want to have a second to kind

of check your dependencies make sure

you're good to go and then start

executing We try and keep that really

lightweight because it's not going to be

right You know we're going to throw it

out halfway because we will have learned

new things

So the moment of planning is helpful

even if you're only going to you know

it's only partially right So that's I

think just expecting that you're going

to be super agile and that there's no

sense writing a three-month road map let

alone a year-long road map because the

technolog is changing underneath you so

quickly We really do try and go like

very strongly bottoms up kind of subject

to our overall directional

alignment We have great people um we

have engineers and PMs and designers and

researchers who are passionate about the

products they're building and have

strong opinions about them uh and are

also the ones building them and so

they're they have a they have a real

sense of what the capabilities are too

which is super important and so I think

you want to be more bottoms up in in

this way and so we operate that way we

are happy making mistakes we make

mistakes all the time it's one of the

things I really appreciate about Sam he

pushes us really hard to move fast But

he also understands that with moving

fast comes uh we didn't quite get this

right or you know we launched this thing

it didn't work we'll roll it back you

know look at our naming our naming is

horrible there's a lot of questions

people had for you the model names yeah

it it it's absolutely atrocious and we

know it um and we'll we'll get around to

fixing it at some point but it's not the

most important thing and so we don't

spend a lot of time on it but it also

shows you how it doesn't matter uh again

chatbt the most popular fastest growing

product in history Uh the models are

it's the number one AI API and model So

clearly it doesn't matter that much And

we name things like 03 mini high

Oh man I love it Um okay so you talked

about road mapping Um and bottoms up and

I'm really curious how you is there like

a a cadence or ritual of aligning with

you or Sam or he or you review

everything that's going out like is

there a meeting every week or every

month where you guys see what's

happening on key projects so we do

product reviews and things like that

like you would expect Um there isn't a

ritual because there isn't uh we we I I

would never want us to be blocked on

launching something you know waiting for

a review with me or Sam If we can't get

there if I'm traveling or Sam's you know

busy or whatever that's a bad reason for

us not to ship So obviously for the

biggest most high priority stuff we have

a pretty close beat on it but we really

try not to frankly um like we want to

empower teams to move

quickly and uh I think it's more

important to ship and iterate So we have

this philosophy that we call iterative

deployment and the idea is like we're

all learning about these models together

So there's a real sense in which it's

way better to like ship something

even when you don't know the full set of

capabilities and iterate together like

in public and we we kind of co-evolve

together with the rest of society as we

learn about these things and where

they're different and where they're good

and bad and weird I really like that

philosophy Um there's also a bit of I I

think the other thing that

that like ends up being a part of our

our product philosophy is uh this sense

of like model

maximalism The models are not perfect

they're going to make mistakes You could

spend a lot of time building all kinds

of different scaffolding around them And

by the way sometimes we do because

sometimes there are things you know

kinds of errors that you just don't want

to make But we don't spend that much

time building scaffolding around the

parts that don't match that because our

general mindset is in two months there's

going to be a better model and it's

going to blow away whatever you know the

current set of limitations are And so if

if you're building and we say this to

developers too If you're building and

and the product that you're building is

kind of right on the edge of the

capabilities of the models keep going

because you're doing something right

because you give it another couple

months and the models are going to be

great and suddenly the the product that

you have that just barely worked is

really going to sing And uh you know

that's that's kind of how you make sure

that you're really pushing the envelope

and building new things I had uh the

founder of Bolt on the podcast uh Stack

Blitz is the company name and he he

shared the story that they've been

working on this product for seven years

behind the scenes and it was failing

Nothing was happening and then all of a

sudden uh it was sorry to mention a

competitor but Claude uh came out or a

Sonnet 3.5 came out and all of a sudden

everything worked and they've been

building all this time and finally it

worked And I hear that a lot with YC

just like things are that never were

possible now are just becoming possible

every few months with the updates to the

models Yeah absolutely Let me actually

ask this I wasn't planning to ask this

but I'm curious if you have any quick

thoughts just why why is uh Sonnet so

good at coding and kind of thoughts on

uh your stuff getting as good and better

at actual coding Yeah I mean kudos to

Anthropic They've built very good coding

models Uh no doubt we uh we we think

that we can do the same Um maybe by the

time this uh podcast is shipped we'll

we'll have more to say but either way uh

all credit to them I

think this intelligence is really

multi-dimensional And so I think there's

the the the model providers it used to

be that OpenAI had this like massive

model lead you know 12 months or

something ahead of everybody else That's

not true anymore You know I like to

think we still have a lead I'd argue

that we do but it's certainly not a

massive one And that means that there

are going to be different places where

you know the Google models are really

good or where anthropics models are

really good or where we're really good

and and our competitors are like "Ah we

got to get better at that." And it

actually is easier to get better at a

certain thing once someone's proved it

possible than it is to you know forge a

path through the the jungle and doing

something brand new So I just think yeah

as an example it was like nobody nobody

could break four minutes in the mile and

then finally somebody did and the next

year 12 more people did it I I think

there's that all over the place and it

just means that competition is really

intense and consumers are going to win

and developers are going to win and

businesses are going to win in a big way

from that It's part of why the industry

moves so fast But um you know all

respect to to the other big model

providers models are getting really good

We're going to move as fast as we can

and I think we've got some good stuff

coming Exciting Uh this makes me also

think about uh in many ways other models

are better at certain things but somehow

Chad Gibbt is like the like if you look

at all the awareness numbers and usage

numbers it's like no matter where you

guys are in the rankings people seem to

just like think of AI and chatbt almost

as as the same What do you think you did

right to kind of win in consumer mindset

at least at this point in awareness in

the world i think being first helps

which is one of the reasons why we're so

focused on moving quickly Um you know we

like being the first to launch new

capabilities things like deep research

Uh we've also our models are very they

can do a lot of things right so they can

they can take real-time video input they

can you have speech to speech you can do

speech to text and text to speech Um

they can do deep research they can

operate on a canvas they can write code

And so Chat GBT can kind of be this

one-stop shop where all the things that

you want to do are possible Um and as we

as we go forward and it you know we have

more agentic tools like operator where

it's browsing for you and doing things

for you on the web you know more and

more you're going to be able to come to

this one place to chat GPT give it

instructions and have it accomplish real

things for you in the world there's like

something fundamentally valuable in that

and so you know we we think a lot about

that we think and it it we we move we

try to move really fast so that we are

always the most useful place for people

to come do What would you say is uh the

most counterintuitive thing that you've

learned after building AI products or

working at OpenAI something I was just

like I did not expect that I don't know

maybe I should have expected this but

one of the things that's been funny for

me is um the extent to which you can

kind of reason when you're trying to

figure out how some product should work

with AI you can often or even why some

AI thing happens to be true you can

often reason about it the way you would

reason about another human and it kind

of works Yeah So maybe a couple examples

When we were first launching our um our

reasoning model right we were the first

to to build a a model that could reason

that could that could instead of giving

you just a quick you know system one

answer right away to every question you

asked It was the third emperor of the

Holy Roman Empire Like you know here's

an answer You could ask it hard

questions and it would reason the same

way that if I asked you to do a

crossword puzzle you couldn't just like

snap fill in everything you would be

well okay on this one across I think it

could be one of these two but that means

there's an A here so that one has to be

this oh way you know like backtrack kind

of step by step build up from where you

are same way you answer any any

difficult uh logistical problem any

scientific problem so this reasoning

breakthrough was big but it was also the

first time that a model needed to sit

and think and that's a weird paradigm

for a consumer product you don't

normally have something where you might

need to hang out for 25 seconds after

you ask a question And and so we were

trying to figure out you know what's the

UI for this because it's also not like

with deep research where the model is

going to go and think for 25 minutes

sometimes It's actually not that hard

because you're not going to sit and

watch it for 25 minutes You're going to

go do something else You're going to go

to another tab or go get lunch or

whatever Uh and then you'll come back

and it's done When it's like 20 25

seconds or 10 seconds it's a long

experience It's a long time to wait but

it's not long enough to go do something

else And so you actually need and you

know so you you can think like if you

asked me something that I needed to

think for 20 seconds to answer what

would I do i I wouldn't just like go

mute and not say anything and kind of um

you know shut down for 20 seconds and

then come back So we shouldn't do that

We shouldn't just like have a slider

sitting there That's annoying But I also

wouldn't just start like babbling every

single thought that I had Um so we

probably shouldn't just like expose the

whole chain of thought as the model's

thinking but you know I might go like

huh that's a good question All right I

might approach it like that and then

think and you know you're sort of like

maybe giving little updates And that's

actually in what we ended up shipping

You have similar things where you can

like you can find situations where um

you get better thinking sometimes out of

a group of models uh that all try and

attack the same problem and then you

have a model that's looking at all their

outputs and integrating it and then

giving you a single answer at the end I

mean sounds a little bit like

brainstorming right like I certainly

have better ideas when I get in a room

and brainstorm with other people because

they think differently than me

So anyways there's just like all these

situations where you can actually kind

of reason about it like a group of

humans or an individual human and it

sort of works which I don't know maybe

maybe I shouldn't have been surprised

but I was That is so interesting because

when I see these models operate I like I

never even thought about you guys

designing that experience Like to me it

just feels like this is what the LLM

does It just sits there and tells me

what it's thinking And I love this point

you're making of like we like let's make

it feel like a human operating and how

does a human operate well they just talk

out loud They think here's the thing I

should explore And I love that deep

sequence like to the extreme of that

right where they're just like here's

everything I'm doing and thinking and

people actually like that too I guess

Was that was that surprising to you like

oh maybe that could work too People seem

to like everything Yeah we learned from

that actually Um because we um when we

first launched it we kind of gave you

like the the subheadings of what the

model was thinking about but not much

more And then DeepSeek launched and

there were it was a lot and we kind of

went you know I don't know if everyone

wants like that There's some novelty

effect to seeing what the model's really

thinking about We felt that too when we

were looking at it internally It's

interesting to see the model's chain of

thought but it's not you know I think at

the scale of like 400 million people you

don't want to see the model kind of like

babble a bunch of things Um and so what

we ended up doing was summarizing it in

interesting ways So instead of just

getting the subheadings you're kind of

getting like one or two sentences about

how it's thinking about it And you can

learn from that So we kind of tried to

find a middle ground that that we

thought was an experience that would be

meaningful for most people but you know

showing everybody like three paragraphs

uh is probably not the right answer This

reminds me of something else you said at

the summit that has really stuck with me

this idea that chat people always make

fun of like chat is not like the future

interface for how we interact with AI

but you made this really interesting

point that may argue the other side

which is like as humans we interface by

talking and the IQ of a human can span

from really low to really high and it

all works because we're talking to them

and chat is the same thing and it can

work on all kinds of intelligence levels

Uh maybe just share maybe I just shared

it but I guess anything there about just

why chat actually ends up being such an

interesting interface for LLMs Yeah I

don't know if maybe I'm uh maybe this is

one of those things I believe that most

people don't believe but I actually

think chat is an amazing interface

because it's so

versatile Um people tend to go "Oh chat

Yeah well that's just like you know

we'll figure out something better." And

I kind of think I kind of think this is

uh it's a it's it's incredibly universal

because it is the way we talk Like I can

talk to you verbally like we're talking

now I can you know we can see each other

and interact uh we can talk on WhatsApp

and you know be texting each other but

all of these things is this sort of like

unstructured

uh you know method of communication and

that's how we operate If I had to end if

I had some more rigid interface that I

was allowed to use when we spoke I would

be able to speak to you about you know

far fewer things and it would actually

get in the way of us having like maximum

communication bandwidth So there's

something magical and and by the way in

the past it never worked because models

there there wasn't a model that was good

at understanding all of the complexity

and nuances of human speech and that's

the magic of LLMs So to me it's like an

interface that's exactly fit to the

power of these things And that doesn't

mean that it always has to be just like

I don't necessarily always want to type

but if you you do want that very

open-ended flexible communication medium

It may be that we're speaking and the

model's speaking back to me but you

still want that like that that the very

sort of lowest common denominator um no

restrictions way of of interacting That

is so interesting That's really changed

the way I think about the stuff is that

point that chat is just so good for this

very specific problem of talking to

super intelligence basically By the way

I think there are like it's not that

it's only chat either Like there are if

you have high volume use cases where

they're more prescribed and the you

don't actually need the full

generality There are there are many use

cases where it's better to have

something that's less flexible more

prescribed faster at a specific task and

those are great too and you know you can

build all sorts of those and u but you

still want chat as like this baseline

for anything that falls out of whatever

you know vertical you happen to be

building for It's like a catch-all for

like every possible thing you'd ever

want to express to a model I'm excited

to chat with Christina Gilbert the

founder of One Schema one of our

longtime podcast sponsors Hi Christina

Yes thank you for having me on Lenny

what is the latest with one schema i

know you now work with some of my

favorite companies like RAMP Vanta Scale

and Watershed I heard that you just

launched a new product to help product

teams import CSVs from especially tricky

systems like ERPs Yes So we just

launched one schema file feeds which

allows you to build an integration with

any system in 15 minutes as long as you

can export a CSV to an SFTP folder We

see our customers all the time getting

stuck with hacks and workarounds And the

product teams that we work with don't

have to turn down prospects because

their systems are too hard to integrate

with We allow our customers to offer

thousands of integrations without

involving their engineering team at all

I can tell you that if my team had to

build integrations like this how nice

would it be to be able to take this off

my road map and instead use something

like one schema and not just to build it

but also to maintain it forever

Absolutely Lenny We've heard so many

horror stories of multi-day outages from

even just a handful of bad records We

are laser focused on integration

reliability to help teams end all of

those distractions that come up with

integrations We have a built-in

validation layer that stops any bad data

from entering your system and one schema

will notify your team immediately of any

data that looks incorrect I know that

importing incorrect data can cause all

kinds of pain for your customers and

quickly lose their trust Christina thank

you for joining us and if you want to

learn more head on over to one schema.c

co that's one schema.co See how

I want to come back to that you talked

about researchers and the relationship

with product teams Uh I imagine a lot of

innovation comes from researchers just

like having an inkling and then building

something amazing and then releasing it

and some ideas come from PMs and

engineers How does how do those teams

collaborate like does every team have a

PM is it a lot of researchled stuff just

like what give us a sense of just where

ideas and products come from mostly it's

an area where we're evolving a lot I'm

really excited about it Frankly I I

think if you go back you know a couple

years when chat GBT was just getting

started

uh obviously I wasn't at OpenAI so

um but uh it we were more we were more

of a pure research company at the time

ChachVT if you remember was a low-key

research preview Um it for many years

Yeah It it wasn't a thing that the team

launched thinking it was going to be

this massive product Oh JPT And it it

was just a way that we were going to let

people you know play with and iterate on

the models Um and so we were we were

primarily a research company a

world-class research company And as Chat

GPT has grown and as we've built our B2B

products and our APIs and other things

now we're more of a product company than

we were I still think we can't we're

OpenAI should never be a pure product

company We need to be both a world-class

research company and a world-class

product company And the two need to

really work together And that's the

thing that's um that I think we've been

getting much better at over the last

like 6 months If you if you treat those

things separately and you know the

researchers go do amazing things and

build models and then they get to some

state and then the product and

engineering teams go take them and do

something with them We're effectively

just an API consumer of our own models

The best products though are going to be

is like I was talking about with deep

research It's a lot of iterative

feedback It's understanding the products

you're trying to solve or the the

problems you're trying to solve building

evals for them using those evals to go

gather data and fine-tune models to get

them to be better at the these use cases

that you're looking to solve It's a huge

amount of back and forth uh to do it

well And I think the best products are

going to be product design and research

working together as a single team to to

build novel things So that's that's

actually how we're trying to operate

with basically anything that we build

It's a new muscle for us because we're

kind of new as a product company but um

but it's one that people are really

excited about because we've seen every

time we do it we build something awesome

and so you know now every product starts

like that How many product managers do

you have at OpenAI i don't know if you

share that number but if you do Not that

many actually I don't

know 25 Um maybe it's a little more than

that but I my personal belief is that

you want to be pretty PM light as an

organization just in general I say this

with love because I am a PM but too many

PMs causes problems You know we'll like

fill the world with decks and ideas

versus execution So I think that the the

I I think it's a good thing when you

have a PM that has uh that that is

working with maybe slightly too many

engineers because it means that they're

not going to get in and micromanage

You're going to leave a lot of of you

know influence and responsibility with

the engineers to make decisions It means

you want to have really product focused

engineers which we're fortunate to have

We have an amazingly product focused

like high agency engineering team But

when you have something like that you

have a team that feels super empowered

You have a a PM that's you know trying

to really understand the problems and

kind of gently guide the team a little

bit but has too much going on to get too

far into the details and you end up

being able to move really fast So that's

kind of the philosophy we take Uh we

want we want producty leads and and

producty engineers all the way through

Um we want not too many PMs but really

awesome high quality ones Um and so far

that seems to be working pretty well I

imagine being a PM at OpenAI is like a

dream come true for a lot of people Uh

at the same time I imagine it's not a

fit for a lot of people There's

researchers involved very product-minded

engineers What do you what do you look

for in the PMs that you hire there for

folks that are like maybe I pro I

shouldn't go work there I shouldn't even

think about that I think I I've said

this a few times but like high agency is

something that we really look for People

that are not going to come in and kind

of wait for everyone else to allow them

to do something They're just going to

see a problem and go do it Um that's

it's just a core part of how we

work I think people that that are happy

with ambiguity because there is a

massive amount of ambiguity here is not

the kind of place and and we have we

have trouble sometimes with um with more

junior PMs because of this because it's

just not the place where someone is

going to come in and say "Okay you know

here's here's the landscape Here is your

area I want you to go do this thing."

And that's that's what you want as a as

an early career

PM We just I mean no one here has time

and the nobody the problems are too

ill-formed and we're figuring them all

out as we go And so um high agency very

comfortable with

ambiguity ready to come in and help

execute and move really

quickly That that's kind of our our

recipe And I think

also happy leading through

influence because I mean it's usual as a

PM people don't report to you Uh your

team doesn't report to you etc But you

also have the the

the complexity of a research function

which is even more sort of

self-directed and it's really important

to build a good rapport with the

research team Uh and so you know that I

think the EQ side of things is also

super important for us I know at most

companies a PM comes in and they're just

like why do we need you and as a PM you

have to uh earn trust and help people

see the value And I feel like at OpenAI

it's probably a very extreme version of

that where they're like why do we need

this person we have researchers

engineers what are you going to do here

yeah I think people appreciate it done

right Um but you got you bring people

along I I think one of the most

important things a PM can do well is be

decisive So it's

it's there's a real fine line You don't

want to be making I mean it's kind of

like I I don't love the PM as the CEO of

the product uh illusion all the time But

but just like Sam in his role would be

making mistakes if he made every single

decision in every meeting that he was in

and he would also be making mistakes if

he made no decisions in any meetings

that he was in Right it's a it's the

it's understanding when to defer to your

team and to like let let people innovate

and when there is like a decision to be

made that people either don't feel

comfortable with or don't feel empowered

to make or a decision that that you know

has too many different like disperate

pros and cons that are spread out across

a big group and someone needs to be

decisive and make a call It's a really

important trait of a CEO It's something

Sam does well and it's it's also a

really important trait of a PM kind of

at a at a more microscopic level And so

because there's so much ambiguity it's

not obvious what the answer is in a lot

of cases And so having a PM that can

come in and like and by the way this

doesn't need to be a PM I'm perfectly

happy if it's anybody else but I kind of

look to the PM to say like if there's

ambiguity and no one's making a call you

better make sure that we get a call made

and we move forward

This touches on a few posts I've done of

just where is AI going to take over work

that we do versus help us with various

work So let me come at this question

from a few different direction of just

how AI impacts product teams and hiring

things like that So first of all there's

all this talk of uh LM doing our coding

for us and 90% of code is going to be

written by AI in a year Dario at

Enthropic said that at the same time you

guys are all hiring engineers like crazy

PM's like crazy You know every function

is dead but you're still hiring every

single one

Uh I guess just first of all let me just

ask this How do you how do you and the

team like say engineers PMs use AI in

your work is there anything that's like

really interesting or things that you

think people are sleeping on and in and

how you use AI in your day-to-day work

we use it a lot I mean every one of us

is in chat GPT all the time summarizing

docs using it to help write docs with

GPTs that you know write product specs

and things like that All all the stuff

that you would imagine I I mean talk

about writing evals like you can

actually use models to help you write

evals and they're pretty good at

it That all said I still don't I'm still

sort of disappointed by by us and this

by I really mean me Um in if I were to

if I were to just like teleport my

5-year-old self leading product at some

other company into my day job I would

recognize it still And I think we should

be in a world certainly a year from now

probably even more now that um where I

almost wouldn't recognize it because the

workflows are so different and I'm using

AI so heavily and I'd still recognize it

today So I think in some sense I'm not

doing a good enough job of that You know

just to give an example

like why shouldn't we be like vibe

coding uh demos right left and center

like instead of showing stuff in like

Figma we should be showing prototypes

that people are vibe coding you know

over the course of 30 minutes to

illustrate proofs of concept and to

explore ideas That's totally possible

today and we're not doing it enough Our

actually our chief people officer Julia

was telling me the other day she

vibecoded an internal tool that she had

at a previous job that she really wanted

to have here at OpenAI and she opened I

don't know wind surf or something and

vibecoded it Like how cool is that and

if our chief people officer is doing it

we have no excuse to not be doing it

more That's an awesome story Okay And

some people may not have heard this term

VIP coding Can you describe what that

means yeah Uh I think this was uh I

think this was Andre's uh term Karpathy

Yeah Andre Karpathy Yeah Um where it's

just so you have these tools like cursor

and windsurf that and GitHub copilot

that are very good at suggesting uh what

code you might want to write So you can

give them a prompt and it'll write code

and then as you go to edit it it's

suggesting what you might want to do And

the the the way that that everyone

started using that stuff was give it a

prompt have it do stuff you go edit it

give it a prompt you know and you're

kind of like really going back and forth

with the model the whole time as the

models are getting better and as people

are getting more used to it you can kind

of just like uh let go of the wheel a

little bit And when the model's

suggesting stuff it's just like tap tap

tap tap tab like keep going Yes Yes Yes

Yes Yes And of course the model makes

mistakes or it does something that

doesn't compile But when it doesn't

compile you paste the error in and you

say go go go And then you you test it

out and it like does one thing that you

don't want it to do So you enter in an

instruction and say go go go go go And

you just kind of like let the model do

its thing And it's not that you would do

that for production code that needed to

be super uh tight today yet But for so

many things you're trying to get to a

proof of concept You're getting to a a

demo and you can really take your hands

off the wheel and the model will do an

amazing job And that's what that's

that's vibe coding That's an awesome

explanation I think like the pro version

of that which is I think the way Andre

even described it is you talk you do

like a there's a step like whisper super

whisper something like that where you're

like talking to the model not not even

typing Yeah totally Oh man So let me let

me just ask I guess when you look at

product teams in the future you talked

about how you guys should be doing this

more instead of designs having

prototypes What do you think might be

the biggest changes in how product teams

uh are structured or built where do you

think things are going in the next few

years i think you're definitely going to

live in a world where you have more um

where you have researchers built into

every product team And I don't even mean

just at at like foundation model

companies because I think the future

actually frankly one thing that I'm sort

of surprised about about our industry in

general is that there's not a greater

use of fine-tuned models Uh like a lot

of people you know these models are very

good So our API does a lot of things

really well but when you have particular

use cases you can always make the the

model perform better on a particular use

case by fine-tuning it It's probably

just a matter of time You know folks

aren't like quite comfortable yet with

doing that in every case but to me

there's no question that that's the

future Every you models are going to be

everywhere just like transistors are

everywhere AI is going to be just a part

of the fabric of everything we do But I

think there are going to be a lot of

fine-tuned models because why would you

not want uh to to more specifically

customize a model against a particular

use case and so I think you're going to

want sort of quasi researcher uh machine

learning engineer types as part of

pretty much every team because

fine-tuning a model is just going to be

part of the core workflow for building

most

products So that's that's one change

that maybe you know you're starting to

see at foundation model companies that

will propagate out to more teams over

time I'm curious if there's a concrete

example that makes that real and I'll

share one that comes to mind as you talk

Sure Which is when you look at cursor

and winds surf something I learned from

those founders is that they they use

like a sonnet but then they also have a

bunch of custom models that help along

the edges that make the specific

experience that's not just generating

code even better like autocomplete and

looking ahead to where things are going

So is that one any other examples of

what you what what what is a fine-tuned

model there that do you think teams will

be building with these researchers on

their teams yeah I mean so when you're

fine-tuning a model one of you're you're

basically giving the model uh a bunch of

of examples of the kinds of things you

want it to be better at So it's it's

here's a problem here's a good answer

here's a problem here's a good answer uh

or here's a question here's a good

answer you know times a thousand or or

10,000 Uh and suddenly you're you're

teaching the model to be much better

than than it was out of the gate at that

particular thing We use it everywhere

internally Um we also we use ensembles

of models much more internally than

people might think Um so it's not here

is I I have 10 different problems I'll

just ask you know baseline GPT40 about a

bunch of these things If we have 10

different problems we might we might

solve them using uh you know 20

different model calls some of which are

using specialized fine-tuned models

They're using models of different sizes

because maybe you have different latency

requirements or cost requirements at

different for different questions They

are probably using custom prompts for

each one Like basically you want the to

teach the model to be really good at you

want to break the problem down into more

specific tasks versus some broader set

of highle

tasks and then you can use models very

specifically to get very good at each

individual thing and then you know you

have an ensemble that sort of tackles

the whole

thing I think a lot of good companies

are doing that today I still see a lot

of companies uh kind of giving the model

single generic broad problems versus

breaking the problem down and I think

there will be more breaking the problem

down using specific models for specific

things including fine-tuning And so in

your case because this is really

interesting is is that you're using

different uh levels of chat GBT like 01

03 and stuff that's earlier there'll be

parts of our internal stack So we do if

you give you an example uh customer

support with 400 plus weekly uh 400 plus

million weekly active users we get you

know a lot of inbound tickets right I

don't know how many customer support

folks we have but it's not very many 30

40 I'm not sure way way smaller than you

would have at any comparable company and

it's because we've automated a lot of

our flows we've got you know most

questions using our internal resources

knowledge base you know uh guidelines

for how we answer questions what kind of

personality etc You can teach the model

those things and then have it do a lot

of its answers automatically or where it

doesn't have uh you know the full

confidence to answer a particular

question it can still suggest an answer

request a human to look at it and then

that human's answer actually is its own

sort of fine-tuning data for for the

model you're telling it the right answer

in a particular case and uh and we're

using at various places you know some of

these places you want a little bit more

reasoning is not super latency sensitive

so you want a little more reasoning and

we'll use one of our oer models in other

places you want a quick check on

something and so you're fine to use like

40 mini which is super fast and super

cheap and in general it's like specific

models for specific purposes and then

you you you ensemble them together to

solve problems by the way Again not

unlike how we as humans solve problems A

company is arguably an ensemble of

models that have all been you know

fine-tuned and based on what we studied

in college and what we have like learned

over the course of our careers We've all

been fine-tuned to have different sets

of skills and you like group them

together in different configurations and

the output of the ensemble is much

better than the output of any one

individual Kevin you're blowing my mind

That sounds exactly correct Uh and also

different people are you pay them less

Uh they they cost less to talk to Some

people take a long time to answer Some

people hallucinate This is I'm telling

you this is like a this is a mental

model that really does work in in

thinking This is great Some people are

visual They want to draw out their

thinking Some people want to talk word

cell Wow This is a really good metaphor

So again coming back to your advice here

because I love that we circled back to

it It's you're finding a really good way

to think about how to design great AI

experiences and LMS I guess specifically

is think about how a person would do

this Well it's it's it's maybe not

always the answer is to think about how

a person would do it But but sometimes

to gain intuition for how you might

solve a problem you think about what an

equivalent human would do in those

situations and use that to to you know

at least gain a different perspective on

the problem Wow this is great There's

just like you know because so much of

this really is talking to a model

there's a lot of prior art because we

talk to other humans all the time and

encounter them in all sorts of different

situations and and so like there's a lot

to learn from that Okay so speaking of

humans I want to chat about the future a

little bit So you have three kids and

someone a community member asked me this

hilarious question that I think it's

something a lot of people are thinking

about So this is Patrick Sil I worked at

him with a mid Airbnb He asks CS says

"Ask what he's encouraging his kids to

learn to prepare for the future I'm

worried my six-year-old by the year 2036

will face a lot of competition trying to

get into the top roofing or plumbing

programs and need a backup

plan." That's funny Um so our kids are

we have a 10-year-old and 8-year-old

twins so they're they're still pretty

young Uh they're they're kind of I mean

it it's amazing how AI native they are

like they just it's completely normal to

them that there are self-driving cars

that they can talk to AI all day long Um

they have full conversations with chat

GPT and Alexa and everything else I

don't know I think who knows what the

future holds I I think you know things

like coding skills are going to be

relevant for a long time Who knows but I

I think if you teach your kids to be

curious to be independent to be

self-confident you teach them how to

think I don't know what the future holds

but I think that those are going to be

skills that are going to be important in

in any configuration of the future And

so you know it's not like we have all

the answers but that's how Elizabeth and

I think about uh our kids And do you

find that AI there's a lot of talk about

AI tutoring is that something you guys

are doing anything you're I know they're

using catchup I love I love all the

photos you post where they're playing

with prompts and stuff but I guess is

there anything there you're you're

experimenting with or you think is going

to become really important this is

something that uh it's maybe the most

important thing that that AI could do

Maybe that's a maybe that's a grand

statement There are lots of important

things that AI can do including like

speeding up the pace of fundamental

science research and discovery which I

maybe is actually the most important

thing AI can do but but one of the most

important things would be personalized

tutoring And it kind of blows my mind

that there is still I know there are

there are a bunch of good products out

there like you know Khan Academy does

great things They're a wonderful partner

of ours Uh Venode Kosla has a nonprofit

that has uh that's doing some really

interesting stuff in the space and is

making an impact But I kind of want like

I'm kind of surprised that there isn't

like a two billion

kid you know AI personalized tutoring

thing because the models are good enough

to do it now And every every study out

there that's ever been done seems to

show that when you have you know

classrooms is still classroom like

education is still important but when

you combine that with personalized

tutoring you get like multiple standard

deviation improvements in learning speed

And so it's just it's

uncontroversial It's good for kids It's

free Chat GPT is free You don't need to

pay for and the models are good enough

Like it still just kind of blows my mind

that there isn't something amazing out

there that you know our kids are using

and your future kids are using and like

people in all sorts of places around the

world that aren't as lucky as our kids

to be able to like have this sort of

built-in solid education Again chat GPT

is free People have Android devices

everywhere Like this could I I really

just think this could change the world

and I'm surprised it doesn't exist and I

want it to exist This kind of touches on

something I want to spend a little time

on which is a lot of people also worry a

lot about AI where it's going They worry

about jobs it's going to take They worry

about you know the super intelligence

squashing humanity in the future What's

kind of your perspective on the on that

and just kind of the optimistic case

that I think people need to hear i mean

I'm a big technology optimist I think if

you look over the last 200 years uh

maybe maybe more technology has driven a

lot of the advancements that have made

us the the world and the society that we

are today It drives economic

advancements It drives um uh

geopolitical advancements quality of

life longevity advancement I mean

technology is at the root of of just

about everything So I I think there are

very few examples where uh where this is

anything but a great a great thing over

the longer term That doesn't mean that

there aren't like temporary dislocations

or where there aren't individuals that

are impacted and that's like that that

matters too So it can't just be that the

average is good You've got to also think

about how you take care of each

individual person as best you can

So uh it's something that we think a lot

about and as we you know work with the

administration as we work with policy

like we we try and help where wherever

we can We do a lot with education Um you

know one of the one of the benefits here

is that chat GPT is also perhaps the

best like reskilling app you could

possibly want It knows a lot of things

It can teach you a lot of things if

you're interested in learning new things

So but I these are these are very real

issues I'm super optimistic about the

long run and we're going to need to do

everything we can as a society to ensure

that we like make this transition you

know a as graceful and as well supported

as we can to give people a sense of

where things might be going That's a big

question a lot of people's minds So

someone asked this question that I love

which is uh AI is already changing

creative work in a lot of different ways

writing and design and coding What do

you what do you think is the next big

leap what should we be thinking is the

next big leap in AI assisted creativity

specifically and then just broadly like

where do you think things are going to

be going in the next few years yeah this

is also an area where I'm I'm a big

optimist like if you if you look at Sora

for example I mean we talked about image

gen earlier and the the the absolute

like fountain of creativity that people

are putting across Twitter and Instagram

and other places I'm I am the world's

worst artist Like the worst Maybe the

only thing I'm worse at than than than

art is singing And I you know I like

give me a pencil and a pad of paper and

I can't draw better than my five than

our 8-year-old You know it's just like

it's but give me give me image gen and

you know I can think some creative

thoughts and put something into the

model and suddenly have output that I

couldn't have possibly done myself

That's pretty cool Even even you look at

um at folks that are really talented I

was talking to a director recently about

Sora someone who's directed films that

that that we would all know And uh and

he was saying you know for for a film

that he's doing like say say take the

example of some sort of sci-fiish you

know think of like Star Wars and you've

got some scene where there's a there's a

plane zooming into some Death Star-L

like thing And so you've got the plane

looking at the whole planet and then you

want to cut to a scene where the the

plane's like you know kind of at the

ground level and all of a sudden you see

the city and everything else right how

are you going to manage that cut scene

and and that transition and he he was

saying you know in in the world of two

years ago I would have paid uh uh you

know a 3D effects company

uh a h 100red grand and they would have

taken a month and they would have

produced two versions of this cut scene

for me and I would have evaluated them

We would have chosen one because what

are you going to do like pay another 50

grand and wait another month and uh and

we would have just gone with it and you

know it would be fine like movies are

great I love them and and um there have

been obviously we can do great things

with the technology that we've had but

you now look at what you can do with

Sora and and his point was now I can use

Sora our video model and I can get 50

different variations of this cut scene

just you know me brainstorming into a

prompt and the model brainstorming a

little bit with me I've got 50 different

versions and and then of course I can

like iterate off of those and refine

them and take different ideas and now

I'm still going to go to that that 3D

effect studio to produce the final one

but I'm going to go having brainstormed

and like had this had a much more

creative approach with a with an outcome

that's much better and and like I did

that assisted by AI So my personal view

on on creativity in general is that it's

no one's going to you don't type into

Sora like make me a great movie It

requires creativity and ingenuity and

all these things but it can help you

explore more It can help you get to a

better final result So you know again I

tend to be an optimist in in most things

but I'm actually I I I think I think

there's a very good story here I know

Sam Alman I think it was him who tweeted

recently the creative writing piece that

you guys are working on where it's say

is very bad at writing creative stuff

and he shared an example where it's

actually really good I imagine that's

another area of investment Yeah there's

there's some exciting stuff happening

internally um with some new research

techniques So uh we'll have more to say

about that at some point but yeah Sam uh

Sam sometimes uh likes to show off some

of the stuff that's coming um which is

know by the way it's like very sort of

indicative of this iterative deployment

uh philosophy We don't have some

breakthrough and keep it to ourselves

forever and then you know bestow it upon

the world someday We kind of just talk

about the things we're working on and

share when we can and launch early and

often and then iterate in public And I I

I I really like that philosophy I love

all these hints at a few things coming I

know you can't say too much You talked

about how there might be a coding leap

coming in the near future maybe by the

time this comes out Is there anything

else people should be thinking about

might be coming in the near future any

things you can tease that are

interesting exciting man this hasn't

been enough for you

Oh only everything is getting better

every day Yeah I'm like man I hope uh I

hope we get some of this stuff out

before the the episode launches This is

your new time box I don't piss people

off Um no Uh it's the the the the

amazing thing to me is we we were

talking earlier about how far models

have come in just a couple years If you

went back to GBT3 you'd be like

disgusted by how bad it was even though

Lenny of two years ago was you know mind

blown by how good these were Um and for

a long time we were iterating every you

know 6 to9 months on a new GPT model It

was like GPT3 GPT3.5 four And now with

this O series of reasoning models we're

moving even faster We're like every

roughly you know 3 months maybe four

months there's a new Oser model And each

of them is a step up in in capability

And

so the capabilities of these models are

are increasing at a massive pace They're

also getting cheaper as as they scale

You know you you look at uh at where we

were even like a couple years ago The

original I think the original I don't

know what was it GPT 3.5 or something

was like 100x the cost of GPT 40 mini

today in in the API So couple years

you've gone down two orders of magnitude

in uh in cost for much more intelligence

And so I like I don't know where there's

another series of trends like that in

the world Models are getting smarter

they're getting faster they're getting

cheaper and they're getting safer too Uh

you know they hallucinate less every

every iteration And so there's just you

know the the Moore's law and and and

transistors becoming ubiquitous That was

a that was a law around doubling the

number of transistors on a chip every 18

months If you're talking about something

where you're getting 10x every year

that's a massively steeper exponential

And it just you know it it tells us that

the future is going to be very different

than today I I I still the the thing I

try and remind myself is the AI models

that you're using

today is the worst AI model you will

ever use for the rest of your life And

when you actually get that in your head

it's kind of wild I was going to

actually say the same thing That's

that's the thing that always sticks with

me when I watch this thing Like you're

talking about Sora and I imagine many

people hearing that are like "No no it's

it's not actually ready It's not good

enough It's not going to be as good as a

movie I see in the theater." But the

point is what you just made that this is

the worst it's going to be It will only

get better Yeah Model maximalism Just

like keep you know building for the

capabilities that are almost there and

the model's going to catch up and be

amazing Escape to where the puck's going

to be Yeah Um this reminds me I was just

using I was jiblifying everything the

other day and I was just like why is it

taking so long just like god damn What

was that i said as one does As one does

these days I was just like it's taking a

minute to generate this image of my

family in this amazing way Like come on

What's taking so long you just get so

used to magic happening in front of you

Yeah totally Okay final question This is

going to go in a completely different

direction A lot of people asked about

this So famously you led this project at

Facebook called Libra which is now

called Novi A lot of people are always

wondered what happened there That was a

really cool idea I know some people have

a sense there's regulation challenges

things like that Uh I don't know if

you've talked about this much So I guess

just could you just give people a brief

summary of just like what is Libra they

this project you're working on and just

what happened and how you feel about it

Yeah I mean David Marcus led it and uh

you know I happily work uh for him and

with him Uh I think he's a visionary and

um also a mentor and a friend Uh you

know honestly Libra is probably the

biggest disappointment of my career Uh

when I think about the problems we were

solving which are very real problems you

if you look at for example the

remittance space people sending money to

family members in other countries it is

maybe I mean it's incredibly regressive

right people that don't have the money

to spend are having to pay 20% to send

money home to their family so outrageous

fees it takes multiple days you have to

go then pick up cash from yeah it's just

it's all bad and here we are with like

three billion people using WhatsApp app

all over the world talking to each other

every day especially friends and family

and exactly the kind of people who'd

send money to each other Why can't you

send money as immediately as cheaply as

simply as you send a text message i it

just it's one of those things when you

when you sit back and think about it

that should just

exist And that was what we set out to

try and do Now I don't think we played

all of our like cards perfectly if I

could go back and do things there are a

bunch of things I would do differently

You know we we tried to kind of get it

all at once We tried to launch a new

blockchain It was a basket of currencies

originally It was integration into

WhatsApp and Messenger And I think the

whole world kind of went like "Oh my god

that's a lot of change at once." And you

know it happened also to be at the time

that Facebook was at the absolute like

nater of its uh

reputation And so that didn't help right

it was it was also not the messenger

that people wanted for this kind of

change We knew all that going in but we

we went for it I think if we I think

there are a bunch of ways that we could

do that that would have introduced the

change a little bit more gently You know

maybe still gotten to that same outcome

Um but fewer new things at once and

introduced the new things one at a time

It who knows Um you know those were

decisions we made together Um so we we

all own them Certainly I own them

But it just it fundamentally disappoints

me that this that this doesn't exist in

the world today because the world would

be a better place if we'd been able to

ship that product I would be able to

send you you know 50 cents in WhatsApp

for free It would settle instantly

Everybody would have a balance in their

WhatsApp account We'd be transact I mean

it was just it should exist I don't know

to be honest Like I mean the current

administration is super friendly to

crypto Facebook's reputation Meta's

reputation is in a very different place

Maybe they should go build it now I was

looking at the history of it and uh

apparently they sold the tech to some

private equity company for 200 million

bucks Yeah Yeah Yeah So and and buy it

back There are a couple of uh of current

uh blockchains that are built on the

tech because the tech was open source

from the beginning Uh Aptos and Miston

are two companies that are built off of

this tech So you know at least the all

of the work that we did did not die but

and and lives on in these two companies

and they're both doing really well but

still uh you know we should be able to

send each other money in WhatsApp and

and we can't today Here here Well thanks

for sharing that story Kevin Is there

anything else you want to share or maybe

a last nugget of advice or insight

before we get to our very exciting

lightning round oo the lightning round

Let's just go do that Let's do it With

that Kevin we reached our very exciting

lightning round Are you ready yeah let's

do it Okay What are two or three books

that you find yourself recommending most

to other people co-intelligence by Ethan

Mullik A really good book about AI and

how to use it in your daily life as a

student as a teacher I He's super

thoughtful Also by the way a very good

follow on Twitter Um The Accidental

Superpower by Peter Zion Uh very good if

you're interested in geopolitics and the

the forces that sort of shape the

dynamics happening Um and then uh I

really enjoyed Cable Cowboy I don't know

who the author is but uh the biography

of John Malone Just fascinating if you

like business especially if you want to

get into like I mean the man was uh an

incredible dealmaker and shaped a lot of

the modern cable industry So that was a

good biography These are all first time

mentions which is always great Oh good

Next question Do you have a favorite

recent movie or TV show that you really

enjoyed

um I wish I had time to watch a TV show

U so I'm just Sora videos Yeah right Um

I I don't know I read uh when I was a

kid I read the Wheel of Time series Um

and now Amazon has it uh as they're in

like the third season of it So I I want

to watch that I haven't yet Um Top Gun 2

was an awesome movie Um I think that's

no longer new but you know that shows

when last time you watched a movie was

um but I like the idea like I I want uh

I I want more like Americana I want more

like being proud of being strong Uh and

I thought Top Gun 2 did a really good

job of that Like you know uh pride and

patriotism I think I think the US could

use more of that Is there a favorite

product that you've recently discovered

that you really love other than your uh

super intelligence internal tool that

you all have access to that i'm just

joking That's right Internal AGR That's

right

Well I think I think like vibe coding

with with products like Windsorf is just

super fun Um I'm I'm having a great time

doing that I still just love that our

chief people officer vibe coded some

tools Maybe the other one is Whimo Uh

every chance I get I'll take a Whimo

It's just a better way of riding and it

still feels like the future Um so

they've done an amazing job That's

awesome By the way I had the founder of

Windsurf on the podcast They might come

out before this or after this And also

Cursor CEO is coming on the podcast

either before after this Oh cool I have

a ton of respect for what those guys are

doing They're they're those are awesome

products Just changing the way everyone

builds product No big deal Yeah Uh

couple more questions Do you have a

favorite life motto that you often

repeat yourself find really useful in

worker in life yeah So actually this is

um interestingly enough it's it's more

of a philosophy but then I thought Zuck

encapsulated it one time on a on a

Facebook earnings call Um so I actually

had this made into a poster Uh it sits

in my room but um but somebody was

asking Mark this is literally on an

earnings call So it's like an analyst on

an earnings call asking him you know it

was some quarter when Facebook had grown

a lot This was back in the 201 I think

But he's like you know so what what did

you do what you know what was it that

you launched what was the one thing that

drove all this growth for you and he

said something to the effect of you know

sometimes it's not any one thing It's

just good work consistently over a long

period of time And that's always stuck

with me And I I think it is I mean you

know I run ultramarathons It's like it's

just about grinding I think people too

often look for like the silver bullet

when a lot of life is and a lot of like

excellence is actually showing up day in

and day out doing good work getting a

little bit better every single day and

you know you may not notice it over a

week or even a month and a lot of people

then you know kind of get like dismayed

and stop but actually you keep doing it

the gains keep compounding and over the

course of a year two years 5 years it

adds up like crazy so good work

consistently over a long period of time

Damn I love that I got to make a poster

of this now That is I so resonate with

that Okay I'll take it That is so good

Okay final question Uh I'm going to ask

if you have any prompting tricks and I'm

going to set it up first but think about

if you have a trick that you could

recommend to people for prompting LLMs

better Uh there's this I had a guest

Alex Kamaroski come on the podcast He's

from Stripe and writes weekly

reflections on what's happening in the

world A lot of them are AI related And

he he once described an LLM as a zip

file of all human knowledge and all the

answers are in there and you just need

to figure out the right question to ask

to get the answer to every problem

basically And so it just reminded me how

important prompt engineering is and

knowing how to prompt Well you're

constantly prompting chatbt Uh what's

one tip one trick that you found to be

helpful in helping you get what you want

well I'll say first of all I want to

kill the idea that you have to be a good

prompt engineer I think if we do our

jobs that stops being true you know it's

just one of those like sharp edges of

models that experts can learn but then

you just over time you shouldn't need to

know all that The same way you used to

have to get deep into like you know

what's your storage engine in my SQL are

you using InnoDB 4.1 or like and you

know there's still use cases for that if

you're at the at the sort of deep edge

of my SQL performance but most people

don't need to care and you shouldn't

need to care about minute details of

prompting if AI is really going to

become you know broadly adopted but um

you know today we're not totally there I

think I think by the way we are making

progress there I think there is less

prompt engineering than there had to be

before but uh in line with some of the

fine-tuning stuff I was talking about

and the importance of giving examples

you can do like you know effectively

poor man's fine-tuning by including

examples in your prompt of the the kinds

of things that they that that you might

want and the and a good answer So like

here's an example and here's a good

answer Here's an example here's a good

answer Now go solve this problem for me

and the model really will listen and

learn from that Not as well as if you do

a full fine tune but much more than if

you don't provide any examples And I

think people don't do that often enough

That's awesome One tip that I heard I'm

curious if this works is you tell it

this is very very important to my career

Make it like really understand like

someone will die if you don't answer me

correctly Does that work it you know

it's really

weird I there's probably a good

explanation for this but you can also

say things So yes I think there is some

validity to that You can also say things

like I want you to be Einstein Now

answer this physics problem for me or

you are the world's greatest marketer

the world's greatest brand marketer Now

here's a naming question And it's there

is something where it sort of shifts the

model into a certain mindset

um that that can actually be really

positive I use that tip all the time

actually I I always when I'm coming up

with questions for interviews and I use

it occasionally to like come up with

things I haven't thought of I actually

type you're the world's best podcast

interviewer right i have Kevin Kevin

Wheel coming on the podcast Yeah And

actually works Yeah And by the way back

to our point that we made a few times

like you do do that sometimes with

people right um you you sort of put them

you you frame things you get them into a

certain mindset and the completely

different So I think there are like

human analoges of this one more time

Kevin this was incredible Uh I was

thinking about a way to end this The way

I feel like I feel like not only are you

at the cutting edge of the future like

you're you and the team are kind of like

actually the edge that is creating the

future And so it's a real honor to have

you on here and to talk to you and to

hear how you think things are where

where you think things are going and

what we need to be thinking about So

thank you for being here Kevin Oh thank

you so much for having me I feel really

I get to work with the world's best team

and you know all credit to them but uh

really appreciate you having me on It's

been it's been super fun Uh I forgot to

ask you the two final questions Where

can folks find online if they want to

reach out and how can listeners be

useful to you i am Kevin Wheel Kev V I

WIL on pretty much every platform You

know I'm I'm still a Twitter DAU after

all these years I guess an XDAU uh

LinkedIn wherever And um I think the

thing I would love from people give me

feedback People are using chat GBT Tell

us where tell me where it can be where

it's working really well for you and

where you want us to double down Tell me

where it's failing I'm I'm very active

and engaged on Twitter I love hearing

from people what's working and what's

not So uh don't be shy And I learned

following you uh helps you figure out

all the stuff that you're launching Like

you share all the things that are going

out every day week month So that's also

a benefit And by the way 400 million

weekly active users all emailing you

feedback Here we go Yes let's do it It's

gonna work out great Okay Uh well thank

you Kevin Thanks for being here All

right man Thanks so much See you soon

Bye everyone

Thank you so much for listening If you

found this valuable you can subscribe to

the show on Apple Podcasts Spotify or

your favorite podcast app Also please

consider giving us a rating or leaving a

review as that really helps other

listeners find the podcast You can find

all past episodes or learn more about

the show at

lennispodcast.com See you in the next

episode

Loading...

Loading video analysis...