Sam, Jakub, and Wojciech on the future of OpenAI with audience Q&A
By OpenAI
Summary
## Key takeaways - **AGI is less than a decade away**: OpenAI's chief scientist estimates that deep learning systems could be less than a decade away from achieving superintelligence, surpassing humans across many critical axes. [04:26] - **AI to accelerate scientific discovery**: The most significant long-term impact of AI development will be its ability to accelerate scientific discovery and the development of new technologies, fundamentally changing the pace of progress. [05:00] - **AI researcher by 2028**: OpenAI aims to have a fully automated AI researcher capable of delivering on large research projects by March 2028, following an intern-level AI research assistant by September 2026. [07:43] - **Value alignment is key safety concern**: The most important long-term safety question for superintelligence is value alignment, ensuring AI fundamentally cares about high-level principles and humanity, especially as systems tackle problems beyond human ability. [08:52] - **OpenAI transitioning to a platform**: OpenAI aims to evolve from providing AI super assistants to becoming a platform where other people and companies can build on top of their technology, creating an 'AI cloud'. [15:11] - **Massive infrastructure investment planned**: OpenAI has committed over $1.4 trillion to infrastructure buildout, with aspirations to build an 'infrastructure factory' capable of producing 1 gigawatt of compute per week to meet future demands. [19:49]
Topics Covered
- Deep learning systems are less than a decade from superintelligence.
- Treating adult users like adults is essential for AI platforms.
- Scaling AGI requires trillions in new compute infrastructure.
- AI resilience broadens safety to an ecosystem-wide rapid response.
- AI will compress centuries of scientific discovery into years.
Full Transcript
Hello, I'm Sam. This is our chief
scientist, Yakob. And we have a bunch of
updates to share today about OpenAI. Um,
obviously the news of today is our new
structure. We're going to get to that
near the end, but there's a lot of other
important context we would like to share
first. Given the importance of a lot of
this, we're going to go into uh an
unusual level of transparency about some
of our specific research goals and
infrastructure plans and product, but we
think it's uh you know sort of very much
in the public interest at this point to
cover all of this. So our mission at
OpenAI in both the nonprofit and our new
PBC is to ensure that artificial general
intelligence benefits all of humanity.
As we get closer to building this, we
have new insights into what that is
going to mean.
There was a time earlier on in OpenAI
where we thought that AI would or AGI
would be sort of this oracular thing in
the sky and it would make all these
wonderful things for us and we now have
a sharper view of that which is we want
to create tools and then we want people
to use them to create the future. We
want to empower people with AI as much
as possible and then trust that the
process that has been working for human
history of people building um better and
better things with newer and better
tools will continue to go on. We we can
now see a vision where we help build a
personal AGI that people can use
anywhere with all of these different
tools, access to all these different
services and systems to help with work
and personal life and their personal
life. And as AI gets better and better,
as AI can even do things like discover
or help discover new science, what
people will be able to create with that
um to make all of society better and
their own lives more fulfilled, we think
should be quite incredible.
There are three core pillars we think
about for OpenAI. Research, product, and
infrastructure. We have to succeed at
the research required to build AGI. We
have to build a platform that makes it
easy and powerful to use. And we have to
build enough infrastructure such that
people can use at a low cost all of this
amazing AI that they'd like. Here's a
little like cartoon of how we think
about our world. So, at the bottom layer
here, we have chips, racks, and the
systems around them, the data centers
that these go into, and the energy. Uh
we'll talk more about the first three
today in energy another time. Then we
train models on top of these. Then we
have an open AI account on top of that.
We have a browser now uh called Atlas
and we have devices coming um in the
next few years that you'll be able to
take AI with you everywhere. And we have
a few first party apps like Chachi Pine
and Sor and we'll have more over time.
But mostly what we're excited about is
this big puzzle piece in the upper
right. We're finally getting to a world
where we can see that people are going
to be able to build incredible services
uh with AI starting with our API with
apps and chatbt new enterprise platform
that we'll have over time uh an open
account and way way more and people will
be able to fit all of the kind of like
current things in the world and many
more into this new AI world and we want
to enable that and we we believe the
world will build just a huge amount of
value for um for all of us. So that's
kind of what we see the economic picture
looking like. But one of the things that
we've thought about for a long time and
we really see happening now or starting
to happen now, glimmers of it, green
shoots, whatever you want to call it, is
the uh impact that AI will have on
science. And um although the economic
impact from that previous slide will be
huge for the long-term quality of life
uh and improvement that and change in
society, AI that can autonomously
discover new science or help people
discover new science uh faster will be I
think one of the most important things
and something that we're really trying
to wrap our heads around. So I'm going
to hand this over to Yakob to talk about
research and as I mentioned uh we're
going to share a lot about our internal
goals and our picture of where thing
where things are.
Thanks Sam. Um at the core we are a
research laboratory focused on
understanding a technology called deep
learning. And so a particular focus of
ours is um understanding uh what happens
as you scale up training deep learning
systems.
Um and one consequence we discuss a lot
there uh is AGI artificial general
intelligence. But we find that um in
some way even this maybe understates a
bit um the magnitude of um the possible
progress and change here. Um and so in
particular we believe uh that it is
possible that deep learning systems are
less than a decade away from super
intelligence. So systems that are
smarter uh than all of us on
uh a large number of of of of critical
axis
and
um this is of course a a serious thing
right there's uh a lot of implications
of this to grapple with and one uh
particular focusing impact of this
technology um and the technologies
leading up to it and something that we
organize our entire research program
around is the potential uh to accelerate
scientific discovery um to accelerate
the development of new technology.
Um
we believe that this will be uh perhaps
the most significant uh long-term impact
of AI development. um and it will
fundamentally change the pace of
progress on on on developing new
technologies.
Um and so thinking about how far along
we are uh uh towards these goals, uh one
good way to think about progress is to
look at the time horizon that it would
take people to accomplish the task that
the models can perform. And so this is
something that has been extending
rapidly over the past few years. Um so
where the current generation of models
is at right now is about five hours. So
um if you you can see this by looking at
the models matching the best people in
competition such as the international
Olympics or informatics
and we believe that this horizon will
continue to extend rapidly and this is
in part as a result of algorithmic
innovation and a part uh just scaling
deep learning further and in particular
scaling um along this um new axis um in
context compute also called test time
compute uh where we really see orders
and orders of magnitude to go. Um so
this is roughly how much time the model
spends thinking right and if you look at
how much time the model's currently
spent thinking about problems and if you
think about how much compute how much
time you would like to spend on problems
that really matter such as scientific
breakthroughs you should be okay using
entire data centers. Uh and so there is
there is really quite a way to go there.
Um and anticipating this progress uh we
of course make plans around it
internally and we want to provide some
transparency around our thinking there
and so we want to take this maybe
somewhat unusual step of sharing our
internal goals and goal timelines uh
towards these very powerful systems and
you know these particular dates we
absolutely may be quite wrong about them
uh but this is how we currently think
this is currently how how how we plan
and organize and So as a research
organization that is working on
automating research, naturally we are
thinking about how does this impact our
own work and uh how will AI systems that
accelerate development of future AI
systems look like? How can they empower
um research like alignment? And so uh we
are making plans around getting to quite
capable AI research interns that can
meaningfully accelerate our researchers
by expanding
a a significant amount of compute um by
September of next year. So we believe
that is actually quite close. Um and
then we uh look towards getting a system
capable of autonomously delivering on
larger research projects and a
meaningful uh fully automated AI
researcher uh by March of 2028.
Um and so of course as we look towards
these very capable systems uh we think a
lot about safety and alignment right and
in fact a lot of our work uh both on
deployments and safety uh but also just
on on on understanding deep learning and
development capability side we can think
of preparation for these very capable
models. Um safety is a multiaceted
problem and so the way we generally
structure our thinking uh are these five
layers ranging from uh factors that are
most internal to the model to ones that
are most external. And so at the core
right and what we believe is the most
important uh long-term safety question
for super intelligence is value
alignment.
So to put this um
this
value alignment uh you can think of as
what is what is really the thing that
the AI fundamentally cares about uh can
it adhere to some high level principles
u what will it do if it's given unclear
and conflicting objectives u does it
laugh humanity
um and the reason we believe that kind
of this this high level uh um
um
objectives or or or or principles
driving the AI are so important is that
as we get to the systems that are
thinking for very long, uh as they
become very smart, as they tackle
problems that are uh um at the edge or
perhaps beyond human ability, uh really
getting to like complete specifications
becomes quite difficult. Uh and so we
have to rely on this on this deeper
alignment.
Um then there's goal alignment. Um, does
the agent uh interact with people? How
does it interact with people? How does
it um do it following instructions? Um,
then reliability. Um, can the AI
correctly calibrate its predictions? Uh,
can it be reliable on easy tasks,
express uncertainty on hard ones, can it
deal with environments that are a little
bit unfamili unfamiliar?
Um then we have adversarial robustness
which is very related to reliability but
it's about adversarial settings. So can
can the AI withstand target attacks from
human or AI adversaries.
And then the outer layer is systemic
safety which are guarantees about the
behavior of the overall system that do
not rely on the AI's intelligence or
alignment. Um so for example this can be
security or what data does the AI have
access to? um or um um uh um what
devices it can use. And so we invest in
multiple research directions across
these domains. And we have seen uh quite
a lot of progress also come from just
the general uh development and improving
understanding of deep learning uh as a
whole. And uh I want to I I want to take
a slightly deeper technical dive here.
Um and talk about a particular
direction. Uh value alignment is a hard
problem, right? It's definitely not
solved yet. Um however, there is a new
promising tool that aids our study of
it. Uh and that is chain of fun
faithfulness. Uh it's something we
invest in very heavily. Um, starting
from our first reasoning models, we've
been pursuing this new direction in
interpretability. And the idea is to
keep parts of the model's internal
reasoning free from supervision. So
don't look at it during training and
thus let it remain representative of the
model's internal process. Um,
so we refrain from from from from kind
of guiding the model to think good
thoughts and and and and and so let it
let let it remain a bit more faithful to
to to what it actually thinks, right?
And this is not guaranteed to work, of
course, right? We cannot make uh
mathematical proofs about deep learning.
And so this is something we study. Uh
but there are two reasons to be
optimistic. One reason is that we have
seen very promising empirical results.
Uh this is a technology we employed a
lot internally. Uh we use this to
understand um how our models u um train
h how their propensities evolve over
training. Uh also we have had successful
external collaborations on investigating
the models propensity scheme for
example.
Um and secondly uh it is scalable and in
the sense that explicitly we make the
scalable objective not adversarial to
our ability to monitor the model. Um
and of course an objective not being
adversarial to the ability to monitor
the model is only half the battle. Um
and you know ideally you want it to to
to get it to help with monitoring the
model. And so this is something we're
we're researching quite heavily.
Um but one important thing to underscore
about train of thought faithfulness is
it's somewhat fragile. Um it really
requires drawing this clean boundary uh
and having this clear abstraction uh and
having restraint in in what ways you can
access the chain of thought and this is
something that is present uh at OpenAI
from algorithm design to the way we
design our products right so so if you
look at the chain of thought summaries
in chat GPT uh if we didn't have the
chain of summarizer if we just make the
chain of f fully visible at all times
right that would make it kind part of
the overall experience over time it will
be very difficult to not subjected to
any supervision.
Um and so longterm we believe that by
preserving some amount of this
controlled privacy for the models uh we
can retain the ability to understand
their inner process and we believe this
can be a very impactful technique uh as
we move towards these very capable
longunning systems. Um and I'll hand
back to Tom.
>> Okay, that's very hard to follow uh with
the rest of this and obviously that's
the most important part of what we have
to say. But um you know just to to
reiterate uh we may be totally wrong. We
have set goals and missed them miserably
before. But with the picture we see we
think it is plausible that by September
of next year we have sort of a intern
level AI research assistant and that by
March of 2028 which I believe is almost
5 years to the month after the launch of
GPT4. um we have like a a legitimate AI
researcher and this is the core thrust
of our research program. There are two
other areas we want to talk about uh
product and then infrastructure. On the
product side,
as we make this incredible progress with
deep learning, we want to make it useful
to people to sort of invent the future
as we mentioned. And um what that's
looked like traditionally for us as an
AI super assistant inside of Chacht, but
we're now really going to evolve to a
platform uh that other people will build
on top of and all of the pieces that
need to fit in uh of the world um you
know will be built by others. Before we
go talk about that, wanted to just show
a quick video of how people are using
GPT5. Some of the ways people are using
GPT5 in Chacht today.
>> I'm a quantum physicist.
>> I'm a nail technician, human
iminologist.
>> I'm a steel worker,
>> a designer and developer, a professor of
economics.
>> I basically go fishing for a living.
>> GPT5 is able to predict the outcomes of
experiments that we haven't even done.
Can you create a part numbering system
that's easy for my guys in the shop?
>> I want to catch dungeonous crab in the
Bay Area. Here I would ask a question
about the application of a certain
quantum operator. This one gives me a
very detailed mathematics.
>> I ask for basically a camera app where I
can draw real time in the air.
>> We have a theme. We have a direction to
go. Let's go from a million or infinite
amount of ideas to like give me 10 or
20.
>> You just start testing it. You know, you
tell it jokes. You ask it questions like
what should economists do?
>> Different baits I can use, different
water depths, all this information that
it would take years to figure out on
your own. A lot of trial and error. It's
tremendous for brainstorming. It's back
and forth.
>> I can kind of easily follow the
reasoning. I don't need to trust the
result. I can just look what did you do?
>> GBD5 just like did this um in one shot.
>> A 101 part numbers. It would have taken
weeks for me to number this and it would
have made me go crosseyed.
>> Uh Yakob will join me back for Q&A in a
little bit, but we're going to have one
special guest on before before the end.
Um we love that. We want much more of
that. We want that everywhere. So we
want open AI to be a platform that
people and companies can build on top
of. We can sort of see our way now to an
AI cloud where this is not just in chat
GPD. This is not just services that we
create, but we want to expose our
technology for as many people to build
the things that people will depend on
and use and create with as possible. I
think this this quote or at least this
like idea is originally from Bill Gates.
um at least that's where I first heard
it that you know you've built a platform
when there's more value created by
people building on the platform than by
the platform builder builder and that's
our goal um and that's our goal like
next year we we really think we can now
take this technology and this user base
and this sort of framework we've built
and get the whole world to build amazing
new companies and services and
applications on top of it
to do that uh there will be many things
that we have to evolve towards but
there's two foundational principles as
we as we move towards being in this
platform that I wanted to touch on. Um,
one is about user freedom. If this is
going to be a platform that all sorts of
people are building, uh, on using, um,
creating with, people around the world
have very different needs and desires
and there will of course be some very
broad bounds, but we want users to have
a lot of u control and customization of
how they of how they use it. Now, I
made, you know, one of my many stupid
mistakes when I tried to talk about this
recently. Uh, I wish I had used an
example other than erotica. thought
there was an understandable difference
between erotica and pornbots. But in any
case, we were trying to show the point
we're trying to get across is that
people need a lot of flexibility and
people want to use these things in
different ways and we want to treat our
adult users like adults. In our own
first party services, we may have, you
know, tighter guidelines, but AI is
going to become such an important part
of people's lives. The freedom of human
expression is going to need to be there.
Along with that, we think that world
will need to think about privacy in a
different way than they have for
previous kinds of technology. Privacy is
important for all sorts of technology,
of course, but privacy for AI will be
especially important. People are using
this technology in a different way than
they've used the technologies of the
past. They're talking to it like they
would to their doctor, their lawyer,
their spouse. Um, they're sharing the
most intimate details of their lives.
And of course, we need strong technical
protections on that privacy of that
privacy. But we also think we need
strong um policy protections of that
privacy. We've talked about concepts
like AI privilege. Um but really strong
protections if AI is going to be this
fundamental platform in people's lives
seem super important to us.
Okay. And then I want to go on to
infrastructure. So I know there's been a
lot of confusion about sort of where we
are in our infrastructure buildout and
we figured we would just be super
transparent about that. So where we are
today, um all of our commitments total a
little bit over 30 gawatts of uh of
infrastructure buildout. Um and that's
about a $1.4 trillion total uh financial
obligation for us over the next many
years. Um this is what we've committed
to so far. We of course hope to do much
more, but given the picture we see
today, given what we think we can see
for revenue growth, our ability to raise
capital, this is what we're currently
comfortable with. This requires a ton of
partnerships. We've talked about uh many
of our great chip partners. Uh there are
people building the data centers for us,
land, energy. Uh there will be chip fab
facilities. This is already getting to
require quite a lot of supply chain
innovation. And we're thrilled to get to
work with uh AMD, Broadcom, Google,
Microsoft Nvidia Oracle SoftBank
many others to really make this happen.
But this is still early. if the if the
work that Yakob talks about comes to
fruition, which we think it will, and if
the economic value of these things um
happen and people want to use all these
services, we're going to need much more
than this. So, I want to be clear, we're
not committing to this yet, but we are
having conversations about it. Our
aspiration is that we can build an
infrastructure factory where we can
create 1 gawatt a week of uh compute and
we aspirationally would like to get that
cost down significantly um to like $20
billion a gigawatt over the 5year life
cycle of uh you know that equipment. To
do this will require a ton of
innovation, a ton of partnerships,
obviously a lot of revenue growth. Um
we'll have to repurpose our thoughts
about robotics to help us build data
centers instead of doing all the other
things. Um, but this is where we'd like
to go and over the coming months we are
going to do a lot of work to see if we
can get here. Um, it will be some time
before we're in a financial position
where we could actually pull the trigger
and get going on this. 1 gawatt is like
a big number, but I figured we would
show a little video to put this into
perspective. Um, this is a data center
that we're building in Abene, Texas.
This is the first Stargate site. We're
doing several of these now around the
country, but this one is the furthest
along. There's like many thousands of
people that work here every day just
doing the construction at the site.
There's probably hundreds of thousands
or millions of people that work in the
supply chain to make all this happen to
design these chips, to fab these chips,
to put them together. Um there, you
know, there's all of the work that goes
into this for energy. There's an
enormous amount of stuff that has to
happen um for each one gigawatt. And we
want to figure out how we can make this
way more efficient and way cheaper and
way more scalable so that we can deliver
on the infrastructure that the research
roadmap requires and that all of the
ways that people will want to use this
need um to enable that uh we have a new
structure. So um maybe you saw before
this like crazy convoluted uh diagram of
all of the open entities. Now it's much
simpler. We have a nonprofit called the
open air foundation that is in control
of uh where the board sits uh or where
uh let's come back to the board uh where
the board also sits and uh owns a slice
of our PBC public benefit corporation
called OpenAI group. So uh nonprofit and
control public benefit corporation sits
under it. Um we hope for the OpenAI
Foundation to be the biggest nonprofit
ever. As I mentioned now uh a few times,
science is one of the ways that we think
the world most improves along with the
institutions that broadly distribute the
benefits of that. So the science will
not be the only thing that the nonprofit
funds, but it will be an important first
major area of the things that we do. The
nonprofit will govern the PBC. It will
initially own about 26% of the PBC
equity, but that can increase over time
with warrants if if we perform really
well. and it will use these resources to
uh pursue what we think are the best
benefits of AI uh given where the
technology is and what society needs.
The PBC uh will operate more like a
normal company. It will have the same
mission. Um it will be bound to that
mission. Um and you know in matters of
safety will only be bound to that
mission but it will be able to attract
the resources that we need for that
gigantic infrastructure buildout to
serve the research and product goals
that we have. So, the initial focus of
the the foundation, we'll do more things
over time, but we want to knock
something out of the out of the park,
uh, hopefully first, is a $25 billion
commitment to use AI to help cure
disease. There are a lot of ways this
can happen. Um, generating data, um, you
know, using a lot of compute grants to
grants to scientists and also for AI
resilience. Um AI resilience is a new
and I think very important area and I'd
like to invite uh our co-founder of
Voyche up to talk about what this will
what this will look like.
>> Hello
glad to be here.
>> Thanks for being here.
>> So um term AI resilience is um little
bit broader than what we historically
thought about AI safety. So in case of
the uh resilience we think that advanced
AI comes with risks and disruptions and
we would like to have an ecosystem of
organizations that uh can help to um
solve a number of these problems. Um so
let me give you an example to better
illustrate it. Um we all believe that AI
uh will advance in biology and as it
advances in biology there is a there is
a risk that some bad actor could use AI
to create manmade pandemics. So the on
the safety level the mitigation would be
to make sure that the models uh block
the queries that have to do with a
viology. Um however um if you consider
the entire entire um AI uh industry it's
very likely that even if open AI blocks
it someone could use uh different models
out there and still uh produce pathogens
and um the in case of resilience we
don't want just to block it but also
have a uh rapid response if the problem
would occur. So when I think about the
uh risks uh and disruptions there are
just many the the mental health is one
of them bio is another one another one
uh job displacement might be another one
and we think that we need the ecosystem
and maybe a good analogy that I like is
cyber security so at the beginning of
internet
uh it was actually it was a place that
people didn't feel comfortable putting
their credit card numbers because it was
so easy to get hacked and uh when there
was a virus people were giving each
other a call to disconnect the computer
from internet and we got a long way at
the moment there's entire infrastructure
of cyber security companies they are
protecting uh um the critical
infrastructure governments corporations
and individual users to such extent that
people are willing to put the most
personal data uh online to have life
savings um be online. Um yeah, so the
the cyber security got really far and we
think that something analogous will be
present for AI that there will be AI
resilience layer and uh I'm really
excited that the uh nonprofit uh will
help out uh to stimulate it to create
such an ecosystem.
>> So am I. I think this is an important
time to be doing this and I'm very
excited to that you're going to like
figure out how how we how we go off and
make it happen. So again, these are not
the only things that the nonprofit will
fund, but we're excited about these as
the first two using AI to develop cures
and uh treatments for diseases and this
new AI resilience effort as we figure
out what the deployment of AGI into
society is going to look like.
So we mentioned that those are our three
pillars, but you know what what if this
all works? We we think it is plausible
that in 2026 we start to see the models
of that year begin to make small
discoveries. By 2028 medium or maybe
even larger discoveries and you know who
knows what 2030 and 2032 are going to
look like. If AI can keep advancing
science as has happened in the past um
we think the future can be very bright.
Of course, we think it's very important
that humans can self-determine our way
through this future. But the open space
that new scientific advances give us is
quite impressive. So, we asked Sora to
help us imagine a radically better
future by looking at the past. And we
are particularly interested in how the
history of science builds on itself
discovery after discovery. This is what
we hope will happen with AI. You know,
this is going to be 200 years of
science. But if you can do these 200
years of compounding discoveries of the
scaffolding building up on each other
not in 200 years but in 20 years or in
two years and if you look at how much
this this is accelerated think about
what could be possible before you can
imagine a world uh where a radically
better future becomes quite possible.
You have a data center here that is
discovering a cure for cancer. A data
center there that's making the best
entertainment ever. A data center here
that's helping you find your future
husband or wife. This one is building
rockets and helping you colonize space.
This one is helping to solve the climate
crisis. So, we did all this stuff the
oldfashioned way. Um, and now with the
help of AI, we'll be able to shape what
comes next, uh, with maybe much more
power. So, we talked a little about AI
medicine. We're very excited about
robots. Um, we really think energy is
very, very important to the world. We
want to figure out what personalized
education can mean, design novel
materials, and probably a ton of other
things that we can't even think of yet.
So as we head into this next phase of
open AI and more importantly than that
this continual progress in deep learning
um we thank you for joining us today and
we're going to try something new now
which is we're going to just answer
questions. If this works it's something
we are we'll try more in the future.
Yakob is going to rejoin for this Q&A.
Thank you very much.
>> Um but uh this is a new format for us.
So bear with us as we try it this first
time again. If if um if this is useful
it's something we'll do uh again a lot
more. and we're going to try to just
answer questions in the order they are
most upvoted. Um,
are we good to go? All right, let's see
how this works. So, uh, you can put
questions in the Vimeo link and we will
just start answering them. So, from
Caleb, we've warned that the tech is
becoming addictive and eroding the tech
is becoming addictive and eroding trust
yet Sora mimics Tik Tok and Chach may
add ads. Why repeat the same patterns
you criticized and how will you build
build rebuild trust through actions and
not just words? We're definitely worried
about this. Uh I worry about it not just
for things like Sora and Tik Tok and ads
and chatbt which are maybe known
problems that we can design carefully
but you know we have certainly seen
people develop relationships with chat
bots that we didn't expect and there can
clearly be addictive behavior there
given the dynamics and competition in
the in the world. I suspect some
companies will offer very addictive new
kinds of products. Um and I think you'll
have to just judge us on our actions.
We'll have to you know we'll make some
mistakes. We'll try to roll back models
that are problematic. If we ship Sora
and it becomes super addictive and not
about creation, we'll, you know, cancel
the product and you'll you'll have to
just judge us on that. My hope and
belief is that we will not make the same
mistakes that companies before us have
made. Uh I don't think they meant to
make them either. It's uh you know,
we're all kind of discovering this
together. We probably will make new ones
though and we'll just have to evolve
quickly and have a tight feedback loop.
We we can imagine all sorts of ways this
technology does incredible good in the
world. also obvious bad ones and um you
know we're guided by a mission where
we'll just continuously evolve evolve
the product. Um one thing that we are
quite hopeful about for in terms of um
what we optim optimize for in in uh
products like chat GPT or or Sora is
thinking about um optimizing for the
very long term which is naturally very
aligned with uh how we think in general
about extending uh the horizon on which
the models can work productively. Um
and so we believe that quite a lot of
development is possible there and we can
eventually get the models that really
optimize for um long-term uh
satisfaction and and and well-being
instead of just short-term signals.
Um okay, next question. Will we have an
option to keep the for model uh
permanently after adult mode is
installed? We don't need a safer models
responsible adults. Um,
we have no plans to sunset 40. Uh, we
are not going to promise to keep it
around till the heat death of the
universe either, but we we understand
that it's a product that some of our
users really love. We also hope other
people understand why um it was not a
model that we thought was healthy for
miners to be using. Um, we hope that we
build better models over time that
people like more. You know, the people
you have a relationship with in your
life, they evolve and get smarter and
change a little bit over time. And we
think that we hope that the same thing
will happen. But yeah, no, no plans to
uh no plans to sunset 40 currently.
Uh wow, we have a lot of for questions.
All right, we're not going to in the
interest of time, we will not go through
uh all of these, but but yeah, we don't,
you know, we want people to have models
that they want to use. We don't want
people to feel like we're routing them
around models. Um, and you know, we we
want adults to make choices as adults
as long as we think we're not, you know,
selling heroin or whatever, which also
you you shouldn't you shouldn't do. Um,
so people that want to have emotional
speech, uh, as we've said, we want to
allow more of that and we plan to. Okay,
here's a good anonymous question for
Yakob. When will AGI happen?
Um
so
I think I think
in in some number of years we'll look
back at these years and we'll say you
know this was kind of the transition
period when AGI happened. Um I think you
know what one way we thought about
um I think as as some said like early on
adopting we thought about AGI kind of
emotionally as this like thing that is
like the kind of ultimate solution of
all the problems and and and um it's
it's this like single point um for which
there is before and after and um I think
um we found that it's a bit more
continuous than that. Um
and and so in particular for like
various kind of benchmarks that you know
seemed at uh um
seem like kind of the obvious like
milestones towards AGI. I think I think
we now think of them as kind of like
indicating like you know roughly how far
away we are in years. And so uh you know
if you look at a succession of of of of
milestones such as computers beating
humans at chess and then at go and then
uh you know computers being able to
speak in natural language and computers
being able to solve math problems right
I think well they clearly kind of get uh
closer together. Um
yeah, I I would say I think it's the AGI
term has become hugely overloaded and as
Jakob said, it'll be this process over a
number of years that we're in the middle
of. Uh but one of the reasons we wanted
to present what we did today is I think
it's much more useful to say our
intention our goal is by March of 2028
to have a true automated AI researcher
um and define what that means uh than it
is to sort of try to you know satisfy
everyone with a definition of AGI
>> and and maybe one other thing to mention
right like I think like one kind of
counterintuitive thing here is that
obviously we're working with like a
pretty complicated technology we're
trying to understand all these
algorithms and maybe initially we kind
imagined that like AGI is the moment
once you where you kind of have figured
out all the answers and and and and it's
kind of the the final thing and I think
now we increasingly realize that you
know there is kind of a some some curve
of intelligence maybe a
multi-dimensional one and you know
humans are somewhere on it and as you
scale deep learning as you develop these
new algorithms like eventually well you
kind of inch closer to that point and
eventually and eventually will surpass
it and you know already have surpassed
on multiple axis um and and and so and
that doesn't actually mean you have
solved all the problems around it which
is something we we need to seriously
think about.
>> Can you give us an idea of how far ahead
internal models are compared to deployed
ones?
>> Um
I think
um
we we have quite strong expectations for
our for for for our next models. Um so I
think I think we we expect quite rapid
progress over the next uh uh couple
months and a year. Um
yeah I think um but we
we haven't been like withholding
something something uh extremely crazy.
>> Yeah. One of the ways this uh kind of
often works in practice is there's like
a lot of pieces that we develop and that
you know they're all kind of hard one
victories and then we know that when we
put them together um we will have
something quite impressive and we're
able to predict that fairly well. Uh
part of our goal today is to say that um
we have a lot of those pieces. It's not
like we're kind of currently sitting on
this giant, you know, thing that we're
not showing to the world, but that we
expect by a year from now, certainly
with this September of 2026 goal
that we have a a like I mean not likely,
we have a realistic shot at a like
tremendously
important step forward in capability.
Um, what is OpenAI? Ronin asks, "What is
OpenAI's stance on partnering with labs
like Anthropic, Gemini, or XAI for joint
research, compute sharing, and safety
efforts?" Uh, we think this is going to
be increasingly important on the safety
front. Uh, labs will need to share
safety techniques, safety standards. Um,
you can imagine a time when the whole
world would say, uh, okay, before we hit
a recursive self-improvement phase, we
really need to all carefully study this
together. Um, we welcome that
collaboration. I think it'll be quite
important. Um one thing to mention on
the on the on chain of thought
faithfulness that I talked about earlier
um we actually have um started talking
about establishing industry arms and and
we started some some joint
investigations with researchers from
Google and entropic uh and and and some
other labs and uh yeah that's something
I'm very excited about and I think that
is an example of something where we can
really benefit from collaborating across
multiple labs.
Anonymous asks, "Will you ever open
source some of your old models like the
original GPT4?" Um, we might do those as
museum artifacts someday, but they're
not like GPT4
is not a particularly useful open source
model. It's big. It's not that good. Uh,
you know, we could probably make
something that is beyond the power of
GPT4 at a very tiny scale. Uh, that
actually would be useful to people. So,
for useful things, I expect more things
like that. For uh for fun museum
artifacts, yeah, someday, who who knows?
Like I think there could be a lot of
cool things like that. Another anonymous
or maybe the same one asks, "Will you
admit that your new model is inferior to
the previous one and that you're ruining
your company with your arrogance and
greed while ignoring users needs?" Um,
I believe that it is inferior to you for
your use case and we would like to build
models that are better for your use
case. on the whole uh we think for most
users it's a it's a better and more
capable model but we definitely have
learned things about the 40 to5 upgrade
and we will try to do much better in the
future both about better continuity and
about making sure that our model gets
better for most users not just sort of
people that are using AI for science or
coding or for whatever
uh Y asks will there ever be a version
of chatbt meant for personal connection
and reflection not only business or
education yeah for sure this is this
We think this is a wonderful use of AI.
Um, we're very touched by how much this
has meant to people's lives. We get all
of us get a ton of emails and outreach
from users about how Chacht has helped
people in difficult personal situations
or to live a better life. And like this
is what we're here for. I mean, this is
like as important as anything that we
do. Like we we love to hear about
scientific progress. We love to hear
about people that, you know, got
diagnosed with the disease and got
cured.
The personal stories are incredibly
important to us and we're thrilled about
that and we absolutely want to offer uh
such a service.
Your safety oh
uh okay two parts of this two questions
that are about tide uh from G and
anonymous. Your safety routing breaks
user trust and workflows by overriding
our choices. Will you commit to revoking
this paternalistic policy for all
consenting adult users and stop treating
us like children? Uh when do users get
control over routing? Where's the
transparency on safety and censorship?
Why can't adults pick their own models?
Yeah, I I don't think the way we handled
uh the model routing was our best thing
ever. There there are some real problems
with forro. Um and we have seen a
problem where people are forming people
that are in fragile psychiatric
situations using a model like 40 can get
into a worse one. um most adult users
can use those fine, but we do, as we've
mentioned, we have an obligation to
protect minor users. And we also um have
an obligation to protect adult users who
are not in a frame of mind where, you
know, we're reasonably likely that
they're choosing what they really want
and we're not causing them harm. as we
build age verification um in and as we
are able to differentiate users that are
having like a true um mental health
crisis from users who are not uh we of
course want to give people more user
freedom as we mentioned that's one of
our our platform principles um so yes
expect improvement there and I don't
think this was our best work and how we
communicated the previous roll out
how we strike the right balance between
protecting people and allowing adults to
speak about difficult things without
feeling policed
you want to say anything there?
Yeah. So, so,
so the definitely there is a problem
where um we we we
aim to lay out the kind of high high
level um
policies and and guidelines for the
model in in in the spec uh that we um
that we develop for for CH GBT. But um
the the the the
space of of of of of situations you can
find yourself in is is is is enormous
and and and um and at some point kind of
like establishing the the right
boundaries really becomes a a tough
intelligence problem. Um and so we are
seeing uh
improved results on this matrix from
reasoning models and from expanding uh
more more reasoning on on thinking about
this uh uh software uh software
questions and and and trade-offs. Um,
of course, this is like a bit more
difficult to train for also than uh math
problems, for example. And and so this
is something that we're uh researching
quite heavily.
Uh Kate says, "When in December will
adult mode come? Will it have more than
just NSFW? when writing even slight
conflict when writing triggers filters
on for uh I don't know exactly when in
December it will ship but yes the goal
is um when you are writing when you are
using open AAI to help chat to help you
with creative writing um you it should
be much more permissive in in many
categories than the previous uh models
are again we want this and we know users
want this too um if this is going to be
your personal tool it should help you
with what you're doing and every time
you hit a content filter for something
that
uh you know feels like it shouldn't. We
we understand how annoying that is. So
uh we are uh we're going to try to fix
that with adult mode. There may be new
problems that we face, but we want to
give people uh more flexibility.
Uh anonymous says, "Why does your idea
of safety require lying to users about
what model they're actually using?" Uh
again, I think we mis rolled this one
out, but the goal here was to let people
continue to use Forro, but in the
situations where Foro has behavior that
we think is actually really harmful
before we have all of the age gain in
that we'd like to kick it uh to to put
the user into a model where they are not
going to have some of the mental health
problems that we faced with for um FORO
was an interesting challenge. It's a
model that uh some users really love and
it was a model that was causing some
users harm that they really didn't want
and I don't think this is the last time
we'll face challenges like this with a
model. Um but we are trying to figure
out the right the right way to balance
that.
Um will we legacy models back for adults
without rewriting rewriting? Yes. Uh, y,
will the December update officially
clarify OpenAI's position on human AI
emotional bonds? Uh, or will
restrictions continue implicitly
defining such connections as harmful
worldwide?
I I don't know what it means to have an
official position like we build this
tool, you can use it the way you want.
If if you want to have like a small R
relationship and you're getting
something like empathy or friendship
that matters to you and your life out of
a model, like it's very important to us
that the model faithfully communicate
what it is and what it isn't. But if you
as the user are finding value in that
support, again, we think that's awesome.
Like we we are very touched by the
stories of people who find
value, utility, a better life in the
emotional support or other kinds of
support they get from these models.
Kylo says, "How is OpenAI increasingly
allowing so many features for the free
version users?" I can answer this from a
product and business perspective, but
Jakob, I think it might be useful for
you to just talk about the incredible
rate at which models are getting more
capable for lower prices and less
amounts of compute. Um yeah, we we are
seeing quite a lot of uh
um
ability as we get to like the new
frontiers of intelligence to to to
reduce the cost for that quite quite
quickly. Um
and so yeah, especially especially with
reasoning models, we've seen uh that
actually quite uh quite cheap models
when allowed some some additional test
time compute can become much more
capable. Um
yeah, and so this is uh this is
something that I expect will continue.
And so um yeah, as we talk kind of about
about um you know, getting to these um
new frontiers and automating research
and so forth, I I I expect that the cost
of of of a lot of that will will keep
falling quite a lot too. So yeah, our
our
we talk a lot about the increase in
model capability and for you know
pushing forward science that's that's
hugely important. One of the most
amazing things that I've observed about
AI is over the last few years, the sort
of price of a particular unit of
intelligence has fallen about 40x per
year for the last few years. Um, so when
we first,
you know, when we first had like GPT3,
we thought it was very cool and it was
at this cost that was kind of hard and
like GP3 scale models now basically run
for free like on a phone or something.
um the cost of a model that's as smart
as GPT4 at the time we launch it
relative now has fallen hugely and we
expect this trend to keep going. Now we
still think we need a ton of
infrastructure because what we continue
to find is the cheaper we can make it
the more people want to use it and I
expect that only to increase but our
goal is to drive the you know cost of
intelligence down and down and down and
have people use it for more and more
things that will allow us to continue to
offer lots of features for free. Um, but
that will also mean, I think, that
people who really want to spend a lot on
pushing AI to the limit to cure a
disease or figure out how to build a
better rocket or whatever will spend a
huge amount. Um, we are committed to
continuing to put the best technology we
can uh as long as we can make the
business model even sort of work into
the free tier and you should expect a
lot more from us there over time.
Uh, okay. Anonymous asks, "Will an age
will an age verification start that
allows users to opt out of the safety
route or a waiver that could be signed
releasing open air from any liability?"
Um, we're not going to like do again,
we're not going to do the equivalent of
selling heroin or whatever, uh, even if
you sign a liability. But yes, on the on
the principle of treat adult users like
adults, if you're age verified, you will
get quite a lot of of flexibility. We
think that's important and clearly it
resonates with the people asking these
questions. Um, anonymous also asks, "Is
chatbt the ask jeieves of AI?" We sure
hope not. We don't think it will be. Uh,
okay. Since we only have 10 minutes
left, we're going to um, and some of
these touch other things that we've uh,
we've already touched on. We're going to
skip down through some of the same
questions and try to get to more. Uh, in
future qu sessions, we can um, do more
of these if we don't get to everything
here. Um,
just as the Macintosh, no asks, just how
the Macintosh was the precursor to the
iPhone, do you see Chacht as the OpenAI
product or do you see it as a precursor
to something much greater that truly
reshapes the world?
Um,
so I would say like as a research lab,
um,
well, we haven't set out to build a a
chatbot originally, although I think
we've since come to appreciate how
aligned this this product is with with
our overall mission. Um, and we of
course expect Chad GPT to continue to
become better and and and and be this
way um for people to interact with
increasingly advanced AI. Um but we do
anticipate that uh eventually AI systems
will be capable of um
creating valuable artifacts of actually
pushing scientific progress forward as
we were discussing and um I believe that
will be the the real lasting legacy of
AI. I I think the chat interface is a
great interface. It won't be the only
interface, but the way that people use
these systems um will
change hugely over time. If you think
about what Yakob shared earlier of the
5-second, 5 minute, 5 hour tasks, um if
you think about a 5year or five century
task that would take something that
would take humans, it's hard to even
think about what that means. But
probably you want a different kind of of
product experience. I also think you
probably will want this to feel more
like a sort of ambient always present
companion. Like right now you can ask
Chacht something. It can do something
for you. But it'd be really nice to have
a service that was sort of just
observing your life and proactively
helping you when you needed it and you
know helping you come up with better
ideas and just I I think we can probably
push very hard in in that direction.
Um Neil asks, "I love GPT 4.5. It's by
far the best on the market for writing
and it's the main reason I pay for Pro.
Could we get some clarity on its future,
please?
Um, we think we're going to have models
that are much better than 4.5 very soon
uh and for writing much much better. We
plan to keep it around until we have a
model that is a huge step forward in uh
writing. But, you know, we'd like to we
don't think 4.5 is that good anymore.
We'd like to offer something much much
better.
>> But yeah, we are definitely not done
with that direction of research. Uh and
yeah, we we expect combining combining
that with with uh other things we're
working on, we'll get wheels are
dramatically better than 4.5 on all
axis. Do you have uh any sense of timing
to share about when you think we have a
model that is dramatically better than
4.5 on on this kind of task like writing
and also anything about like how far
that's going to go?
>> Um next year I think is definitely
what I expect.
Um, when is Chach Atlas for Windows
coming? asks Lars. Uh, I don't know an
exact time frame. Some number of months,
I would guess. Uh, it's it's definitely
something we want to do. And more
generally, this idea that we can build
experiences like browsers and new
devices that let you take AI with you
that get towards this sort of ambient
always helpful uh assistant rather than
something you just query in response. uh
this will be a very this will be a very
important direction for us to to push
more on.
Will you disclose the documents with the
opinions of the 170 experts so that
we'll have some transparency regarding
the new justifications for model
behavior? Um I will ask Fiji Simo uh how
she'd like to handle that, but I think
we could I don't know exactly what we'll
be able to share, but I think we should
do something there. And I I think more
um more transparency there is a good
thing. Anonymous says, "I've been a
prouser since month two. As a researcher
and fiction writer, I feel GPT helps me
think clearer. I lost the question." Um,
sorry, that was a really good question.
Let me try to find it again.
Uh, clearer but not freer. Has
imagination become an optimization
casualty?
What do you think?
Um
it is
I think it is definitely possible for
current systems that uh you know like I
think if you compare a model like 4.5 to
a model like 03 I would expect that like
there will be trade-offs there. I I
think there are definitely like
transitory as we like figure out our way
around these technologies and so again
like I expect this will this will get
better. Yeah, I
I I think there are going to be
population scale like one one of the
sort of strange things I've noticed is
people in real life talking in like
CHBTE
uh where they sort of use some of the
quirks of of things ChachiBT says and I
I think there will be other things like
this where there's like this
co-evolution of people and the
technology in ways we can't totally
predict but my expectation is over time
people are much more capable, much more
creative, think much more expansively
and much more broadly than they do
today. Um, and we start
we certainly see examples of this uh
where people are just like I never would
have been able to keep this in my head.
I never would have been able to have
this idea. And then we hear other
examples where people say, you know,
I've outsourced my thinking and I just
do what this thing tells me. And
obviously we're much more excited about
the former than the latter.
Um, can you help us understand why you
build emotionally intelligent models and
then criticize people for using using it
for accessibility reasons when it comes
to processing emotions and mental
health? Uh, again, we think that's a
good thing. We want that. Uh, we're
happy about that. There the same model
that can do that can also be used to
encourage delusions and mentally fragile
users. And what we want is people who
are using these models intentionally.
The model is not deceiving the user
about what it is and what it isn't. the
model's being helpful, the model's
helping a user accomplish their goals.
We want more of that and less of
anything that would feel like the model
tricking a user, for lack of a more
scientific word. Um, I totally get we
totally get the the frustration here
like to, you know, whenever you're
trying to stop something that is causing
harm, you stop some perfectly good use
as well. Um, but please understand the
place we're coming from here is trying
to
provide a service to adults that are
aware of it and that are getting real
value from it and not cause unintended
harm to people who who don't want that
along along the way. Um, all right.
Given that we have just a couple of
minutes left, let's see if there's any
questions in very other directions that
we should try to get to.
Okay. Uh, when do you think massive jobs
loss will happen due to AI from Razi?
So I think I think
already we are at a point where I think
a lot of the um
um the gap that stops that stops present
models from like being able to perform a
lot of um intellectual jobs. uh it's
more about um integrations and
interfaces
uh than maybe raw intellectual
capability. Um and so I think we
definitely have to
uh
think about that um uh think about
automation of of of a lot of jobs is
something that um will be happening over
the next years. And I think this is a
big uh
thing for for for for us to to
collectively think about like what are
uh
what are the jobs that will will replace
those and uh and what are the kind of
new pursuits that we'll that that we'll
all engage on.
>> This is a question from me not from uh
not from the live stream. What what do
you think meaning will look like? Uh
what do you think the jobs of the future
will look like? How do you think when AI
automates a lot of the current things
like how do you think we'll derive our
fulfillment and spend our time?
>> Um
I I expect um
yeah well I I think this is a this is a
quite um
philosophical question. I think I I
think can go in many directions but some
things I expect I think the high level
goal setting right like picking what
pursuits uh um
we're we're chasing that is something
will remain human and and
uh and I think that that is something
that a lot of people will derive meaning
from. Um, I think also just the ability
to
understand so much more about the world,
the wide the incredible variety of of of
new knowledge and new also
entertainment, but al also
uh
um just intelligence that that that that
that
will be in the world. I think I think
will uh will provide quite a lot of
meaning and fulfillment for people.
>> Okay, rapid fire, two minutes. Uh Shindi
says when GBT6
um
I think I think I think uh in some ways
maybe that's more of a question for you
uh in that like um I think I think with
with with GPD with GPD 5 uh we have uh
with previous models right like uh like
GPD4 GPD3 like we've kind of like kept
very tight uh connection of like how h
how we're training new models like what
are the products that we ship
And as I was just saying like I think
right now there's a lot to do on the
kind of integration side and on on the
uh so for example with GPT5 is the kind
of the the first time we really bring
reasoning models as kind of our our
flagship our main flagship model. Um and
so uh we're we're not coupling like
these releases and these products as
tightly to our research program anymore.
>> Yeah. Um I don't know either exactly
when we'll call it that but I think a
clear message from us is say 6 months
from now probably sooner we expect to
have huge steps forward in model
capability. Felix asks is an IPO still
planned and how would the structure then
look like? Are there rules in place for
increasing capital? Um we don't have
like specific plans or this is exactly
when it's going to happen but I think it
it's fair to say it is the most likely
path for us given the capital needs that
we'll have in sort of the size of the
company. Um, but you know that's not
like a top of- mind thing for us right
now. Alec asks, "You mentioned being
comfortable with the 1.4 trillion of
investment. What level of revenues would
you need to support this over time? What
will be the largest revenue driver? It
can't just be a per user subscription."
Um, you know, eventually we need to get
to hundreds of billions of a year in
revenue and and we're on a pretty steep
curve towards that. I expect enterprise
to be a huge revenue driver for us, but
I think consumer really will be too. And
it won't just be the subscript
subscription, but we'll have new
products, devices, tons of other things
there as well. Uh, and this says nothing
about like what it would really mean to
have AI discovery and science and all of
the revenue uh um possibilities that
that would unlock. And as we see more of
that, we will uh we will increase spend
on infrastructure. Okay. Um, we are out
of time. Thank you all very much for
joining us and the questions. and we
will try to learn from this format and
iterate on it and keep doing uh these
sorts of Q&A. Thank you very much.
Loading video analysis...