OpenAI is in "CODE RED" (Did Gemini win that hard??)
By Theo - t3․gg
Summary
## Key takeaways - **OpenAI Declares Internal Code Red**: Sam Altman has declared a code red internally, signaling OpenAI is scared by Gemini 3 Pro launch and Deepseek open-weight models, a shift from historically ignoring competition like Anthropic, Amazon, and Google. [00:00], [00:42] - **ChatGPT Traffic Drops 6-7%**: ChatGPT traffic has declined by up to 6% recently per Similar Web, coinciding with Gemini 3 launch improving Google's AI search results, which is terrifying as 6% of users exceeds many competitors' entire bases. [11:38], [12:13] - **OpenAI Lags in Model Leadership**: For the first time, OpenAI doesn't have the best model for anything; Anthropic leads in code with Opus 4.5, Google crushes benchmarks with faster Gemini 3 Pro at 75 TPS vs. OpenAI's slow 32-50 TPS. [09:41], [11:12] - **Google's AI Search Dominates Apps**: Google's AI overview on search is the most popular AI app ever, posing the biggest risk to OpenAI as it integrates seamlessly into everyday Google searches and can be monetized via ads for free access. [07:42], [09:03] - **Ads Unviable for OpenAI Revenue**: OpenAI can't compete in ads like Google and Meta due to lacking data, infra, and targeting; example shows 2.3M views yield only $6,800 ad revenue, far less than AI inference costs at ~2 cents per request. [21:42], [23:57] - **Delaying All But ChatGPT Focus**: Per the memo, OpenAI is delaying advertising, AI agents for health/shopping, and Pulse personal assistant to enhance ChatGPT's speed, reliability, and personalization, recognizing it as their core product. [03:10], [05:22]
Topics Covered
- Google's Search AI Crushes ChatGPT
- OpenAI Loses Model Lead
- Google Chips Enable Cost Wins
- Ads Fail AI Chats
- ChatGPT Must Dominate
Full Transcript
It seems like OpenAI is finally feeling the heat. Sam Alman has just declared a
the heat. Sam Alman has just declared a code red internally. This is kind of crazy. Historically, OpenAI hasn't
crazy. Historically, OpenAI hasn't really thought too much about the competition. Almost all of my
competition. Almost all of my conversations with them just really didn't seem to think much about what other companies were doing. They didn't
care much about Anthropic and they really didn't seem to care about companies like Amazon and Google. That
has changed. combination of the Gemini 3 Pro launch as well as recent developments in openw weight models, especially from Deepseek seems to have OpenAI quite a bit scared. This is
reminiscent of the explosion that happened internally at Meta when Deepseek originally dropped that resulted in like 3/4 of the Llama team getting fired or leaving.
This is a big deal. This is not the type of thing Sam normally does. This is not the type of thing OpenAI normally spends their time on, but they are clearly scared more so than ever. Now, there's a
lot to read into in this memo. I don't
know how much of it has actually leaked, but we'll figure that out going forward.
We need to discuss why OpenAI is doing this, why they're feeling so much pressure, what their long-term plans are, and what they're going to do before the end of the year, because I have a feeling they're going to drop something
very soon. As always, OpenAI is not
very soon. As always, OpenAI is not paying me for any of my coverage or content. There is a company paying me
content. There is a company paying me though, so let's hear from them really quick. There's a lot of data that I want
quick. There's a lot of data that I want to be able to store for my users, but might not want to have in my database as clear text. Like if users are bringing
clear text. Like if users are bringing their own API tokens for a service like T3 chat, I don't want to keep those API keys as plain text in my DB. I want to store them in a way that only that user
can access, but ideally is still owned and managed by me. Which is why today's sponsor has been awesome. Turns out work OS has a thing just built in for this, and now we're using it all over T3 chat.
Their vault product is super cool. You
store the data wherever you choose to.
And what you're storing in work OS is just a key to unlock it. They give you the primitives you need to access the key and run it against your data. So
that you don't have to have work OS as your database. They're just there as
your database. They're just there as your O layer. And that lets users access data and read the actual contents of that data without having to build all this complex infrastructure yourself. If
that's all Work OS offered, it would have been worth us making the move. But
there is so much more. From the admin portal that makes it way easier for companies to onboard to your platform to their MCP off stuff that they're industry leading in. Turns out making O for MCP is annoying because the standard
sucks. They fixed it. Okit, which is all
sucks. They fixed it. Okit, which is all the UI layer and components that you'll need to onboard things and get it added to your codebase. User management,
enterprise SSO, Radar has been awesome, too, having an actual user level thing checking the validity of a user. There's
a reason every company is making the move to work OS. small companies like OpenAI and big ones like T3 chat.
Actually though, Verscell, Cursor, Indeed Plaid Loom Socket FAL Source Graph, Cardo, Web Flow, and so many more. This is a wild list of companies.
more. This is a wild list of companies.
It's crazy that we all picked this solution, but once you try it, you'll probably understand why. Stop losing
enterprise customers. Move to work OS today at swive.link/workos.
So, let's start by reading what news coverage exists on this. Then we'll dive into all of the fun things OpenAI is cooking in order to try and compete with their increasing competition. OpenAI CEO
Sam Alman has set off a code red alert to employees to improve its flagship product Chat GPT and delay other product developments according to Wall Street Journal. Interesting. I didn't know this
Journal. Interesting. I didn't know this was specific to Chat GPT. I wonder if the best AI chat ever made might be getting them a little bit scared. By the
way, if you want every single model in the best possible UI for eight bucks a month, check out T3 Chat. When I say every model, I mean it. Including image
gen models, which I've been using for memes. Obviously, I have my thoughts on
memes. Obviously, I have my thoughts on image gen, but it's good for this type of thing. And if you want your first
of thing. And if you want your first month for just a dollar at checkout, apply the code code red. Yes, code red at checkout for your first month for $1.
And if you're Sam Alman watching this video, by the way, uh we might not be a dollar to acquire, but if you do want us to fix chat GPT, you can get my phone number pretty easily. Let's chat.
Seriously though, if Anthropic is buying Bun to fix their developer stuff, I think it makes a lot of sense for OpenAI to buy us to fix their user experience stuff. We're kind of the guys who do
stuff. We're kind of the guys who do that. Anyways, Wall Street Journal
that. Anyways, Wall Street Journal reported that Altman sent an internal memo to staff on Monday saying more work was needed to enhance the artificial intelligence chatbot's speed, reliability, and personalization
features. Do you know who's really good
features. Do you know who's really good at speed and reliability for chatbots?
Sam, let's chat. This week marks 3 years since OpenAI first released chat GPT, sparking global fascination in a commercial boom in generative AI technology and giving the San Francisco based startup an early lead. The company
faces increased competition with rivals including Google which last month unleashed Gemini 3, the latest version of its own AI assistant. Opening I
didn't immediately respond to a request for comment on Tuesday. Tech news outlet the information also report on the memo.
They have 800 million weekly users but they also are valued at 500 billion.
They don't make a profit and they've committed to more than a trillion dollars in financial obligations in cloud computing over the next 10 years.
The risk OpenAI won't make enough money to fulfill the expectations of backers like Oracle Nvidia has amplified investor concerns about the AI bubble and OpenAI VP and its head of Chibbt post on social media Monday that online search is one of the product's biggest
areas of opportunity as the company focuses on making Chashibbt more capable and even more intuitive and personal.
They currently make revenue from premium subscriptions, but most users are on the free version. They also have their own
free version. They also have their own web browser. Apparently, in the memo,
web browser. Apparently, in the memo, they said they're delaying work on advertising, AI agents for health and shopping, and a personal assistant called Pulse. Pulse, if you don't know,
called Pulse. Pulse, if you don't know, is a thing that popped up in chat GPT that was like, "Here are some things that we think might be interesting to you today." And I have some screenshots,
you today." And I have some screenshots, I don't feel like finding them, of really cringe things it would say to me based on stuff I had done in the past with it. It was It was weird. It was
with it. It was It was weird. It was
like bringing up text stacks that haven't been relevant for a year cuz I asked about them forever ago. Very
interesting. The ads thing is particularly fascinating because some people were reporting not too long ago, like literally 3 days ago, that the
Android app for ChatgPT had ads features added to its like reference and like XML dump. So, they were planning on
dump. So, they were planning on introducing some ad stuff, which makes sense when you realize that Google makes almost all of their money from ads still. The only two real players in the
still. The only two real players in the advertising space are Google and Meta because they have all of the data, all of the infra, all of the targeting, all of the tooling, and all of the things needed to do advertising right. They're
the only companies that can make real amounts of money through ads. Everybody
else is stuck with whatever margins Google's willing to give up when you integrate their ad stuff. So, if OpenAI wanted to do this, they'd have to build their own solution for almost all of that because there's no world in which Google's giving them a good deal on
AdSense. Just not going to happen. So if
AdSense. Just not going to happen. So if
they want to compete with Google, the king of ads, they need to win in like every single vertical. Artificial
analysis did this report earlier this year. Artificial analysis publishes
year. Artificial analysis publishes reports on the state of AI relatively regularly. And there's an interesting
regularly. And there's an interesting chart I use a lot in here. Go to
artificial analysis if you want to see the whole thing. I'll link the site in the description. You can find it
the description. You can find it relatively easily there. This chart's a big part of why OpenAI is currently scared. There are four major verticals
scared. There are four major verticals that all of these companies are competing on. apps, models, cloud
competing on. apps, models, cloud inference, and the actual hardware the inference is done on. So to give a basic example for someone running their own inference, you're probably going to be
using NVIDIA GPUs. You're probably going to be hosting on something like AWS or Azure. Maybe you're using anthropic open
Azure. Maybe you're using anthropic open directly. The foundation models, these
directly. The foundation models, these are the models you're actually using. So
for basic example, accelerator hardware could be NVIDIA. Cloud inference could be Amazon. The model could be sonnet.
be Amazon. The model could be sonnet.
And the app could be T3 chat. Just as a simple example of what each of these is, Google is very uniquely positioned here because they have their own apps. The
biggest AI app ever made is actually Google. Literally Google. Google.com.
Google. Literally Google. Google.com.
The most popular AI app ever made. You
Google search something. Oh, look at that. I didn't even This is the thing.
that. I didn't even This is the thing.
New try AI mode. Search whatever's on your mind and get AI powered results. I
see why Enthropic is scared. That's not
even what I was trying to do here. But
you know what? Let's turn it on. Is this
what happens now when I Google? It
isn't. Interesting. I have to click AI mode to go to that.
Interesting. It doesn't know about T3 chat. That means it's using the new
chat. That means it's using the new Gemini because it it knowledge is somewhat dated. Why is it linking the
somewhat dated. Why is it linking the Git Nation podcast episode of all things?
That perform well. Oh, 16k plays. Yeah,
I think it's one of the most popular videos on their channel.
The point I'm making is that that little blurb at the top of Google, the search generation AI overview thing, this is the most popular AI app in the world.
This is, in my opinion, the single biggest risk to OpenAI. Like terrifying
that your core product is now just the thing that happens when you Google search. That would scare the [ __ ] out of
search. That would scare the [ __ ] out of me. Especially because Google can make
me. Especially because Google can make money off ads which allows them to provide this for free. That's really
scary. And that's the application layer.
Historically, OpenAI has been the winner cuz the average person when they think of AI, they're thinking of chat GPT. My
parents get confused when I tell them about what I'm building because they think of AI as chat GPT. So like, oh, you're building into chat GPT? Like is
OpenAI paying you? They're just
confused. Google is the only company I can think of that has a real chance of disrupting here. Although I did take an
disrupting here. Although I did take an Uber yesterday from somebody who was insisting that Meta's AI chat is the best and that he was going to make a really successful company using it, but then he went into some really crazy conspiracy theories. So, I don't think
conspiracy theories. So, I don't think he's a great reference as anything other than a normal kind of weird person. And
then there's the foundation models.
Historically, OpenAI has maintained the lead or very close to it. But for the first time ever, in my opinion, right now, OpenAI doesn't have the best model
for anything. That's weird. GBD 5.1 Pro
for anything. That's weird. GBD 5.1 Pro is still incredible and can solve things I haven't seen other models solve, but it's not available over API. I can't
really benchmark it. It's only usable through the chatbt site, so it's hard to to know the extent of how good it is beyond playing with it and getting fun answers, which but 5.1 Pro isn't really
measurable the way other models are.
When it comes to code, Anthropic is currently in the lead in my opinion by quite a bit with Opus 4.5. I've been
very impressed with that model. When it
comes to general benchmark performance and things like ArcGI and stuff, Google is crushing all of the benchmarks.
Google also has better search and a much faster model. This is one of the most
faster model. This is one of the most annoying things and I I can't imagine OpenAI not addressing this with future
releases. GBT 5.1 is not a fast model.
releases. GBT 5.1 is not a fast model.
It's up to 50 TPS and if you look at codeex, it's 32. I don't think Codex is even on the API yet. It's only available via codecs like the CLI. And I think they gave it access to a couple other
companies, but 32TPS is very, very slow, especially when you compare that to Gemini 3 Pro, which is at 75ish.
That's 2 to 3x difference, and it feels like a whole different world. Opus 4.5
is in the 60s pretty consistently.
And Sonnet 4.5 has actually dropped a little. It's in the It's hitting the 60s
little. It's in the It's hitting the 60s consistently. The point here is that the
consistently. The point here is that the experience I get actually using the new models that are being provided by OpenAI feels not great because it just it's slow. And this is a very good point from
slow. And this is a very good point from Michaela from Zed. 5.1 is for completions and responses. 5.1 CEX is only for responses. Codex max is only
CLI. And 5.1 Pro is only chat. Yeah,
CLI. And 5.1 Pro is only chat. Yeah,
apparently chat GPT traffic has declined by up to 6% recently. The theory is that this has
recently. The theory is that this has happened since the Gemini 3 launch, which would improve the like AI search results on Google, but that's pretty brutal. Hard to know with these numbers,
brutal. Hard to know with these numbers, like we don't actually have access to those numbers. This is from Similar Web,
those numbers. This is from Similar Web, which is a analytics firm that tries to track usage across things. They are
seeing a 7% decline over the last couple weeks. That is terrifying. 6% of chat
weeks. That is terrifying. 6% of chat GBT users is more users than we have by quite a bit back here. So, OpenAI now
has real computational applications.
People still associate AI with chat GBT and they even like call it chat. I
thought that was fake. I thought people were referring to chat like how I refer to my Twitch chat. No, normies refer to chat GBT as chat and they'll just say like, "Have you asked chat about that?"
Or, "So, I was talking to chat earlier."
Insane to me. Like, actually,
despite that, Google now has a real edge and is starting to take cuts out of OpenAI's traffic. the foundation model
OpenAI's traffic. the foundation model side which we were just discussing.
OpenAI doesn't have the best model for any one thing right now. They have
really good general models but they're quite a bit slower and not super pleasant to use. The cloud inference thing is also very interesting because historically OpenAI has been the frontier lab that is the most reliable
cloud. Anthropic is so unreliable that
cloud. Anthropic is so unreliable that you basically have to use a service like Open Router to diversify your traffic to other places. Thankfully, Anthropic has
other places. Thankfully, Anthropic has partnerships with both Google and AWS and recently even Azure, so you can use their models on other clouds with other GPUs and not have to worry about Anthropics provisioning. Anthropics
Anthropics provisioning. Anthropics struggled a lot with their own infrastructure for quite a bit now.
Their stuff's down hilariously often, but despite that, they've managed to get their revenue up. According to CNBC, they only had a thousand enterprise customers two years ago, and now they
have over 300,000. and they define those as customers that spend more than 100k a year on the infra on their stuff. Cool
to know that I'm one of their large accounts now with T3 chat. Jesus Christ.
Yeah, we spend a lot of money on anthropic. So on the infant side, OpenAI
anthropic. So on the infant side, OpenAI was one of the better options. The only
other place you can use OpenAI models is currently Azure because of the weird partnership between Microsoft and OpenAI that's slowly falling apart. Google, of
course, with Google Cloud, you're good.
I'm not going to say Google Cloud's great. The differences between AI Studio
great. The differences between AI Studio and Vertex are [ __ ] obnoxious. I'm so
tired of fighting those two platforms and I've been yelling at the AI Studio team for a while now. I actually am meeting with them in a few weeks and I'll be yelling at them even more. There
are so many things that are broken with it. It is actually hilarious. Uh, fun
it. It is actually hilarious. Uh, fun
fact, you can't over API force Nanobanana to only do a certain number of images. It is possible to force Nano
of images. It is possible to force Nano Banana Pro to generate multiple images over API even if you try to restrict it with your system prompt.
Yeah, I have feelings. So, OpenAI is okay here, but they're only as okay as their models are because you're not going to use their infra for other models. You might use GCP for other
models. You might use GCP for other models. You are going to use Microsoft
models. You are going to use Microsoft for other models because Microsoft's models are garbage. Amazon did just put out their first ever decent model, but you're not using Bedrock for that right now. You're using Bedrock for anthropic
now. You're using Bedrock for anthropic support almost certainly. So, OpenAI
needs their models to be the best for their cloud inference to be a real player and their API usage is still not a big portion of their revenue. For
OpenAI, 73% of their revenue is currently chatg subscriptions and only 27% is the API and multimodal services.
So, they need the subscriptions to be their main source of revenue. They can't
really compete otherwise. The only thing they could do is ads. And we've already discussed why that's going to be rough for them. And then there's the
for them. And then there's the accelerator hardware. This is one of
accelerator hardware. This is one of those weird edges that puts Google far ahead. Google is the only lab that's
ahead. Google is the only lab that's making good models and making good chips. Google doesn't need Nvidia to
chips. Google doesn't need Nvidia to win. Google is actually selling their
win. Google is actually selling their own chips now to Meta to help them win with Llama. Google doesn't have to rely
with Llama. Google doesn't have to rely on the Nvidia monopoly to succeed. The
only other companies coming close to Google's chip architecture stuff are these three at the back here. Gro,
Cerebras, and Samanova. These companies
are effectively building ASIC chips similar to like for the Bitcoin mining stuff and they're just doing this for inference. So that's why on Grock you
inference. So that's why on Grock you can get absurd speeds. If we look at Kimmy K2 for example, the non-thinking version, most hosts are getting between 40 and 80 TPS. Some are struggling a lot
with like 16 TPS, but if you look at Grock with a Q, the ones making their own chips, they're hitting 330 TPS comically faster because all they're doing is building these chips that are
really fast. Rather than selling the
really fast. Rather than selling the chips, they're selling you access to their servers so that you can get the way faster speeds. Grog with the Q is awesome. I really like these guys, but
awesome. I really like these guys, but those companies are building chips and architecture which makes them compete on a different level. None of them are trying to build a chat app or a search product. And none of them are trying to
product. And none of them are trying to train their own models. They're trying
to provide APIs that we can hit to get better performance using chips that they make that are way faster and use less energy. Open is not competing there at
energy. Open is not competing there at all. So, they're at the whim of Nvidia.
all. So, they're at the whim of Nvidia.
That's why they have to make those crazy trillion dollar commitments because they have to buy chips in order to fight because they're not producing their own.
A long last time ago, they were actually considering buying Cerebrris to benefit them like before they even shipped Chat GPT. But now it's too late. Going to be
GPT. But now it's too late. Going to be a rough play for them to make. And any
attempt to compete here is going to sever their relationship with Nvidia.
And they need that relationship to stay strong or they're [ __ ] So open is kind of caught in between a rock and a hard place here. Other models are getting better enough that the
competition is ramping. If other models are better and a company like us with T3 chat can provide a better interface, you're losing surface area that you can compete on. Bunch of places open can be
compete on. Bunch of places open can be competitive in best model, best chat interface advertising
revenue infrastructure APIs developer experience. So this is like coding with
experience. So this is like coding with things like codecs and whatnot. Mind
share. These are all the places that in my opinion OpenAI needs to be thinking about and competing in historically they've maintained best model.
Historically they've had a good enough chat interface that no one was thinking twice about it. It had its rough edges, but since they hired Nean and a few others, they've been able to make it better. Still nowhere near as stable as
better. Still nowhere near as stable as T3 chat, but better ad revenue. We
discussed it can't really compete in infer APIs are only as good as the model is. These two are very dependent for
is. These two are very dependent for them because you're only using their inferent APIs for OpenAI models.
Developer experience, they put a lot of work in at the start of the year, but now there's real risk happening here with all the developments occurring at Enthropic, especially with the acquisition of Bun and Mind Share. This
is the one place I think they're really winning. Best model, I think it depends
winning. Best model, I think it depends right now. Anthropic and Google are both
right now. Anthropic and Google are both crushing it right now. Best chat
interface, T3 chat. Seriously though,
uh, OpenAI is the big lead right now. As
in like of the major labs, OpenAI has by far the best chat app. Cloud.AI
miserable experience. No one actually uses it. Seriously, it's so bad. It's
uses it. Seriously, it's so bad. It's
every time I use it, I find new bugs.
And the Gemini, like I was going to try and use Gemini for a week. I survived
two hours before getting so frustrated.
It's so bad. I cannot fathom how bad it is. It's so bad that I don't even want
is. It's so bad that I don't even want to work there. Like, if Google was to give us an offer to come in and replace the Gemini app with T3 chat, it would have to be a lot of money for me to deal with all the [ __ ] Like, there's no
world in which there's a good work environment in that level of [ __ ] shipped. And then there's AI Studio,
shipped. And then there's AI Studio, which is somehow even [ __ ] worse. It
It's insane just how bad the interfaces are on Google's side. They cannot ship software. This is particularly funny
software. This is particularly funny because right now I'm live filming my videos on Twitch, which is what I usually do. I also am live on YouTube,
usually do. I also am live on YouTube, but I can't be live on YouTube with the chat on right now because they broke it.
Yes, if I have YouTube chat on my YouTube live stream, it oos my tab. Chat
interfaces are hard. Interfaces that
have text that is formatted with fancy animations and stuff. Updating thousands
of times per second is a difficult problem. I worked really hard on this
problem. I worked really hard on this with a really talented team at Twitch.
I'm still friends with a lot of them.
Building a good performant chat interface is not as easy as people seem to think. And if you hack one together
to think. And if you hack one together and you're impressed with how fast it is, add syntax highlighting, add math and latte, add animations, add complex rendering of any form, add image
rendering that resizes.
It falls apart fast, I promise you. And
even and YouTube can't even make a live chat that updates once every 20 seconds without failing. Google doesn't have the
without failing. Google doesn't have the engineering expertise to compete in the user experience side, which is why they're leaning so heavily into the Google experience instead. Gemini's chat
app will never be a popular solution because it will never be [ __ ] good.
Hopefully, I don't need to convince you guys that Google is incapable of shipping good software. I think it's pretty apparent, but that matters now.
So, the edge that OpenAI has over Google is the quality of the software that they're shipping. OpenAI still has the
they're shipping. OpenAI still has the lead here, but they're at a real risk now. As silly as it is, something like
now. As silly as it is, something like T3 chat is a huge risk for them because T3 chat's great if you want to try lots of models and have a slightly better chat experience, but a lot of people
just go back to chat GBT. I would
honestly guess that half of our subs are still maintaining a chat GBT subscription as well and switch between the two. If we get better enough and the
the two. If we get better enough and the other competing models get better enough, there's less and less reason to keep that chatbt subscription and more and more reason to try out the other
models on something like T3 Chat. And
god forbid if anthropic was to buy T3 chat and we were to go fix the claude website, all of a sudden they are really screwed. By the way, Enthropic, it seems
screwed. By the way, Enthropic, it seems like you're interested in companies that are in the JavaScript world. I know
we've had our issues in the past.
There's a lot of ways that we can bridge those issues. All it would take is open
those issues. All it would take is open source cloud code and a a light acquisition. I'm sure we can work out an
acquisition. I'm sure we can work out an 8-digit number. Anyways, advertising
8-digit number. Anyways, advertising revenue. This is, as I said before,
revenue. This is, as I said before, where they cannot win. Google and Meta are the only winners here. There have
been a lot of people suggesting that AI chats would do so much better if you just added ads to it. Do you have any idea how much this [ __ ] costs? Do you
have any idea how little money you make off of ads? Just to give you an idea, these are numbers I'm not supposed to share. I don't care. I get 2.3 million
share. I don't care. I get 2.3 million views or so a month. That's pretty good.
My audience is developers, which is really good. My ad revenue is $6,800
really good. My ad revenue is $6,800 for the month. 2.3 million views, $6,800.
Do you understand how [ __ ] that ratio is? I pay my editor more than that.
is? I pay my editor more than that.
We're doing a video every day. Ad rev is [ __ ] It's so much lower than people think. It is very hard to make a lot of
think. It is very hard to make a lot of money on ads, especially when the average request costs money. I don't
feel like going in our dashboard and figuring out how much the average request costs with GBT 5.1, but um just for an example with Sonnet, Sonnet 4.5
costs $3 per million input tokens and $15 per million out. Assuming the
average request has 2,000 input tokens and I don't know 800 out, that's about two cents per request. I know that
sounds small, but imagine you make a 100 requests in a couple sessions. That's
$2. And every time you make a new request in a thread, the amount of input tokens goes up and it gets more expensive as follow-up requests happen.
Long threads are very expensive unless you do crazy compaction [ __ ] that makes the quality of the responses worse. If
the average user doing a thing cost 2 cents, like if we were to just look at these numbers for views on YouTube, 2.3
million requests times that 0.02, that's $46,000. Assuming every single
that's $46,000. Assuming every single one of my viewers made one request that was around the average cost, it would cost me $46,000 and my ad revenue was
under $7,000.
Ads cannot sustain you here unless you have ways to reduce costs a ton and maximize return from those advertisements. The only two companies
advertisements. The only two companies that can maximize return on advertisements are Google and Meta. The
only one of these companies that has the ability to reduce costs is Google.
Google has their own chips, which means they can make things run faster and cheaper than their competition. Google
has the service of google.com which means if you and I both Google the same thing and they do an AI result for it, you can get the cache version if I got the generated one which saves them a ton
of money. It's possible that a huge
of money. It's possible that a huge percentage of those Google searched AI result things are cached results so that we don't have to cost them more money for each one. Reading from a cache is
comically cheaper than generating a new response. Google can make inference
response. Google can make inference cheaper. They can reduce the amount of
cheaper. They can reduce the amount of inference and they can make more money per ad than anyone else can. Advertising
revenue is only a viable path for Google right now. In my opinion, Meta is so
right now. In my opinion, Meta is so desperate to reduce their costs that they are inking a crazy contract with Google to be the only other company using Google's chips in their server
factories and warehouses. That's crazy.
Then we have the infrastructure and API side. OpenAI is the only major lab
side. OpenAI is the only major lab that's cloud is competent other than arguably Google. But Google Cloud, I've
arguably Google. But Google Cloud, I've been there.
I would argue for better or worse, Google and AWS are winning. Azure could
be a real competitor here, but they have to put a little bit more work in.
They're getting there, but Google and AWS are the ones that have the best surface for this right now because you can use other models on their clouds.
They're clouds that have other value to the developers. And as nice as OpenAI's
the developers. And as nice as OpenAI's infra is, if you're not using their models anyways, you can just move to Google or AWS. And even Azure, which was their one alt bet, now supports
anthropic models as well. And developer
experience, cloud one, openai was catching up, but cloud code has continued to accelerate. And on top of that, Opus 4.5 is groundbreaking, and the bun acquisition positions them
incredibly well, much more so than a CLI that breaks half the time that was written in Rust. Yeah, the the rewrite of codecs in Rust will be remembered as one of the dumbest things OpenAI has ever done. I told them ahead of time to
ever done. I told them ahead of time to not do it. I yelled at Fouad for doing it real early. They committed because they wanted one binary and look how that's going for them. Nobody wants to contribute to it anymore. And then we
have mind share. This is OpenAI's last frontier. Google can beat the chat
frontier. Google can beat the chat interface with the search interface.
Google can beat the subscription revenue with advertising revenue. Google can
beat the infrastructure with their own chips. Google probably can't beat
chips. Google probably can't beat developer experience. That's why they
developer experience. That's why they acquired Windsurf. They're desperately
acquired Windsurf. They're desperately trying to win there and that's not going to happen. So Enthropic will have that
to happen. So Enthropic will have that one edge there. Mind share is what Open AI has left. The fact that Chachibbt is the Xerox, the Kleenex, the the name brand of AI for most people is their
final frontier. That is what they have
final frontier. That is what they have left. They have to maintain that. Chat
left. They have to maintain that. Chat
GBT needs to be the best chat experience or OpenAI will lose. If they don't maintain that perception, they don't have the position to keep fighting and to keep winning. They don't have the income to justify the trillions of
dollars they've committed to spending.
They don't have the users that they need to make money and win. Which leads me once again to remind you, Sam, that buying us is a really cheap hedge to make it more likely you maintain this.
Hit me up.
I think this does a good enough job of putting my thoughts down on paper. It's
a weird market right now, and I do not envy the people who have to think about this and fight it every day, but at the same time, very thankful to be a layer above. I know everybody loves memeing on
above. I know everybody loves memeing on the AI rapper services, but it's never felt better to be a rapper because T3 chat is still making money. We're in a really good spot. If Google or Enthropic
or somehow Amazon end up winning the model wars, all of them are so bad at UI that we win by default. We're in a really good position right now. And it
would be really hard to convince me to give that up for almost anything. The
moment we have our mobile app ready to go, which by the way, coming in hot, competition's going to heat up. So, what
can they do now? What strats can OpenAI pull to position themselves better?
There are some crazy things they could do like buy Cursor. That would be nuts.
Cursor is a very, very valuable company.
And that acquisition would be weird for a bunch of reasons. In particular, the agreement that currently exists between Microsoft and OpenAI where all OpenAI IP is officially licensed to Microsoft
would mean that all the work that's gone into Cursor would be accessible to Microsoft and then they can all fold that into co-pilot in VS Code. makes
that an unlikely path unless they can sever that relationship. Another path,
and this is what I'm almost certain they're taking, drop a model ASAP. I
honestly kind of feel like they were holding the Codex Max model in case things ramped up too hard with Gemini 3 Pro. They were clearly sitting on that
Pro. They were clearly sitting on that for a bit and then dropped it all of a sudden because they wanted to stay in the lead on the leaderboards. It's also
somewhat clear that GBT 5 and 5.1 aren't huge models. They don't have the big
huge models. They don't have the big model smell. They are slow enough that I
model smell. They are slow enough that I would be surprised if they were small models, but they're not as big as many of the old OpenAI models were. That's
why they made the price as cheap as they did. They can almost certainly put out a
did. They can almost certainly put out a bigger model. GBD 4.5 is the biggest
bigger model. GBD 4.5 is the biggest they ever released. Five is smaller for sure, but if they have a big model they're sitting on that can crush benchmarks, I wouldn't be surprised if they drop that soon. I don't know how it
will be branded or named, but they've got to be cooking there. I've heard a lot of rumors all over Twitter.
Something's going on. What else can they do? they can win on integrations. They
do? they can win on integrations. They
were doing a pretty good job with Apple for a bit where Siri could search with chat GPT, but now Google has a partnership with Apple where Apple is buying a model from Google to host on
their own infrastructure. Very unlikely
that OpenAI can do that anymore now that Google realized the risk there and is squeezing a win out of it. OpenAI should
try and I'm sure they are in order to maintain that relationship with Apple. I
wish them luck because Google knows better than anybody how valuable it is to have Apple use you as a default. One
other thing they could do and I suspect they'll do this at some point. I don't
think it's as big of a thing unless they're already winning hardware. Not
making their own chips or CPUs or GPUs or TPUs or whatever, but actually selling devices that are useful AI stuff that an average chatbt user would consider actually using. If they can
crush the equivalent of like an Amazon Echo or whatever the [ __ ] the like Alexa stuff is called, there's a real potential angle there. I don't know how big it is, but I know those things are
really popular, so I would be surprised if the Joanie IV stuff isn't going in that direction. But the biggest thing by
that direction. But the biggest thing by far is to make sure that chat GPT wins.
OpenAI doesn't really matter beyond Chat GPT in the general marketplace right now.
That's not where they're investing their money. Their money is going into
money. Their money is going into research infrastructure GPUs partnerships, advertising, and all these other things. I would be surprised if
other things. I would be surprised if more than 5% of OpenAI's spend was directly related to chat GPT. A new
model comes out and chat GPT gets it and is betterish, but I've honestly been surprised recently. It feels like when
surprised recently. It feels like when new models drop, T3 chat supports them better than chat GPT does. Yeah, they
need to quadruple or more their investment in chat GPT and go really hard to make sure it wins. There are
certain ways they can do that, many of which I've hinted throughout here. But
the big thing is the investment. They
need to think about chat GPT less as a random surface that is being hit by users and more as a core part of their business. They released it as a demo of
business. They released it as a demo of their models originally and has gone way further than they expected. But now it's time to quadruple down and make sure ChachiBt wins and it seems like judging
from the reporting we were reading earlier. This is exactly how Sam feels
earlier. This is exactly how Sam feels as well. As the memo said, they're
as well. As the memo said, they're delaying work on pretty much everything else in favor of investing more on chatbt. That to me feels like the right
chatbt. That to me feels like the right call. As fun as these other experiments
call. As fun as these other experiments are, I haven't seen anyone using one of those chatbt apps yet. I know one person who kind of uses Atlas as one of their like five browsers. Their agent stuff
just isn't that great and I'd rather roll my own. The advertising they can't really win in. Held and shopping is cool demos, but I don't think they're going to do particularly great in those
spaces. They got to just make the best
spaces. They got to just make the best chat experience. I think this is most of
chat experience. I think this is most of my thoughts covered. Curious how y'all feel though. Am I way overreacting to
feel though. Am I way overreacting to this or is there actual danger for open AI? Curious how y'all feel. And until
AI? Curious how y'all feel. And until
next time, peace nerds.
Loading video analysis...