TLDW logo

What’s Next for AI? OpenAI’s Łukasz Kaiser (Transformer Co-Author)

By The MAD Podcast with Matt Turck

Summary

## Key takeaways - **AI Progress Smooth Exponential**: Fundamentally, if you look at AI progress, it's been a very smooth exponential increase in capabilities. This is the overarching trend like Moore's law, happening through new developments in addition to increased compute and better engineering. [02:46], [03:05] - **Reasoning New Paradigm Shift**: It's not like pre-training fizzled out. It's just we found out a new paradigm that at the same price gives us much more amazing development, and this paradigm is still very new. [05:09], [05:17] - **Chain-of-Thought RL Training**: A reasoning model generates tokens as chain-of-thought before answering, trained with reinforcement learning on verifiable domains like math and code to push towards better answers. This leads models to correct their own mistakes and verify what they say. [12:07], [13:35] - **Frontier Models Fail 5-Year-Old Puzzle**: None of the frontier models is able to solve a simple dot-counting exercise from my five-year-old daughter's math book that you would solve in 10 seconds, due to weak multimodal reasoning and context learning. [48:10], [48:21] - **Pre-Training Alive via Distillation**: Pre-training still scales smoothly but economics demand smaller models; distillation transfers knowledge from huge teacher models to cheaper ones, stacking well with RL for better capabilities. [35:52], [36:42] - **Translation Trusts Human Loop**: Even though models translate well, the translation industry has grown because for high-stakes outputs like newspaper listings, you still want a human who speaks the language to check before publishing due to trust. [01:01:14], [01:02:10]

Topics Covered

  • AI Capabilities Follow Smooth Exponential Curve
  • Reasoning Paradigm Unlocks Gains at Same Cost
  • Reasoning Models Think Before Answering via RL
  • Frontier Models Fail Simple Visual Reasoning
  • General RL Key to True Generalization

Full Transcript

There is this thing that's happening in AI and in AI every week now a lot is happening. Fundamentally if you look at

happening. Fundamentally if you look at AI progress it's been a very smooth exponential increase in capabilities.

This is the overarching trend. It's not

like pre-training fizzled out. It's just

we found out a new paradigm that at the same price gives us much more amazing development. And this paradigm is still

development. And this paradigm is still very new. I think one of the biggest

very new. I think one of the biggest things that I would say people kind of know on the inside and others don't is that already right now it's not about the progress. There are so many things

the progress. There are so many things Chad or Gemini any LLM can do for you that people just don't realize. You can

take a photo of something broken ask how to repair it. It may tell you you can give it a college level homework and it will do it for you.

>> Hi, I'm Matt. Welcome to the Matt podcast. My guest today is Lucas Kaiser,

podcast. My guest today is Lucas Kaiser, one of the key architects of modern AI who has quite literally shaped the history of the field. Luc was one of the co-authors of the attention is all you need paper, meaning he's one of the

inventors of the transformer architecture that powers almost all the AI that we use today. He's now a leading research scientist at OpenAI, helping drive the second major paradigm shift

towards reasoning models like the ones behind GPD 5.1. This episode is a deep exploration of the AI frontier. Why the

AI slowdown narrative is wrong, the logic puzzles that still stump the world's smartest models, how scaling is being redefined, and what all of that tells us about where AI is heading next.

Please enjoy this fantastic conversation with Luc.

>> Lush, welcome.

>> Thank you very much. there was a uh narrative at least in some circles maybe outside of San Francisco throughout the year that um AI progress uh was slowing

down that we had maxed out pre-training that scaling laws were hitting a wall yet we are recording this at the end of a huge week or couple of weeks with

release of GBD 5.1 GPD 5.1 Codex max GVD51 Pro as well as Gemini Banana Pro um for Gro 4.1 Almo 3. So this feels

like a a major violation of of that narrative. What is it that people in

narrative. What is it that people in Frontier AI labs know about AI progress that at least parts of the rest of the world seem to not understand?

>> I think there is there is a lot to unpack there. So so so I I want to go a

unpack there. So so so I I want to go a little slower. There is this thing

little slower. There is this thing that's happening in AI and in AI every week now a lot is happening. you know,

new model, coding, doing slides, self-driving cars, images, videos, you know, there there's it's it's a it's a nice field that that doesn't make you be

bored for a long time. Um, but through all of this, it's sometimes hard to see the fundamental things that are happening. And fundamentally, if you

happening. And fundamentally, if you look at AI progress, it's been a very smooth exponential increase in capabilities. This this is the

capabilities. This this is the overarching trend and there has never been much to make me at least and I think my

my colleagues in the labs believe that this trend is not happening. It's a

little bit like Mors law, right? We we

Morse law happened through decades and decades and arguably you would say it's still very much going on if not speeding up with the GPUs. But of course it did not happen as like one technology was

bringing you there for 40 years. there

were was one technology and then another and another and another and another and and and this went on for decades, right?

So, so from the outside you see a smooth trend but from the inside of course you know progress is made through new developments in addition to the increase

of computer power and better engineering and so so all of these things come together and in terms of language models I think there was a big pivotal point I

mean one point was of course the transformers when it started but the other point was reasoning models and that happened I think one preview was a

bit a year and a month ago or something like that. So we started working on it

like that. So we started working on it maybe 3 years ago but but you you you know it's very recent if if you think of it as a as a paradigm that that's a very

recent thing. So, so it's always like

recent thing. So, so it's always like these S-curves, right? It it starts then it gives you amazing growth and then it flatlines a little bit though. Yeah, we

we'll get to the pre-training, right? I

feel pre-training in some sense it's on the upper part of the S, but it's not like scaling for pre-training don't work. They they totally work like if you

work. They they totally work like if you what scaling clause says is that your loss will log linearly decrease with with your compute. We totally see that and clearly Google sees that and all

other labs. The problem is, you know,

other labs. The problem is, you know, how much money do you need to put into that versus the gains you get and it's just a lot of money and people are

putting it. But with the new paradigm of

putting it. But with the new paradigm of reasoning, you can get much more gains for the same amount of money because it's on this like lower and like they're just discoveries to be made and these

discoveries unlock insane capabilities.

So, so it's not like pre-training fizzled out. It's just we found out a

fizzled out. It's just we found out a new paradigm that at the same price gives us much more amazing development.

And this paradigm is still very new. It

happens so fast. I think you if you blink you may miss it cuz it was basically you had chat 3.5 right GPD 3.5 in chat and it would give you answers

and it used no tools, no reasoning. It

would answer you something. And now you have chat and you know if you were not into it you may have blinked and it also gives you answers and you may say okay it's more or less the same except the

chat now will you know go look on some websites reason about it and give you the right answer instead of something it memorized in its weights. I very much

used to like this example of you know when what time does the SF zoo open tomorrow like the old chat would tell you right totally hallucinate from its

memory an hour that it read zoo opens on probably the zoo's website from 5 years ago and it didn't know what's today or tomorrow so it would just assume it's a weekday chat now knows what's today

because it's in the system prompt it goes to the zoo website reads it extracts the information if it's ambiguous probably checks three other websites just to confirm and then gives you the answer. But if you blink, you

may think it's the same, but no, it's it's dramatically better. And you know, as a consequence, since it can read all the websites in the world, it can give

you answers in in stuff that it wouldn't be able to even touch before. So, so

there is tremendous progress, right? And

it happened so fast it it may even be missed. I I think I think one of the

missed. I I think I think one of the biggest things that I would say people kind of know on the inside and others don't is that already right now it's not

about the progress like there are so many things Chad or Gemini any any LLM can do for you that people just don't realize like you can take a photo of

something broken ask how to repair it may tell you you can give it a college level homework and it will do it for you probably so so That that's absolutely amazing.

>> So there is an education gap to some extent.

>> Well, it just happened like I mean if you think you you said codeex, right?

You know programmers are conservative a little bit. I I I still use Emacs from

little bit. I I I still use Emacs from time to time. All the coding tools like okay it will complete one line for me but but people are very like this is my editor. I write code here now people are

editor. I write code here now people are like no this is codeex. I ask it to do stuff. I will fix it later. Right? But I

stuff. I will fix it later. Right? But I

think it's the recent few months when the transition happened from you know people using it sometimes but rarely to now basically this being how a lot of people work in coding that that's quite

big. I'm not sure everyone's aware of it

big. I'm not sure everyone's aware of it but but it's also like if you don't do programming why would you be aware of it? I I do believe though that this will

it? I I do believe though that this will come to more and more domains >> to the point of all of this being very new and somewhat sudden. Uh something

that you or I hear from time to time when talking to people um is that part of the reason why people are so optimistic is that there is a lot of

lowhanging fruit very obvious things to improve uh for those models in the next few months. First of all, do you agree?

few months. First of all, do you agree?

And second, can you give us some examples like obvious thing that you need to fix next and that the industry will fix?

>> Yes, there is a ton of extremely obvious things to fix. A larger part of this ton is is just hard to talk about on a podcast because it is in the engineering

part. You know, every lab has their own

part. You know, every lab has their own infra uh and their own bugs in the code.

Machine learning is beautifully forgiving in some sense in contrast to old software engineering which would just yell at you when you made a mistake. You know, our Python code

mistake. You know, our Python code generally probably run except much slower and give you worse results if you run it wrong. So you realize, oh no, it was wrong. And you improve it and the

was wrong. And you improve it and the results get better. These are huge distributed computing systems. They're very complex to run. So So there is a huge amount to improve and fix and

understand in in the process about just how to train your model and how to do RL because RL is more finicky than pre-training. It's harder to do really,

pre-training. It's harder to do really, right? So every day this is our

right? So every day this is our day-to-day work. On top of that, there

day-to-day work. On top of that, there is data. You know, we used to train on

is data. You know, we used to train on just like common crawl basically. It's a

it's a big repository of the internet that people just scraped without regard of what and and some things came in, some didn't. It it was a mess. So now,

some didn't. It it was a mess. So now,

of course, every larger company has a team that that tries to filter this and and improve the quality. Um it's a but it's a lot of work to to really extract

better data. Now synthetic data is

better data. Now synthetic data is becoming a thing but when you generate synthetic data it really matters how you do it with what model what you the whole how the in the engineering aspects of

everything it's such a new domain that the the the you know it was done somehow it works it's beautiful but there is just so much to do better that that people I don't

think have any doubts that there is a lot there and on top of that there are the big things like multimodal I mean language models they they are now as as

I as I'm sure you know and and most people realize actually vision language models because and they can also do audio. So they're multimodal models to

audio. So they're multimodal models to some extent but the multimodal part still lags behind the text part to a large extent. So that's one big area

large extent. So that's one big area where obviously you need to do better and it's not a huge secret how you can do better. You know there there are some

do better. You know there there are some methods that maybe will make it even amazingly better but there are some very simple methods how you can do just better but you know this maybe requires retraining your whole base model from

scratch and that takes few months and it's a huge investment so we need to organize it so so so there is a lot

of just work that that will undoubtedly make things better. I think the big question that

better. I think the big question that that people have in their mind is how much better will it make them?

>> So I'd love to do uh a little bit of a deep dive slash educational part on um the whole reasoning model aspect uh because uh as as you just mentioned

since it's so new uh some people truly understand how those work many people don't. At a very simplistic level, what

don't. At a very simplistic level, what is a reasoning model and how is that different from um your sort of base LLM?

>> So, a reasoning model is like your base LLM, but before giving you the answer, it it thinks what people call in the chain of thought, meaning it generates

some tokens, some text that's meant not for you to read, but for the model to give you the better answer. And while it does this these days, it is also allowed

to use tools. So it can for example in its thinking pro so-called thinking process go and browse the web and to give you a better answer. So so that's

the superficial part of the thinking models. Now the deep part is that you

models. Now the deep part is that you start treating this thinking process as part of the model basically. So it's not something the model generates and and

it's an output for you. It's something

you want to train right. you want to tell the model you you should think well you should think so that the answer after this is good in in whatever way and this leads you to a very different

way of training the model because models were in usually trained with just gradient descent the way deep neural networks are trained meaning you say

predict the next word and you do a gradient you differentiate your function from the model they're not fully differentiable but you approximate it

and you train your weights to to do that and that it was quite amazing that doing just that you could make a chat but with a reasoning model you you can do that because there is this reasoning part you

can differentiate through that so we train this with reinforcement learning and reinforcement learning basically you okay there's just this reward and you need to do a bit of tries and reinforce

meaning push the model towards doing more of the things that lead to better answers and this kind of training is a bit more like it has more restrictions than the training we used before. The

training we used before, you took all of the internet, put it in, even if you didn't filter it very well, it would mostly work. Reinforcement learning, you

mostly work. Reinforcement learning, you need to be careful like you need to tune a lot of things, but you also need to prepare your data very carefully. So

currently and current for at least the most basic ways we use it currently, it needs to be fairly verifiable. So there

isn't is your answer correct or not? You

prepare data for that. You can do that in mathematics, coding very well. Uh you

can do this in science to some extent right you can have test questions they're correct or not but you know if it comes to like writing poems is this poem good or not it's for now the

reasoning models are really shine in in domains like science and they've brought some improvements to to non-science domains but but it's not quite as huge

maybe yet as it as it could be I mean at least compared to mathematics and coding then there is the multimodal question.

How do you do reasoning in multimodel? I

think this is starting like uh I saw some Gemini creating images in the reasoning part that quite exciting but it's very very yeah >> retraining and reinforcement learning

part is is um particularly interesting from an educational point of view again because um it seems that people have have come to the conclusion that there is the pre-training world and then the

post-training world and the post-training world is mostly reinforcement learning but this idea that there is reinforcement learning in the pre-training uh I don't think is as understood by everyone

>> at the beginning of chat let's say there was pre-training uh people did not do RL right then but then you couldn't really chat with it so chat was RLHF applied to

a pre-trained model but the RLHF was was different kind of RL of sorts right it was like very small and it was human preference that was telling you what is

better that's what the HF is human feedback right you You showed people like pairs of stuff, you learned a model that says, "Well, people seem to prefer this as an answer." You trained with

that, it would very quickly what we today say hack this model. Like if you trained the RLH of too long, it would start giving things that satisfy this model that seems supposed to model human

preferences. So it was a bit of a

preferences. So it was a bit of a brittle technique, but it was a bit of RL that was extremely crucial to to making the models chat. These days, I think most people move towards this big

RL. It's still not as big as

RL. It's still not as big as pre-training in the scope, but but it it says you have a model that says whether this is correct or not, or if it's a

preference, it it it's a very strong model that analyzes things and says you should prefer that and you have data that's restricted to a domain where you can do this well. And then you can also

put some human preferences on top, but but you make sure that that you can run a little bit longer without making this whole grading fall apart. But again,

this is this is around today. I I do believe the role of tomorrow will will be broader. It will work on general data

be broader. It will work on general data and maybe then it will expand to like domains that that go beyond where where it shines today. Now will it shine there? is a is is a different question.

there? is a is is a different question.

Do you really need to think very much doing some of the things? Maybe not. But

maybe yes, we maybe we think maybe we do more thinking and reasoning than we than we kind of consciously call thinking.

>> What would it take for RL to generalize?

Is that better evaluations? like you you guys released GDP val a few weeks or months ago uh to sort of measure performance against sort of broad

economic sectors. Is that is that part

economic sectors. Is that is that part of what the system needs?

>> I think this is a small part of it. I I

I think that's one part. But but if you think of economic tasks, you know, making slides is important there, following instructions, doing

calculations. It's not math, but it's

calculations. It's not math, but it's still very verifiable, right? What what

I'm thinking about is when you do pre-training, you take the internet and you just say ask what's the next word.

You know, you could think before you ask what's the next word. Obviously, now you don't want to think before every word probably, but I don't know if you've ever looked at the training data for

like a real pre-training run cuz I think people mostly don't realize how bad this is. Like hotels.com is a great website

is. Like hotels.com is a great website compared with the average chunk of 2,000 words from the internet. It's it's a mess, right? And also a miracle that

mess, right? And also a miracle that from this the pre-training process gets you something reasonable. So you

probably don't want you know imag imagine you have a hotel website telling you know it's a beautiful vacation you don't necessarily want to have a very long chain of thought before that right if it was written by a person there was

probably some kind of thinking that went into it maybe maybe not as elaborate as the math and coding thinking but maybe there was something going on so maybe you want a little bit of thought before

at least some of the text and that our models can't do very well yet I I think they're starting I think it it there is a lot of generalization in in this reasoning. If you learn to think for

reasoning. If you learn to think for math, you can you will sometimes do some you know some strategies are they transfer very much like look up on the web and see what they say and and use

that information. So some of these

that information. So some of these things are very generic and they start to transfer. I feel like summer maybe

to transfer. I feel like summer maybe not yet especially thinking in in the visual domains is very undertrained I believe

but you know we work so so so we will try to try to push for for more of that.

>> Going back to chain of thought how does that actually work? How does the model decide to create that chain of thought and is what we see so the little um you

know intermediary steps that we see on the screen as users uh the the chain of thought that's exposed to us is that what's actually being processed by the

model or is there a deeper longer broader chain of thought that happens behind the scenes >> so in the current chat GPT you will see a summary of the chain of thought on the site so there is another model that

takes the full chain of thought shows you a summary because the full ones are usually not very nice to read. They're

more or less say the same thing just just in more messy words. So So it's better to have a more readable summary.

When you start with a chain of thought, the the first paper about chains of thought, you basically just ask the model please think step by step and it would it would think. So if you just pre-train a model on the internet and

ask it to think step by step, it will give you some chain of thought.

Interesting and most important point is you don't stop there. You say okay so you have you start with some way of thinking and then you say sometimes this leads to a correct answer and sometimes

it leads to a wrong answer. So now I'm telling you I have some training examples you will think a 100 times and say 30 lead you to the correct answer then I'll train you on these 30 examples say this

is the way you should be thinking that's the reinforcement learning part of training changes dramatically how the models think. We see this for math and

models think. We see this for math and coding. But but the big hope is it could

coding. But but the big hope is it could also change how the models think for many other domains. Even for math and coding, you you start seeing that the models start correcting their own

mistakes. Right? Earlier if the model

mistakes. Right? Earlier if the model made a mistake, it generally just tell you what it did and insist that the mistake was right or something like that. With with with the thinking, it's

that. With with with the thinking, it's like, oh, I often make mistakes, but I need to verify and correct myself to to give the correct answer. So, so this just emerges from from this

reinforcement learning which is beautiful, right? It it's clearly a good

beautiful, right? It it's clearly a good thinking strategy to to verify what you want to say. And if you think it may be an error, then think again.

That that's what that's what the model learns on the most abstract level.

>> Great. Thank you for this. All right. As

a quick uh detour and we'll go back to more Frontier AI topics. I'd love to talk a little bit about um your story. I

mean you have the incredible distinction of uh having been at the forefront of this industry both the transformers paper which was the the birth of one paradigm and now you're very much

leading the charge on the reasoning model part which is another paradigm so this just incredible story how did you become an AI researcher >> I was a mathematician and a computer scientist but but in theoretical

computer science >> and that started in high school as a kid >> yes I was definitely very into math math in high school and into computers also

later in high school. Yes, I did my studies in Poland. I went for a PhD in Germany. It was the theoretical computer

Germany. It was the theoretical computer science and mathematics PhD. So I very much I'm a mathematician. Yeah. I I I was always fascinated by you know how

how is this thinking going? What is

intelligence? As as a child I always wanted to like emulate the brain. They

thought well okay maybe higher level explanations are more interesting. I did

research in logic but a little programming but then there was this opportunity to join Google just as the deep learning was starting off. I

already had my tenure position in France and the French system has this beautiful thing that you can take a leave of 10 years.

>> Yes.

>> And and you can still return anytime you want from the so so it's a no risk.

>> So at some point when you when you've solved AGI you may return back to France and be a professor. Well, if you solve AGIA, they may take you anyway.

The the nice part about the leave is that they will take you back even if you don't. So,

don't. So, but it's actually very important. I a

number of I I think there's a number of Nobel Prize winners who who took this leave to just try something more risky and you know, sometimes it works, sometimes it doesn't. there there's a

lot of luck in in in science and research but it's very good to have this opportunity to take it so I came to Google that was Google brain at the time you said right >> I came to Ray Curtzswwell's group he was

my first manager he interviewed me and was very inspiring I I I was my first interview was to join like the YouTube UI team and I was like okay I'm not going and and then I had an interview

with Ry and I knew him of course from his books and and he's a very inspiring person so I was like Okay, let's go. The

team was separate from Google Brain at that time. Then I moved to Google Brain,

that time. Then I moved to Google Brain, worked with Ilia Discover, another very inspiring person. There's an amazing

inspiring person. There's an amazing number of great people in in AI in the Bay in general.

>> I have to ask you at this point about the Transformer paper story, how it all came about. The eight of you, right?

came about. The eight of you, right?

Seven or eight of you. How did you all get together?

>> Well, we never got together.

>> You never got together. Okay. I I

recently got a photo on Twitter of a photo session of all eight of us and it it it's it was saying it was fake but I knew it was fake because I don't think all eight of us were ever in the same

physical room. These ideas developed

physical room. These ideas developed from many sites before and after like Jakob were work and Tilia Pushukin worked on attention like self attention

of course attention was there from the encoder decoder side >> and maybe maybe one minute for the broad public on what attention actually means since it's such a fundamental concept.

So, so, so attention is the mechanism that tells the model as as you're doing the next thing, look into your past and find the most similar things that you

see in the past to to what you are seeing right now. It came from the machine translation times where people wanted to align words in one language

with words in another. They were like, okay, so this word where in this previous sentence would it be? It it

it's it's an analog of alignment for deep learning. It's now called attention

deep learning. It's now called attention and in in AI just says you know think of like what comes to your mind as you are here now in in this environment what

things from the past are are similar to it and this mechanism was already used in in in deep learning translation before but it was used there was like one encoder model and the decoder would

be looking at the encoder but never at its own states. main novelty of Transformer was self attention. But

Transformer is more more than just this idea. I I think that's important. I

idea. I I think that's important. I

think that's the the beauty of these eight weird people somehow came together even though not physically to to do it is that we all approached it from

different sides. So so so there was

different sides. So so so there was people working on the attention idea.

There is the you know you need to put this in the network that needs to have a lot of knowledge. So, so there is the feed forward layer that expands and then contracts. Zunam was working on this and

contracts. Zunam was working on this and nowadays used mixtures of experts which actually came before the transformer.

So, so how do you store knowledge in neural networks is is another important question and it's part of this model too. And then you know in deep learning

too. And then you know in deep learning people laugh that ideas are cheap making them work is is the hard part. So

>> how do you write the systems and the code and the baselines to actually make this train? And and this is funny to say

this train? And and this is funny to say now because nowadays you can take any deep learning framework and say you know X equals transformer X train and it will

basically work but back then it totally did not. So you need things like

did not. So you need things like learning rate warm up or or tweaks to the optimizer that that were just work.

And I did a lot of coding and at that time was working on TensorFlow and parts of the framework. And I I remember distinctly that people were like, "So you want to use the same model for a few

different tasks?" Like why do you even

different tasks?" Like why do you even do that? Like if you have a different

do that? Like if you have a different task, like if you do translation, you train one model, if you do parsing, you train another. If you do image

train another. If you do image recognition, you train a third. Like you

never train the same model for three different tasks. Why do you even write

different tasks. Why do you even write like APIs to to do multiple tasks on one model? And I was like, "No, no, we're

model? And I was like, "No, no, we're going to do all tasks in one model." And

people were like, "No, no."

>> So there was a lot of push back against the idea.

>> Not against the idea. Google was also an amazing place at that time that they would very happily let you work on whatever you wanted. But but I don't think there was widespread belief in

doing multiple tasks with the same model. Not to mention,

model. Not to mention, you know, this idea that you I I still find this idea that you take basically the same model as transformer. Like now

there is a bunch of changes to it, but you could in principle take the same architecture as the decoder from the paper, train it on all of the internet and it will basically start chatting with you. It would have back then

with you. It would have back then definitely sounded as a worthy dream. We

maybe had as a dream but not reality that you'd expect 5 years later. It's

very lucky that it actually works so well right?

>> Talk about the transition from uh from Google to open AI and perhaps how those two cultures are different.

>> So, I cover was my manager at Brain.

Then he he went on to to found OpenAI.

He asked me a number of times whether I would like to join in in the years. I

found it a little bit too edgy at the time. Then Transformers came. So, we had

time. Then Transformers came. So, we had a lot of work with that. And then COVID came and and COVID was a tough time for for for the whole world, right? But

Google totally closed. Google was

reopening extremely slowly. So So one part of me was I I find it very hard to do remote work. I much prefer to to work with people directly. That was one

reason. But the other was also Google

reason. But the other was also Google Brain when I joined it was a few dozen people, maybe 40, something like that.

When when I left it was 4,000 3,000 people. spread across multiple offices.

people. spread across multiple offices.

It it's very different to work in a small group and to work in a in a huge company. So with all this, IA was like,

company. So with all this, IA was like, you know, OpenAI though is in a much stabler state. We're doing language

stabler state. We're doing language models. You know, something about this

models. You know, something about this that may look like a good match. And I

was like, okay, let me try. I've never

worked in any company other than Google before, other than the university. So it

was quite a change to the small startup group, but I I I I like working in smaller groups. It's it it it has its

smaller groups. It's it it it has its pleasure, right? It has a little bit of

pleasure, right? It has a little bit of a different in intensity sometimes. In

general, I I I I found it very nice. On

the other hand, Google, you know, has merged made the Gemini and I hear it's also a very nice place. I I think in in general the the tech

labs are more similar to each other than people think. There are some differences

people think. There are some differences but but I think if if I look from at it from the world you know from the university in France the difference between this university and any of the

tech labs is much larger than than between one lab or the other.

>> How are the research teams organized within OpenAI?

>> Um they're organized they're not very organized. I mean we do organize them but some you know people have managers and we sometimes talk to

them. Um no but mostly people find like

them. Um no but mostly people find like projects there there are things to do right like improve your multimodel models improve your reasoning improve

your pre-training improve whatever this part of the infrastructure people work on it you know as we go through these parts right there is infrastructure pre-training reasoning I think the parts

are the same for for most of the labs so so there will be teams doing these things and and then sometimes people um change teams sometimes new things emerge

There is always some smaller teams doing like more adventurous stuff like diffusion models at times. Then you know some of the more adventurous stuff like

video models gets big and and and then maybe they need to grow.

>> Do people compete for GPU access?

>> I don't think it's so much people that compete. I think it's more projects that

compete. I think it's more projects that compete for GPU access. There's

definitely some of that. On the other hand, like on the big picture of GPU access, a lot of this is just determined by how the technology works, right?

Currently, pre-training just uses the most GPUs of all the parts. So, it needs the most GPUs, right? RL is growing in the use. Now, video models of course use

the use. Now, video models of course use a lot of GPUs too. So, you need to split them like this. Um, then of course, you know, people will be, "Oh, but my thing would be so much better if I had more

GPUs." And I've certainly said that a

GPUs." And I've certainly said that a number of times too. Um so then you kind of push like you know I really need more and then some people may say well but you know there's only so much there is

never enough GPUs for for everyone. So

there is some part of the competition but the big part is just decided by the fact how the technology works currently.

>> Great. What is um next for pre-training?

We talked about uh data, we talked about engineering a big uh GPU compute aspect to this. Uh what happens to pre-training

to this. Uh what happens to pre-training in the next year or two? pre-training as

I said I I think it has reached this upper level of the scurve in terms of science but it can scale smoothly meaning if you put more compute you will

get better losses if you do things right which is extremely hard and that's valuable um you don't get the same payoff as as

as pushing carell but it generally just makes them all more capable and that's certainly something you want to do I think what people underestimate a little

bit in in in the big narrative is you know open AAI 3 four years ago I I joined even before that was a small

research lab with a product called API but you know it it was not such a big there was no GPU constraint on the product side for example all GPUs were

just used for training so it was very easy as a decision for the people to say you know we're going to train GPT4 this will be the smartest and largest model ever. And what do we care about small

ever. And what do we care about small models? I mean, we care of them as to

models? I mean, we care of them as to make like to debug the training of the big model, but that's it. So, GPT4 was the smartest model and it it was great,

right? But then it turned out, oh, there

right? But then it turned out, oh, there is this chat and now we have a billion users and you know, people want to chat with it a lot every day and you need GPUs.

So you train the next like huge model and it turns out you cannot satisfy this like people will not want to pay you enough to chat with the bigger model. So

you just economically need the smaller model. So and this happened of course to

model. So and this happened of course to to all the labs because like the moment the economy arrived and it became a product you had to start thinking about

price much more carefully than before.

So I think this caused the fact that instead of just training the you know largest and largest thing you can for for the money you have we said well no we we're going to train like the same

thing but same quality but smaller cheaper the pressure to to to give the same quality for less money is very large in in some senses as a researcher almost

makes me a little sad I have you know a big love for these huge models that um you know people say human brain has 100 trillion synapses or orders of magnitude of course are not that exactly

calculated but our models don't have 100 trillion parameters yet so maybe we should reach it I I would certainly love it but then you need to pay for it so I think this is this may be why people

kind of think that pre-training has paused because a lot of effort went into training smaller models now on the side people kind of rediscovered how amazing distillation is distillation means you

can train a big model and then put the same knowledge from the big the big is a teacher to the little one. People knew

about distillation. It's a paper or a long time ago but somehow at least for open AI I think maybe it was more in Google's DNA when Oriel was

there. Uh but people kind of

there. Uh but people kind of rediscovered how important that is for the economics but now it also means that oh training this huge model is actually good because you distill all the little

ones from it. So now maybe there is a bit more of a return to like it's also matter you know once you realize you have the billion users and you need the GPUs you need to invest into them and of

course we're everyone sees this there's a huge investment but the GPUs are not online yet so when they come back online and I think this may play into this what people call resurgence of pre-training

we both understand that you can distill this amazing big model and there is now enough GPUs to actually train it so so so it's resurging But all of this fundamentally happens on the same

scaling curve, right? It's it's not like we didn't know that you could do this.

It's more like the different, you know, different requirements of different months have have changed sometimes changed the priorities. But I think it's good to step back from it and and and

think of the big picture, which is that pre-training has always worked. And and

the beautiful thing is it even stacks with RL. So if you run this thinking

with RL. So if you run this thinking parallel process on top of a better model, it works even better than if you run it on top of a of a smaller model.

So So >> one question that I find fascinating as I hear you speak and the evolution of the modern AI systems has been this combination of LLM plus RL plus a lot of

lot of things going on. It used to be at some point and maybe that was back in the deep learning days that uh people would routinely say that uh they they

understood how AI worked at a at a micro level like the matrix multiplication aspect but didn't fully understand once you had everything together what really

really happened at the end of the day in the model and I know there's been tons of work done on uh interpretability over the last couple of years in in particular uh But particularly for those

very complex systems, is it increasingly clear what the models do or is there some element of blackbox that persists?

>> I would say both. There is a huge progress in understanding models.

Fundamentally, I mean think of the model that is chat. It it talks to a billion people about all kinds of topics. It

gets this knowledge from reading all of the internet. Obviously, you cannot

the internet. Obviously, you cannot identify like I cannot understand what's going on in there. I don't know the whole internet. What we can identify is

whole internet. What we can identify is there was a beautiful paper just I think last week with from open AI about if you tell the model that lots of its weight should be zeros it should be very sparse

then you can really trace when it's thinking about one particular thing then you can trace what's actually what it's actually doing. So if you say limit

actually doing. So if you say limit ourselves to this and you really study this inside a model then then you can get a lot of understanding and and there there's circuits in the models entropic

had great papers on that.

The understanding of what the models are doing on a higher level has progressed a lot. But then it's still an

lot. But then it's still an understanding of what smaller models do, not the biggest ones. But it's not so much that these patterns don't apply to bigger models. They do. It's just the

bigger models. They do. It's just the bigger models just do so many things at the same time that there is some limit to what you can understand. But I think this limit is is a bit more fundamental

than people think. It's like every very complex system. You can only understand

complex system. You can only understand so many things and then you don't.

Right.

>> Thank you for all of this. I'd love now to talk about 5.1 and do a little bit of a deep dive on all the latest stuff that you guys have released in the last couple of weeks which has been very uh

impressive. In particular as a user

impressive. In particular as a user think that the 5.1 moniker doesn't do justice uh to the evolution between 5.1

and five. It it feels like a much larger

and five. It it feels like a much larger improvement than the number would would indicate from uh again my perspective as a as a user. Walk us maybe through the

evolution of uh from GD4 to 5 to 5.1.

What has actually changed?

>> That that's a very tough question. I

think less than you think it's um no I mean the from GPD4 to 5 I think the biggest thing that changed is

reasoning meaning RL and synthetic data as I told you the pre-training part in that time frame was mostly about making things cheaper not making things better so of course the price has changed

dramatically too right thousand times I think or some of these orders of magnitude the the main improvements from four to five is adding reasoning with

reinforcement learning and this allowed to generate synthetic data which also improves the model. So that's the that's the big picture. In addition to that, CH

GPT is is now a product used by a lot of people. So the the post- training team

people. So the the post- training team has learned a tremendous number of lessons and it's added you know things clearly experimented wanted the model to be very nice to you then it turned out

to be too nice. Then now when a lot of people use it, you need to be really careful about safety, right? There may

be people that are in distress using the model. The model needs to do something

model. The model needs to do something reasonable in these cases. It was not trained for it before. Now it is and and it makes the model much better. But you

know at the same time you don't want to refuse to answer any question that has any sign of anything. So, so as you work on these things, you make the model much better in use, not not just for the

people in distress, but but for everyone who wants questions answered, but the answers to be reasonable. And you know, there was these things called hallucinations. It's still with us to

hallucinations. It's still with us to some extent, but dramatically less than than two years ago. Some part of that is because reinforcement learning can now

use tools and gather data and it also encourages the model to to like know you know verify what it's doing. Uh so

that's an emergent thing from from this reinforcement learning of reasoning but also you just add data because you realize sometimes the model should say I

don't know. So you add this to the post-

don't know. So you add this to the post- training data. You say like well we

training data. You say like well we really need to give it a thought how the model should answer people in in various situations. The 5 to 5.1 is is mostly

situations. The 5 to 5.1 is is mostly this kind like it's mostly a post-training improvement.

>> Yeah. So to double click on this because this super interesting. Indeed. Um, as

part of 5.1, there's, uh, the ability to choose different kind of styles from nerdy to professional. And, uh, that's, uh, I guess in reaction to the fact

that, um, some people were missing the sequantic aspect of earlier models when chat GBD when when GBD 5 came out. And

so, so adding more tones, that's all post trending stuff. So you tell the model those are examples of like how you should be responding which is more like

a like supervising kind of a paradigm or is that uh is that RL like right or wrong with rewards? How does that work?

>> I don't work on post training and it certainly has a lot of quirks but I think the main part is is indeed RL where you say okay is this response cynical is this response like that and

and and you say okay if you were told to be cynical this is how you should response. If you were told to be funny,

response. If you were told to be funny, try this one. So I I do think there is RL is is is a big part of >> in between models uh or or or different

versions of the models. Are the are the releases aligned with pre-training efforts or sometimes you have like one big pre-training effort and like several models that that come out based on that.

There used to be a time not that long ago be half a year distant past where where where the models were did have an

alignment with with technical stuff right so they would align whether with either with RL runs or pre-training runs that's why you had um a beautiful model

called 40 which was aligned with a pre-training run which uh was obviously worse than the 03 aligned with an RL run that was the follow-up to 01 naturally

because you couldn't use the name O2 but it was slightly better than the O4 mini because that one was mini and you know we had this beautiful model picker and

people kind of thought this was not the best naming for some whatever reason so um no I mean it was fairly obvious that

this was very confusing right so so so now the naming is by capability right GPD5 is a capable model 5.1 is a more capable model. Mini is the smaller model

capable model. Mini is the smaller model that's slightly less capable but faster and cheaper and the thinking models are the one that do more research. Right?

And that sense the naming is detached from any technical in particular you know 5.1 may be just a pre-training uh sorry post-training thing but maybe 5.2

towards the newly pre-trained model or maybe not but but but the naming has detached from the technology which also gives some you know as open has grown there is a number of projects right

there is RL and pre-training and there maybe you know something just to make slides better or whatnot um and with distillation you have the ability to put a number of projects into one model it's

kind of nice that you don't need to wait on all of them to complete at the same time and you can try to periodically put together actually make sure that as a product it's nice to the users and good

and and do this separately from you know waiting on the new full pre-training run that takes months and so on. So I feel like even though you know a little tear in my eye goes for the times where it

was that pre-trained model number that that was the number as it's a product serving a billion users it's maybe inevitable that you should name it by what the user should expect from it

rather than >> in 5.1 uh you have additional granularity uh in terms of uh telling the model how

long it should think by default. How

does the model decide how long it should think?

>> So the model sees the task. It will

decide on its own a little bit how long it should think. But you can put give it an additional it's trained with an additional information that can tell it think even harder and and then it will think longer. So you have now the

think longer. So you have now the ability to steer that I I still think it is important to realize.

So this is the fundamental change that came with reasoning models that using more tokens to think increases your capability and it increases it given the

computation way faster than pre-training right so so if you give GP5 the ability to think for long it can solve tasks

that are you know we had these gold medal at mathematical olympiad and computer science olympiad so so amazing abilities.

At the same time, the the fundamental training method of of reasoning is very limited to science data. So, so it's not as broad as the

data. So, so it's not as broad as the pre-training, which I I think like pre-training models felt kind of almost uniformly good or bad at things. I mean,

this was still not uniform because it's not like teaching humans, right? But the

reasoning models are even more people call it jagged, right? They have amazing abilities somewhere and then close by not so much. And that can be very

confusing. It It's something I I always

confusing. It It's something I I always love that it's weird because you can say the model is amazing at mathematical olympiad. At the same time, I have a

olympiad. At the same time, I have a math book for um I have a first grader daughter in the first grade. She's 5

years old. I took one exercise from this math book and none of the frontier models is able to solve it. and you

would be able to solve it in 10 seconds.

So that's something to keep in mind.

Models are both amazing and there are tasks that they cannot do very well. I I I can show you this as an

very well. I I I can show you this as an example. I think it's quite interesting

example. I think it's quite interesting to keep in mind. Let me start with Gemini 3 just to blame the competitors.

>> Yes, please.

>> So it has you see two two groups of dots on both sides and the question is is the number of dots even or if you look at it you see oh they're like two identical things. So that would be even. That's

things. So that would be even. That's

what the 5-year-old is supposed to learn. But there is one dot that's

learn. But there is one dot that's shared. So So now that that must be odd

shared. So So now that that must be odd for this simple one which has like you know I don't know 20 dots or so. Gemini

3 actually does it right? That it it finds out that that that it's an even number of dots and and it says that and that's great. And then you have another

that's great. And then you have another puzzle which is very similar except now there are two mountains of dots and there's also one dot shared at the bottom now and right in context right

after that you ask okay how about this one and then it it it it does some thinking and it just totally misses that there is a shared dot and it says the number is even and it's like in context

where you've seen this first example how would you ever miss that you know and here is the same the exact same prompt for GPT 5.1 point when thinking and it also solves the first it it it sees the

dot. It says it's odd and then it sees

dot. It says it's odd and then it sees the mountains and somehow it doesn't see the dot and it says it's even. The the

the nice thing is if you if you let it think longer this is like or if you just let it think again it will see it. So so

if you use GPT5 Pro it takes 15 minutes.

So you know this is the the human 5-year-old takes 15 seconds. The GPD51

Pro will run Python code to extract these dots from an image and then it will count them in a loop. So that's not quite >> and and why why is that? What trips up the model?

>> I I think this is mostly multimodal part. The models are just they're

part. The models are just they're starting like you see the first example they managed. So, so they've clearly

they managed. So, so they've clearly made some progress, but they have not yet learned to do good reasoning in multimodal domains and they have not yet

learned to use one reasoning in context to do the next reasoning. What is

written in context is is you know learning in context happens but learning from reasoning in context is still not very strong. All of these though are

very strong. All of these though are things that are very wellnown and like the models are just not trained enough to do this. it's it's just something we know we need to add into training. So I

think these are things that will generally improve. I do think there is a

generally improve. I do think there is a deeper question whether so you know like multimodel will improve this will improve like we keep finding these examples. though as the frontier will

examples. though as the frontier will move it will certainly move forward some things will smoon but the question is will it still be just other things that you don't need to you know teach the

human like every you know okay now you know how to use a spoon um and a fork but now if the fork has four instead of

three ends then you need to learn a new that would be a failure of machine learning you know I am fascinated by generalization I think that's the most important topic I always thought this was the key topic

in machine learning in general and in understanding intelligence. Pre-training

understanding intelligence. Pre-training is a little different, right? Because it

increases the data together with your increase in model size. So it doesn't necessarily increase generalization. It

just uses more knowledge. I do believe that reasoning actually increases generalization, but now we train it on such narrow domains that that it's may still be to see. But I think the big

question in all of AI is is reasoning enough to increase generalization or do you need like more general methods? I

think the first step is to make reasoning more general. As we talked before, I that that's my passion. That's

also what I work on. There is still something there, right? We we push the models. They they learn they learn

models. They they learn they learn things that are around what we teach them. they they still have limitations

them. they they still have limitations because they don't live in the physical world because they're not very good at multimodal because reasoning is very young and there's a lot of bugs in how

we do it yet. But once we fix that there there will be this big question is that enough or is there like something other big to to make models generalize better so we don't need to you know teach it

every particular thing in the training data that it just learns and generalizes. I think that's the most

generalizes. I think that's the most fascinating question. But I also think a

fascinating question. But I also think a good way to approach a question like that is to first solve everything that leads up to it. You know, you cannot know whether there is a wall or not

until you come close to it because otherwise there you know we AI is moving very fast. Someone said it's like

very fast. Someone said it's like driving fast in a fog. You never know how how far or close you are. So so we we're moving we are learning a lot. And

does that does that mean so so that that central uh question of basically learning with with very little data the way a child would and the fact that the child is able to do things that even the

most powerful model cannot do. So this

as you said to unpack this uh making progress on reasoning and showing how far we can get into generalization with reasoning and then the separate question is as you as you said whether we need an

entire different architecture and that's where we get into for example Yan Khan's work do you see uh promising fundamental

architectural changes outside of transformers that have caught your attention and and feel like they could um like a serious path to explore in the future. I

future. I >> I think there is a lot of beautiful work that that people are trying out. You

know, the arc challenges inspired one set of people. The their models now that are very small and solve them very well, but with methods that I'm not sure are actually general. We'll need to see. Yan

actually general. We'll need to see. Yan

Leon has been pushing for for other methods. So, I feel like his approach is

methods. So, I feel like his approach is is more towards the multimodal part. But

maybe the maybe if you solve multimodal, right? Maybe if you do jetpa it also

right? Maybe if you do jetpa it also helps your other understanding. There is

a lot of people pushing fundamental science still. It's maybe not so much in

science still. It's maybe not so much in the news as as as the you know the things that that push but whatever you do you know it will probably run on some

GPU. If you get a trillion dollar of new

GPU. If you get a trillion dollar of new GPUs, the the old GPUs will be much easier to get also for for so I think this growth in LLMAI on on the more

traditional side is also helping people to to have an easier time to run more experimental research projects on on various things.

So, so, so I think there is a lot of exploration, a lot of ideas. It's still

a little hard to implement them at a higher scale. the engineering part is

higher scale. the engineering part is the biggest bottleneck. I mean GPUs are a bottleneck too when you scale really up but implementing something that's larger than one machine it's an

experimental research project so you don't have a team to do that I think that's still harder than it should be but you know codecs may may get there or

coder this is the thing where AI researchers have great hope to help themselves and and also other researchers is that if you could just say hey codeex this is the idea and it's

fairly clear what I'm saying please just implement it so it runs fast on on this eight machine setup or or 100 machine setup. That would be amazing. It's not

setup. That would be amazing. It's not

quite capable to do of doing that yet.

But, you know, it's capable of doing this more and more. I think that's what what OpenAI says is they say, you know, we we say we'd like an AI intern by the end of next year. That's how I understand this. You know, can can

understand this. You know, can can someone help us? And is is is part of Codeex's um the path for Codeex to be able to do some of this uh does that

revolve around how long it can run?

Context behind the question being that uh again like two days ago as we record this you guys released GPT 5.1 C codeex max described as a frontier agent coding

model trained on real world software engineering tasks designed for longunning workflows and using compaction to operate across multiple context windows in millions of of of

tokens. So I'd be interested in

tokens. So I'd be interested in unpacking some of this. What does that mean to run for a very long time? Is

that an engineering problem or a model problem? And then maybe a word on

problem? And then maybe a word on compaction.

>> So it is an both an engineering and model problem. Um you know you want to

model problem. Um you know you want to do some engineering task like write a you have some machine learning idea. You

want Codex to implement it for you. Test

it on some simple thing. Find the bugs.

So it needs to run this thing. This is

not something you would do in an hour, right? That's it's something you'd spend

right? That's it's something you'd spend a week on. So the model needs to spend a considerable amount of time because it needs to run things, wait for the results, then fix them. The model is not like it's going to come up with the

correct code out from thin air, right?

It's just like us. It needs to go through the process and often times in the process since it was not trained on anything very long in its in its training or maybe very few but but

nothing certainly nothing that went on for a week, it can get lost. it it can start doing loops or doing something weird. That's of course not something

weird. That's of course not something you want. So, so, so we try to train in

you want. So, so, so we try to train in a way that makes it not happen, but but it does. So, so you know, how can you

it does. So, so you know, how can you make the model actually run a process that requires this larger feedback loop without tripping up? And

the other thing is transformers have this thing called context. So they they remember all the things that they have seen in in the current run and that can

just exceed the memory available for your run and and the attention matrices are are n byn where n is this length so so they can get huge. So instead of

keeping everything, you say, well, I'm going to just ask the model on the side to summarize the most important thing from the past, put it in context a new and forget some part of it. Right? So

it's a very basic form of forgetting the compaction. Right? And that allows you

compaction. Right? And that allows you to run for much longer if you do this repeatedly. But but again, you need to

repeatedly. But but again, you need to train the model to do that. And and when you do it, it works to some extent. I

don't think it works well enough to replace an AI researcher yet. It made a fair bit of progress. I I think another part of progress that that's a little understated on the research side but is

very important is allowing the model to connect to all of these things. So

models now use tools like web search and Python routinely but to run on a GPU to have access to a cluster. It's hard to train models with that because then you need to dedicate

for the model to use and and and that has security problems and this thing like how do models connect with the external world? It's a fundamentally

external world? It's a fundamentally very hard problem because you know when you connect in an unlimited way you can break things in the real world and we don't like models to break things for us. So that's a part where people work a

us. So that's a part where people work a lot. This it overlaps with security,

lot. This it overlaps with security, right? You need to like have very good

right? You need to like have very good security to allow models to go on and train on the things they need to train.

One theme that people like me, VCs and uh you know founders and startup think about a lot as we see all the progress at OpenAI is as the models keep getting

more general with more agent capabilities the ability to run for a very long time you know going to areas like science and uh math and you know

recently it was reported that um there were some investment bankers hired to help improve the model's capability to do grunt investment banking uh work all

of that taken into account uh is is there a world where basically models or maybe just one model does uh everything and I don't know if that's a GI let's not necessarily go into that debate uh

but what's left for uh people that build products that sit on top of models >> I just showed you a 5-year-old exercise that the model doesn't do I I I I think

we need to keep that in mind >> so you're saying there's hope There is hope the next model will do it right >> that hope okay >> well for for me yes I I still think we

have some way on the models progress has been rapid so so there is good hope there will be less and less of this but but on the other hand for now it's you

don't need you know to do a deep search to find things where you'd really want a human to to do that task because the model's not super good on the other hand I you know transformer paper started

with translation I recently went to a translation industry conference. Uh the

translation industry has grown considerably since then. It has not shrunk. There's more translations to be

shrunk. There's more translations to be done. Translators are paid more. The

done. Translators are paid more. The

question is why would you even want a translator if the model's so good in most of the cases. The answer is sometimes imagine you do a listing for a newspaper but in a language you don't

know and GPD5 will almost certainly translate it correctly for you. If it's

into Spanish or French or any high resource language, would you still publish it without having a human who speaks that language look at it? Would

you publish it if it's a, you know, UI of Chat GPT that a billion people are going to see? It's a question of trust probably, right? But if you have a

probably, right? But if you have a million users, a billion users, maybe you will pay the $50 for someone to just take a look over it before you you

translate it. So this is in an industry

translate it. So this is in an industry that fundamentally is totally automated, right? There is still the question of

right? There is still the question of trust and I think that's a question we will grapple with for a long time. There

are also just things you want a person to do like I I I don't think we will have no things to do but but that

doesn't mean that some things we do may not dramatically change >> uh and that you know that can be very painful for people who do these things and and so this is a serious topic that

happy people are engaging with. But I

don't think like there will be this global lack of anything for people to do.

>> And maybe as a last question to help us uh get a sense for what people at the frontier of AI are currently thinking about or working on, you know, some of

the topics that uh one may see are things like continual learning, world models, uh robotics, uh embedded intelligence. What do you

personally find very interesting in addition to what you mentioned a prompt multimodel uh but what do you personally find really interesting as a research area? Well, you know, I I always I find

area? Well, you know, I I always I find this general data reinforcement learning is is my my pet peeve and what I work on

luckily that for example, robotics is probably just an illustration that we are not doing that well in multimodel and that we're not doing that well in

general RL in general reasoning yet. The

moment we do really well in multimodel and we manage to generalize reasoning to to the physical domains that the robot needs, I think it will see amazing progress. When it does, I have a feeling

progress. When it does, I have a feeling given that you know a lot of companies are launching hardware that's kind of teleoperated or glove operated or or or something. So my suspicion is by the

something. So my suspicion is by the moment we make this progress which you know maybe maybe it will be next year maybe it will be in a few more years but

the hardware will may be there by then and having a robot in a home may be like a big visible change

more visible than you know chat I mean given how quickly we got used to the self-driving cars in San Francisco maybe it will be only visible for like the first two days and then be Yeah,

sure. The robots's there. It's always

sure. The robots's there. It's always

been cleaning since I can remember like the last three months.

It's It's stunning to me how quickly we get used to these things, right? The the

self-driving cars in San Francisco are something people got used to so quickly.

So, maybe this will happen for robots, too. Nevertheless, I do think it will be

too. Nevertheless, I do think it will be quite dramatic in in our perception of of the world when it happens. It

hardware is hard though, right? Robots

may have accidents in the house. you

need to be very careful. So, so maybe it will take longer to to deploy them and actually make it a scalable business.

We'll see. It is amazing that you know we are at this point where we can start thinking like yes maybe that will come soon.

>> Lucas, it's been absolutely wonderful.

Uh thank you so much for spending time with us today.

>> Thank you very much Matt. Thank you for the invitation. Great to talk to you.

the invitation. Great to talk to you.

>> Hi, it's Matt Turk again. Thanks for

listening to this episode of the Mad Podcast. If you enjoyed it, we'd be very

Podcast. If you enjoyed it, we'd be very grateful if you would consider subscribing if you haven't already, or leaving a positive review or comment on whichever platform you're watching this or listening to this episode from. This

really helps us build a podcast and get great guests. Thanks, and see you at the

great guests. Thanks, and see you at the next episode.

Loading...

Loading video analysis...