Sergey Brin on Gemini 1.5 Pro (AGI House; March 2, 2024)

AttentionX
3 Mar 202436:59

Summary

TLDRThe speaker introduces an AI model called Gemini 1.5 Pro, explaining it performed much better than expected during training. He invites the audience to try interacting with the model and asks for questions. When asked about problematic image generation, he admits they messed up due to insufficient testing. He acknowledges text models can also say peculiar things if prompted aggressively enough. He claims Gemini 1.5 Pro text capabilities should not have such issues except the general AI quirks all models exhibit. Overall, he is excited about Gemini's potential for long context understanding and multimodal applications.

Takeaways

  • ๐Ÿ˜Š The chat reveals behind-the-scenes info about the AI model Gemini 1.5 Pro, saying it performed better than expected during training.
  • ๐Ÿค“ Gemini is experimenting with feeding images and video frame-by-frame to the models to enable them to talk about the visual input.
  • ๐Ÿ˜Ÿ The speaker acknowledges issues with problematic image generation and text outputs from AI models.
  • ๐Ÿง Efforts are ongoing to understand why models sometimes generate concerning outputs when prompted in certain ways.
  • ๐Ÿ‘ฉโ€๐Ÿ’ป The speaker personally writes a little bit of code to debug models or analyze performance, but says it is probably not impressive.
  • ๐Ÿค” In response to a question, the speaker says today's AI models likely can't recursively self-improve sophisticated systems without human guidance.
  • ๐Ÿ˜Š The speaker is excited about using AI to summarize lengthy personalized information like medical history to potentially enable better health diagnoses.
  • ๐Ÿ˜• The speaker says detecting AI-generated content is an important capability to combat misinformation.
  • ๐Ÿค” When asked if programming careers are under threat, the speaker responds that AI's impacts across many careers over decades is difficult to predict.
  • ๐Ÿ˜€ The speaker expresses optimism about AI advancing healthcare through better understanding biology and personalizing patient information.

Q & A

  • What model was the team testing when they created the 'goldfish' model?

    -The team was experimenting with scaling up models as part of a 'scaling ladder' when they created the 1.5 Pro model they internally referred to as 'goldfish'. It was not specifically intended to be released.

  • Why was the 1.5 Pro model named 'goldfish' internally?

    -The name 'goldfish' was meant ironically, referring to the short memory capacity of goldfish. This was likely meant to indicate the limits of the 1.5 Pro model's memory and context capacity at the time.

  • What issues did the speaker acknowledge with the image generation capabilities?

    -The speaker acknowledged that they 'definitely messed up' on image generation, mainly due to insufficient testing. This upset many people based on the problematic images that were generated.

  • What two issues did the speaker identify with the text models?

    -The speaker identified two issues with text models - first, that weird or inappropriate content can emerge when deeply testing any text model. Second, there were still bias issues specifically within Gemini models that they had not fully resolved.

  • How does the speaker explain the model's ability to connect code snippets and bug videos?

    -The speaker admits they do not fully understand how the model can connect code and video to identify bugs. They state that while it works, it requires a lot of time and study to deeply analyze why models can accomplish complex tasks.

  • What are the speaker's thoughts on training models on-device?

    -The speaker is very positive about on-device model training and deployment. They mention Google has shipped models to Android, Chrome, and Pixel phones. Smaller models trained on-device can also call larger cloud models.

  • What healthcare applications seem most promising to the speaker?

    -The speaker highlights AI applications for understanding biological processes and summarizing complex medical literature. Additionally, personalized patient diagnosis, history analysis, and treatment recommendations mediated by a doctor.

  • How does the speaker explain constraints around self-improving AI systems?

    -The speaker says self-improving AI could work in very limited domains with human guidance. But complex codebases require more than long context, needing retrieval and augmentation. So far there are limits to totally automated improvement.

  • What lessons did the speaker learn from the early Google Glass rollout?

    -The speaker feels Google Glass was released too early as an incomplete prototype rather than thoroughly tested product. Personally lacking consumer hardware expertise then, the speaker wishes expectations were properly set around an early prototype.

  • Despite business model shifts, why is the speaker optimistic?

    -The speaker feels that as long as AI generates tremendous value and productivity gains displacing human labor time and effort, innovative business models will emerge around monetization.

Outlines

00:00

๐Ÿ˜Š Introducing the AI model and its capabilities

Paragraph 1 is an introduction by the speaker about the AI model Gemini 1.5 Pro that they are demonstrating. He explains that it is more powerful than expected, with impressive capabilities, but still requires more testing. He welcomes questions from the audience.

05:03

๐Ÿ’ป Discussing video chat abilities, code contributions, and training costs

Paragraph 2 covers whether the AI could do video chat, with the speaker saying they have done some multimodal experiments. When asked if he writes code, the speaker admits to only minor debugging contributions. He acknowledges that training costs for models are high but the long-term utility is much higher.

10:03

๐Ÿค” Considering recursive self-improvement and AI understanding itself

Paragraph 3 involves a discussion about recursive self-improvement and reflective programming, where AI systems can modify their own code. The speaker sees potential but doesn't think we are at the stage yet where complex code bases could totally improve themselves without human guidance.

15:05

๐Ÿ”ฎ Predicting impacts on industries and potential for on-device training

Paragraph 4 has questions about what verticals will be most impacted by AI advances, which the speaker says is hard to predict, especially with multimodal abilities. He talks positively about the potential for on-device model training and the capabilities of smaller models calling cloud-based models.

20:08

๐Ÿ˜• Discussing transformer limitations and the need for alternate architectures

Paragraph 5 covers whether there are bottlenecks to reasoning abilities with transformer models. The speaker acknowledges theoretical transformer limitations but notes that contemporary versions don't always meet assumptions. He expects continued incremental changes but also anticipates exploration of non-transformer architectures.

25:08

๐Ÿคฅ Considering model hallucination, misinformation generation and detection

Paragraph 6 discusses model hallucination and misinformation generation. The speaker is optimistic that innovations will continue reducing hallucinations but breakthroughs can't be counted on. He notes misinformation is complicated, with issues around political bias, but says detecting AI-generated content is important.

30:10

๐Ÿค– Comparing humanoid robotics now versus the new AI wave

Paragraph 7 covers humanoid robotics, which the speaker worked on previously. He finds software and AI advancing incredibly quickly compared to hardware. Rather than being distracted by today's hardware, he wants to focus on the next level of AI that future hardware will support.

35:11

๐Ÿ’€ Joking about immortality while acknowledging molecular AI progress

The final Paragraph 8 involves a lighthearted mortality question. The speaker admits he is not the expert but has seen huge progress in molecular AI. He expects continued AI benefits for complex health areas like epidemiology, delivering novel hypotheses over time.

Mindmap

Keywords

๐Ÿ’กAI

AI or Artificial Intelligence refers to computer systems that can perform tasks that typically require human intelligence. The video discusses advances in AI with the reveal of Anthropic's conversational AI model Claude. The narrator is excited about the capabilities of models like Claude for powering applications.

๐Ÿ’กconversational AI

Conversational AI involves developing AI systems that can have natural conversations with humans. Claude is introduced as an example of a powerful conversational AI that can understand context and give human-like responses.

๐Ÿ’กlong context

Long context refers to the ability of AI models like Claude to take in large amounts of textual context and use that background knowledge to have more intelligent conversations. The narrator is excited about experimenting with long context in Claude.

๐Ÿ’กmultimodal

Multimodal refers to AI systems that can process different modes of data beyond just text, including images, audio, and video. The narrator discusses Anthropic's interest in developing multimodal capabilities for Claude in the future.

๐Ÿ’กAGI

AGI stands for artificial general intelligence and refers to AI systems with more broad capabilities that approach human-levels of intelligence. When asked if he wants to build AGI, the narrator responds enthusiastically about the exciting advancements towards AGI.

๐Ÿ’กself-improvement

Self-improvement refers to AI systems that can modify or improve their own code or prompts over time. The narrator discusses ideas around self-improvement but feels we are not yet at the stage where AI can deeply self-improve without human guidance.

๐Ÿ’กmisinformation

Misinformation refers to false information spread intentionally or unintentionally. The narrator acknowledges issues around AI hallucinations and the need to bring them down over time to avoid generating misinformation.

๐Ÿ’กrobotics

Robotics involves creating intelligent physical robots and machines. When asked about humanoid robotics, the narrator feels software and AI advances are currently on a faster trajectory than hardware robotics innovation.

๐Ÿ’กhealthcare

Healthcare is discussed as a key industry that could benefit from AI advances. Specific potentials include better understanding biology, improving diagnoses, and providing more personalized recommendations.

๐Ÿ’กimmortality

Immortality involves radical life extension. When jokingly asked about immortality, the narrator does not have a clear answer but feels AI can benefit medical research around extending lifespans.

Highlights

We internally called it goldfish. I don't actually know why - because goldfish have very short memories.

When we saw what it could do we thought hey, we don't want to wait - we want the world to try it out.

I'm grateful that all of you here are here to give Gemini a go.

We definitely messed up on the image generation.

If you deeply test any text model out there...it'll say some pretty weird things.

I invite all of you to try the updated model - it should be at least 80% better.

I've just seen people do...dump in their code and a video of the app and say here's the bug - and the model will figure out where the bug is.

I honestly don't really understand how the model does that.

The long-context queries do take a bit of compute time but you should go for it.

You can learn to understand these models. We can look at where the attention is going at each layer.

I feel like if I get distracted making hardware for today's AIs, that might not be the best use of time compared to what the next level of AI is going to be able to support.

While software and AI are getting so much faster at such a high rate...that feels like the rocket ship.

As computer scientists, seeing what these models can do year after year is astonishing.

AI is the neighborhood getting much better at answering specialized questions - where not many people have written about it already.

Basic information accessed for free, supported by advertising - I think that's great. It gives equal access to a kid in Africa as to the President.

Transcripts

play00:00

um an AI hakon to be this huge that's

play00:05

pretty exciting times well thank you all

play00:07

for coming first of all thanks so much

play00:09

for uh giving Gemini go um what should I

play00:13

say we actually have people who know

play00:15

what they're talking about I think uh

play00:17

Sim that okay Simon okay how's going all

play00:20

right good good um I was worried I would

play00:23

have to say something that uh not quite

play00:25

up to speed on

play00:28

uh I'll just quickly say look it's very

play00:30

exciting times uh this model that um I

play00:33

think we're playing with 1.5 Pro we

play00:37

internally called goldfish I'll just

play00:38

tell you a little secret um I don't

play00:40

actually oh I know why it's because

play00:43

goldfish have very short

play00:46

memories it's kind of an ironic name uh

play00:50

but uh when we were training this model

play00:54

we didn't expect it to have come out

play00:56

nearly as powerful as it did it have all

play00:58

the capabilities that it does um in fact

play01:02

it was just part of a scaling ladder

play01:05

experiment uh but when we saw what it to

play01:08

do we thought hey we don't want to wait

play01:11

uh we want the world to try it out and

play01:13

uh I'm grateful that all of you here are

play01:16

here to give it a

play01:17

go uh what else am I say let me

play01:20

introducing somebody else what's

play01:22

happening next uh I think a people

play01:24

probably have a lot of questions oh okay

play01:26

quick questions I'm probably going to

play01:27

have to defer to the technical experts

play01:29

on those things but uh fire away any

play01:31

questions yeah go

play01:36

ahead don't be

play01:42

afraid so what are your Reflections on

play01:45

the Gemini art happening with the Gemini

play01:51

art yeah with okay I wasn't I expect to

play01:55

talk about but

play01:57

uh um you know we definitely messed up

play02:01

uh on the image generation um

play02:04

and

play02:06

um I think it mostly due to just like

play02:10

not thorough testing um and uh it

play02:14

definitely for good reasons upset a lot

play02:17

of

play02:18

people um uh on the images as you might

play02:21

have seen um I think the the images

play02:27

prompted a lot of people to really

play02:28

deeply test the base text models um and

play02:32

the text

play02:33

models

play02:35

um have two separate effects going uh

play02:39

you know one thing is just quite

play02:40

honestly if you deeply test any text

play02:42

model out there whether it's ours Chach

play02:44

grock what have you um it'll

play02:48

say you know some pretty weird things

play02:50

that are out there that uh you know

play02:52

definitely feel far left for example um

play02:55

and kind of any model if you try hard

play02:58

enough can be prompted to in that regime

play03:00

but also just to be fair um there's

play03:04

definitely work in that model um that

play03:07

once again we haven't fully understood

play03:10

uh why them's left in many cases um and

play03:13

that's not our intention uh but if you

play03:16

try it starting um over this last week

play03:19

it should be at least um 80% better of

play03:22

the test cases that we've covered um so

play03:25

I'd invite all of you to try it

play03:28

um this should be a big effect uh the

play03:32

bottle that you're chying the Gemini 1.5

play03:35

Pro which isn't in the sort of public

play03:38

facing app the thing we used to call

play03:40

Bart um shouldn't have much of that

play03:43

effect except for that General effect

play03:45

that if you sort of red team any AI

play03:48

model you're going to get weird Corner

play03:50

cases um but we're not um even though

play03:54

this one hasn't been sort of Thoroughly

play03:56

testing that way we don't expect it to

play03:58

have uh strong particular leadings um I

play04:01

suppose we can give it a go um though

play04:04

we're more excited today to try the long

play04:05

context and some of the technical

play04:07

features thank

play04:09

you correct yeah uh with all the recent

play04:13

developments modalities have you

play04:15

considered like a video chat

play04:17

GPT um a video chat GPT we probably

play04:21

wouldn't call it

play04:25

that

play04:27

um uh but uh no I mean

play04:31

multimodal both in and out is very

play04:34

exciting with video

play04:38

audio um I we' run early experiments

play04:42

um and uh I mean it's an exciting field

play04:46

even the little you guys remember the

play04:48

Duck video that kind of got us in

play04:49

trouble though to be fair it was fully

play04:52

disclaimed in the video that it wasn't

play04:54

real time but um but that is something

play04:58

that we've actually done is FedEd images

play05:02

and uh you know in like frame by frame

play05:06

and how to talk about it so um yeah

play05:09

that's super exciting I don't um I don't

play05:13

think we have anything like real time to

play05:15

present uh right now

play05:18

today yeah are you personally writing

play05:21

code for some projects um I have't

play05:24

actually write in code to be perfectly

play05:25

honest um it's not like code that you

play05:28

would be very impressed

play05:30

by um but yeah every once in a while

play05:35

just a little like kind of debugging or

play05:36

just trying to understand for myself um

play05:40

how a model works or um you know to just

play05:45

analyze the performance in this slightly

play05:46

a differently or something like that uh

play05:49

but little bits and pieces that make me

play05:50

feel connected it's once again I don't

play05:53

think you would be very technically

play05:54

impressed by it uh but it's nice to it's

play05:57

nice to be able to play with that and

play05:59

sometimes I'll use the AI Bots to write

play06:01

the code for me uh because I'm Rusty and

play06:03

they actually do a pretty good job uh so

play06:06

I'm very pleased about that okay

play06:08

question the

play06:10

back first yeah okay uh so prei

play06:14

simulation sorry pre AI the closest

play06:16

thing we got to simulators was game

play06:18

engines um what do you think the new

play06:20

advances in the field mean for us to

play06:22

create better games or game engines in

play06:24

general you have a view on that

play06:28

um

play06:31

sorry it wasn't like a sigh because of

play06:33

disapproval or anything um I think okay

play06:36

I mean I um what can I say about game

play06:41

engines I think obviously like on the

play06:44

graphics you can do new and interesting

play06:46

things with game engine but I think

play06:48

maybe the more interesting is the

play06:49

interaction with the other you know

play06:52

virtual players and things like that

play06:54

like whatever the characters are um I

play06:58

guess I guess these these days you know

play07:00

you can call people who are bland NPCs

play07:02

or whatever but in the future maybe NPCs

play07:05

will be actually very colorful and

play07:06

interesting yeah um so I think that's a

play07:11

really rich

play07:13

possibility uh probably not not enough

play07:16

of a gamer to think through all the

play07:18

possible Futures with AI but uh I it

play07:22

opens up any

play07:25

possibilities yeah what kind of like

play07:28

applications are you excited about

play07:30

people buing on on

play07:32

J yeah what kind of applications I'm

play07:35

most excited about um I mean I I think

play07:38

just

play07:39

ingesting uh right now for the version

play07:42

we're trying to tell you the you know

play07:43

1.5 Pro one context is something we're

play07:46

really experimenting with and whether

play07:50

you dump a ton of code in there or video

play07:54

I mean I've just seen people do I I

play07:57

don't think the model could do this to

play07:58

be perfectly honest like but people will

play08:00

like dump in their code and do a video

play08:03

of the app and say Here's the bug and

play08:06

the model will figure out where the bug

play08:07

is of the code which is kind of

play08:09

mind-blowing that that works at all I

play08:11

honestly don't really understand how

play08:13

model does that um but I'm not saying

play08:16

you should do exactly that thing uh but

play08:18

yeah experimenting with things that

play08:20

really require the long context um do we

play08:23

have the servers to support all these

play08:25

people here banging on it on we we have

play08:27

the people on the service here well okay

play08:30

my phone is buzzing everybody's really

play08:31

stressed out you guys

play08:34

Tes um because you know the million

play08:37

context queries do take a bit of comput

play08:39

in time but you should go for

play08:42

it yeah you mentioned a few times that

play08:46

you're not sure how this model works or

play08:48

you weren't sure that this could do the

play08:50

things that it does do you think we can

play08:52

reach a point where we actually

play08:54

understand how these models work or will

play08:55

they remain black boxes if we just trust

play08:58

the makers of model to not mess up um no

play09:02

I I I think you can learn to understand

play09:04

it I mean you know the fact is that when

play09:08

uh we train these things there are a

play09:11

thousand different capabilities you

play09:13

could try out so on the one hand it's

play09:17

very surprising that it can do it on the

play09:20

other hand if it's any particular one

play09:22

capability you can go back and you know

play09:25

we can look at uh where the attention is

play09:28

going at each layer between like the

play09:31

code and the video no we can't deeply

play09:34

analyze it um and personally done that I

play09:36

know how far along the researchers have

play09:38

gotten into doing that kind of thing

play09:41

but um you know it takes a huge amount

play09:44

of time and study to really slice apart

play09:47

a why model is able to do some things

play09:50

and honestly most of the time that I see

play09:52

slicing it's like why it's not doing

play09:54

something um so I guess I would say

play09:57

it's it's it's mostly because I I think

play10:01

we could understand it and people

play10:03

probably are uh but most of the effort

play10:07

is spent figuring out where it goes

play10:09

wrong not where it

play10:11

goes

play10:13

um uh yeah so in computer science

play10:16

there's this concept of reflective

play10:18

programming where like a program can

play10:21

look at its own source code maybe modify

play10:23

s source code and then in AGI literature

play10:26

there's like recursive self-improvements

play10:29

so what are your thoughts on the

play10:30

implications of extremely long context

play10:33

windows and a language model being able

play10:35

to modify its own prompts and what that

play10:38

has to do with like autonomy and

play10:40

building towards AGI

play10:42

potentially yeah I think it's very

play10:45

exciting to you know to have these

play10:47

things actually improve themselves um I

play10:50

remember when I was

play10:52

um I think in grad school I wrote this

play10:54

game where like it was like a wall NES

play10:57

you were flying through but when you

play10:58

shot the walls the walls corresponded to

play11:01

bits in memory and it would just like

play11:03

flip those bits and the goal is to crash

play11:05

it as quickly as possible um which

play11:08

doesn't really answer your question but

play11:10

that was an example of self- modifying

play11:13

code uh I guess not for a particularly

play11:16

useful purpose but I i' help people you

play11:18

know play that until the computer

play11:20

crashed anyhow on your positive example

play11:25

uh I see today people just using a talk

play11:28

about

play11:30

um I think you know open loop could it

play11:33

work for certain I think for certain

play11:35

very limited domains today like if you

play11:38

without the human intervention to guide

play11:40

it uh I bet it could actually do some

play11:44

kind of continued

play11:47

Improvement um but I don't think we're

play11:49

quite at the stage where

play11:51

for I don't know real serious things I

play11:54

first of all knowing context is not

play11:55

actually enough for big code bases uh to

play11:58

to to turn on the entire code basee uh

play12:01

but you could do like retrieval and then

play12:03

augmentation editing um I guess I

play12:06

haven't personally played with enough

play12:08

but I I haven't seen it be at the stage

play12:12

today where a complex sort of piece of

play12:17

code will just it or totally improve

play12:20

itself

play12:21

um but uh but but it's it's a great tool

play12:25

and like I said with human assistance we

play12:27

for sure do

play12:29

I mean like I will use Gemini to like

play12:32

try to do something with a Gemini code

play12:35

even today um but not

play12:39

very open loop deep sophisticated things

play12:42

I

play12:42

guess are you try let me get somebody in

play12:46

back just because yes

play12:48

yeah well you first and then the lady

play12:51

behind thank you um so I'm curious

play12:54

what's your take on some ultimate

play12:56

decision or plan at Le to raise 7

play13:01

trillion

play13:04

right I'm I'm just curious like how do

play13:06

you see that from

play13:08

obvious um you know look I saw the

play13:12

headline uh I didn't get too deep into

play13:14

it I assumed it was sort of a

play13:16

provocative headline or statement or

play13:18

something I don't know I don't know I he

play13:21

hasn't asked me for some

play13:25

trillion um I think it was it it was

play13:28

meant for like chip development or

play13:30

something like that I I don't I don't

play13:32

get I'm not an expert in chip

play13:35

development but I don't get the sense

play13:37

that it's just something you can like

play13:38

sort of pour money like even huge

play13:41

amounts of money in outcome

play13:43

chips I'm not an expert in the market

play13:46

though um let's see let me try somebody

play13:49

way in the back was there okay yes s

play13:52

yeah so we the training cost of model is

play13:57

so how we can

play13:59

oh the training cost of balls are super

play14:02

high um yeah the training costs are

play14:05

definitely High um

play14:09

and uh you know that's something

play14:11

companies like us have to cope

play14:14

with but I think you know the long-term

play14:18

utility is incomparably higher like if

play14:21

you kind of measure it on a human

play14:24

productivity level uh you know if it

play14:27

saves somebody an hour of work over the

play14:30

course of a week you know that hour is

play14:32

worth a lot you there are a lot of

play14:33

people using these things or will be

play14:35

using them um but you do it's a big bet

play14:39

on the future um cost less than $7

play14:42

trillion

play14:46

right what's your thoughts on model

play14:48

training on device model training on

play14:51

device on device oh model running on

play14:54

device uh yeah model running on device

play14:57

um we've shipped it to

play14:59

I think Android and chrome and yeah or

play15:02

pixel phones I think even Chrome runs a

play15:05

pretty decent model these days um we

play15:07

just open sourced Gema which was pretty

play15:09

small a couple billion parameters I

play15:11

can't remember right

play15:13

now yeah um yeah I mean it's

play15:17

really useful uh you know it can be low

play15:20

latency you're not dependent on

play15:22

connectivity and uh the small models can

play15:25

call bigger models in the cloud too so

play15:29

uh the H Theon device is a really good

play15:31

idea yeah yes um what are some vertical

play15:36

industry that you feel like this gen way

play15:39

going to have a big impact on and

play15:41

startups should consider hacking on

play15:43

those Industries which Industries do I

play15:46

think have a big

play15:48

opportunity I think it's just like very

play15:50

hard to predict I mean there are sort of

play15:52

the obvious industries that people think

play15:54

of sort of customer service or um kind

play15:58

of just like you know

play16:00

analyzing I don't know like different

play16:03

lengthy documents and kind of a workflow

play16:06

automation I guess those are obvious but

play16:08

um I think there are going to be non-

play16:10

obvious ones which which I can't predict

play16:13

uh especially as you look these sort of

play16:15

multimodal models and the surprising

play16:18

capabilities that they have I I feel

play16:20

like I mean that's why we have all view

play16:22

here you guys are the creative ones to

play16:24

figure that

play16:26

out okay sir hello my name is Alex we

play16:29

run dos of thousands of customer service

play16:32

chats every day and on lmms andly G4 was

play16:35

the only thing to really work and now it

play16:37

seems that jimmi is another thing that

play16:40

really works thank you so much for this

play16:42

and it

play16:45

thank like it's way more chip while it

play16:48

works even better sometimes so the

play16:50

question is will it stay same chip or

play16:53

are you just planning to raise prices at

play16:56

some point or who knows

play17:00

um we're not I I I'm actually not on top

play17:04

for the pricing thing I don't expect

play17:06

that we will raise prices however

play17:09

because uh I mean there are

play17:11

fundamentally a couple of Trends one is

play17:15

just

play17:16

that these um you know there's just

play17:19

optimizations and things around

play17:21

inference that are just constantly like

play17:23

all the time some says I have this 10%

play17:25

idea this 20% idea and like after month

play17:29

that adds up um I think our tpus are

play17:33

actually pretty damn good um at um

play17:35

inferencing and not the thinging the

play17:37

gpus but um but for certain inference

play17:40

workloads they're just configured really

play17:43

nicely uh and the other big effect is

play17:46

actually we're able to make smaller

play17:50

models more and more effective just with

play17:52

new generations just whatever

play17:54

architectural changes training changes

play17:56

um all kinds of things like like that U

play17:59

so the models are getting more powerful

play18:01

even at the same time of size so I would

play18:03

not expect prices to go up yes maam um

play18:07

what are your prediction for how AI is

play18:10

going to impact healthare and biotic and

play18:12

some things you're excited about that oh

play18:14

AI Healthcare and

play18:16

biotech um well I I think there are a

play18:19

couple very you know different ways you

play18:22

know on the biotech side people look at

play18:25

um things like Alpha fold and things

play18:28

like that just like understanding the

play18:31

fundamental mechanics of life and I

play18:34

think you'll see AI do more and more of

play18:36

that whether it's actual physical

play18:40

um molecule kind of bonding things or

play18:44

reading and summarizing journal articles

play18:46

things like that um I also think for

play18:49

patients and this is kind of a tough

play18:51

area honestly uh because we're

play18:53

definitely not prepared for our just AI

play18:56

like go ahead ask that any question like

play18:58

like we're not you know AI make mistakes

play19:00

and see things like that but I

play19:03

think there's a future when you if you

play19:05

can overcome those kinds of issues where

play19:08

an AI can much more deeply spend time on

play19:12

an individual person and their history

play19:15

and all their

play19:16

scans uh maybe mediated by a doctor or

play19:19

something but actually give you just

play19:21

better diagnoses better recommendations

play19:23

things like

play19:25

that and are you focusing on any other

play19:29

non- transformer architectures for like

play19:31

reasoning planning or any of to get

play19:33

better at their okay question are we

play19:35

focusing on any non- Transformer

play19:39

architectures I mean I think there's

play19:41

like so many sort of uh variations um

play19:45

but I guess most people would argue are

play19:47

still kind of Transformer based um I

play19:50

mean I'm sure somebody in the company

play19:52

there speak more to it uh would be

play19:55

looking but uh

play19:58

yeah as much progress as Transformers

play20:00

have made over the last whatever six

play20:03

seven seven eight years I guess

play20:07

um there's you know there's nothing to

play20:11

say there's not going to be some new

play20:12

revolutionary uh architecture and it's

play20:15

also possible that just you know

play20:17

incremental changes for example sparcity

play20:20

and things like that um that are still

play20:23

kind of same Transformer also bring

play20:26

revolutions to I I don't magic cancer

play20:29

but is there some bottleneck for like

play20:31

reasoning kind of questions bottleneck

play20:34

using this Transformers um I mean

play20:36

there's been lots of theoretical work

play20:38

showing the limitations of Transformers

play20:40

and you know can't do this kind of thing

play20:43

this many layers and things like that

play20:47

um I I I don't know how to extrapolate

play20:50

that to

play20:51

like contemporary Transformers that

play20:54

usually don't meet the assumptions of

play20:55

the theoretical Works um so may not

play20:58

apply

play21:00

but um I probably HED my butts and try

play21:04

other architectures all being

play21:06

equol thank

play21:11

[Laughter]

play21:21

you Google has a Google Glass but now

play21:23

Apple has Vision Pro um I I think Google

play21:26

Glass may be a little bit early would

play21:28

because they like try that another shop

play21:31

um yeah like I messed up Google

play21:36

Glass no no but I I I feel like I made

play21:38

some bad decision it yeah it was for

play21:40

sure early and early in two senses of

play21:42

the word uh maybe early in the overall

play21:46

evolution of Technology but also I think

play21:50

I like in hindsight I tried to push it

play21:53

as a product when it itself was sort of

play21:55

more of a prototype and I should have

play21:56

set those expectations around it um and

play22:00

I personally didn't know much about sort

play22:01

of consumer harder Supply chains back

play22:03

then and a bunch of things I wish i' had

play22:07

done differently um but I personally am

play22:10

still a fan of kind of the lightweight

play22:13

kind of minimal display that that

play22:14

offered that you could just like wear

play22:16

all day versus the big heavy things that

play22:18

we have today um that's my personal

play22:20

preference uh but the the Apple vision

play22:23

and the Oculus for the matter they're

play22:25

very impressive like having played with

play22:27

them

play22:28

um I mean I'm just impressed with what

play22:30

you can have in front of your screen but

play22:32

that wasn't what I was personally going

play22:36

for back then um yes ma um so do you see

play22:41

um Gemini expanding capabilities into

play22:43

like 3D head down the line of spatial

play22:46

Computing in general uh or simulation of

play22:48

the world in general of that and

play22:50

especially Beyond Google who already

play22:53

have several product that's really in

play22:54

the area right like Google MTH stet view

play22:56

air core all of that do you see all of

play22:58

those have some synergies between them

play23:01

wow that's a good question to be honest

play23:03

I haven't thought about it but now that

play23:05

you say it yeah there's no reason we

play23:09

can you know put in more sort of three

play23:11

like it's kind of another mode you know

play23:13

3D data um so probably something

play23:18

interesting would happen I mean I I

play23:19

don't see why you

play23:21

wouldn't uh try to put that into a model

play23:24

that's already I've got all the smarts

play23:25

of the text model now can turn on

play23:28

something else

play23:29

too and by the way maybe somebody's

play23:31

doing it at Gemini I don't know oh

play23:35

yes I'm noty to it or I forgot about

play23:39

doesn't stop happen okay yes question

play23:41

the back there are you optimistic that

play23:43

we be able to re in text generating

play23:46

models ability to hallucinate and what

play23:48

do you think about the ethical issue of

play23:50

potentially

play23:51

spreading uh problem right now um no

play23:55

question about it um I mean we have made

play23:59

them hallucinate less and less over

play24:02

time

play24:04

um but I would definitely be excited to

play24:07

see a breakthrough that brings it to

play24:09

near zero um I don't you know that's not

play24:12

you can't just like count on

play24:14

breakthroughs um so I think we're going

play24:15

to keep going the incremental kinds of

play24:18

things that we do to just like bring all

play24:20

the H stations down down down over time

play24:23

like I said I think breakthrough would

play24:24

be good um misinformation

play24:28

you know misinformation is a complicated

play24:30

issue I think um I mean obviously you

play24:33

don't want your AI Bots to be just like

play24:36

making stuff up um

play24:39

but they can also be kind of tricked

play24:44

into like I me there's a lot of I

play24:48

guess complicated uh

play24:53

political issues in terms of what people

play24:56

consider what different people consider

play24:58

misinformation versus not and it gets

play25:00

into kind of a broad social debate

play25:05

um I suppose another thing you could

play25:08

consider is about them sort of

play25:10

deliberately generating this information

play25:12

on the behalf of another actor

play25:15

um from that point of view I mean

play25:17

unfortunately it's like it's very easy

play25:20

to make a lousy AI um like one that

play25:24

hallucinates a lot um and you can make

play25:28

you know any open source text model and

play25:31

probably tweak it to generate

play25:34

misinformation of all kinds and if

play25:36

you're not concerned about

play25:39

um um you know the accuracy it's just

play25:42

like kind of an easy thing to do so I

play25:45

don't know I I guess now I think about I

play25:47

mean detecting AI generated content is

play25:49

an important field and something that we

play25:51

work on and so forth so you at least can

play25:54

maybe tell if something coming at you

play25:56

was AI generated

play26:00

yeah Alexandra so the CEO of Nvidia said

play26:05

that basically the future of writing

play26:08

code as a career is

play26:11

[Laughter]

play26:18

dead okay

play26:20

yeah um I mean it's

play26:24

um like we don't know where the future

play26:27

AI is going broadly I wouldn't you know

play26:31

we don't know you know it seems to help

play26:33

across a range of many careers whether

play26:36

it's graphic artists or customer support

play26:39

or doctors or um or Executives or you

play26:44

know what have you um I mean so I don't

play26:48

know that I would be like singling

play26:51

out um programming in particular um it's

play26:54

it's actually probably one of the more

play26:56

challenging tasks for an llm

play27:00

today but if you're talking about for

play27:03

you know decades in the future what

play27:05

should you be kind of preparing for and

play27:07

so forth I mean it's it's hard to say I

play27:10

mean the AI could get quite good

play27:12

programming but you can say that about

play27:14

kind of any field of human

play27:16

endeavor so I guess I probably wouldn't

play27:19

have singled that out as like saying

play27:21

don't study specifically programming um

play27:25

I don't know if that's

play27:30

answer okay hand the

play27:32

back lot people start these agents to

play27:34

write I'm wondering how that's going to

play27:36

impact it security You could argue that

play27:39

like the code might become worse or like

play27:41

less check for certain issues or you

play27:43

could argue that like we're get better

play27:45

at invting test Suites which cover all

play27:47

the cases what are your opinions on this

play27:49

like is maybe for the outage programmer

play27:51

like it security way to go because like

play27:54

the code is going to be Britten but

play27:55

someone still needs to check it

play27:58

oh wow you guys are all trying to choose

play28:00

career based

play28:04

on I don't know I think you should use a

play28:06

fortune teller for that

play28:08

General line of

play28:12

questions but I I do think

play28:16

that you know using an AI today to write

play28:20

let's say unit tests is pretty

play28:22

straightforward yeah like that's the

play28:24

kind of thing the AI does really quite

play28:26

well

play28:27

uh so I guess my hope is

play28:31

that AI will make code more secure not

play28:34

less secure I mean it's kind of it's

play28:38

usually insecurity is to some extent the

play28:41

effect of people being lazy and the one

play28:43

thing that AI is kind of good at is you

play28:45

know not being

play28:47

lazy so if I had to bet I would say

play28:50

there's probably a net benefit to

play28:52

security with

play28:54

AI um but I wouldn't discourage you from

play28:58

pursuing clear and I Security based on

play29:00

that I think pretty

play29:03

much um

play29:05

okay do you want to build AGI do I

play29:10

want

play29:12

yeah yeah yeah I mean I think it's um

play29:15

you know different people mean different

play29:16

things about that but uh to me the

play29:19

reasoning aspects are really exciting

play29:21

and amazing and um you know I kind of

play29:25

came out of retirement just because of

play29:27

the Vector of AI is so exciting and as

play29:30

computer scientists just seeing what

play29:31

these models can do year after year it

play29:34

is astonishing so yes any efforts on

play29:38

like humanoid robotics or these because

play29:40

there was so much progress in Google X

play29:42

like in 2015 16 oh humanoid robotics um

play29:46

boy we've done a lot of humanoid

play29:49

robotics over the years and sort of

play29:50

required and they sold a bunch of

play29:52

companies humanoid robotics um and now

play29:56

there are bilian sorry not there quite a

play29:59

few companies doing humanoid Robotics

play30:02

and internally we still have groups that

play30:04

work on Robotics and

play30:06

varying uh varying forms so what are my

play30:09

thoughts about that I don't know you

play30:12

know in general I worked on X prior to

play30:16

this sort of new AI wave and that there

play30:19

the focus was more Hardware projects for

play30:21

sure uh but honestly I guess I found a

play30:26

the hard way open

play30:27

Hardware is much more difficult um kind

play30:31

of on a technical basis on business

play30:33

basis and every way so I'm not

play30:35

discouraging people from doing it we

play30:37

need people for sure to do it um at the

play30:40

same time while the software and the AIS

play30:44

are getting so much faster at such a

play30:47

high

play30:49

rate I guess to me that feels like

play30:52

that's kind of the rocket ship um and I

play30:57

feel like if I get distracted in a way

play31:00

by making hardware for today's

play31:03

AIS um that might not be the best of use

play31:06

times compared

play31:07

to what is the next kind of level of AI

play31:10

going to be able to support and for that

play31:12

matter we'll it design a robot for

play31:15

me that's my person there are a bunch of

play31:17

people at the Google and alphabet who

play31:20

can work on hard yes thanks uhing

play31:24

advertising R is really important for's

play31:26

your how advertising will be

play31:30

disrupted way the question about

play31:34

advertising yeah I um of all people not

play31:38

too terribly concerned about business

play31:40

model shifts I mean I think it's a

play31:41

little bit

play31:45

um I think it's wonderful that we've

play31:47

been now for 25 years or whatever uh

play31:51

able to give just world class

play31:54

information um search uh for free to

play31:58

everyone and that's supported by

play32:00

advertising um which in my mind is great

play32:03

it's great for the world you know uh

play32:06

well you know kid in Africa has just as

play32:08

much access to basic information as the

play32:10

President of the United States or what

play32:12

have you um so that's good um at the

play32:17

same time I expect business models are

play32:20

going to evolve over time and uh and

play32:25

maybe they'll still be advertising

play32:26

because whatever the advertising kind of

play32:28

works better um the AI is able to tailor

play32:31

it better or you like it but even if it

play32:34

happens to move to you know now we have

play32:36

u g Advanced um other companies have

play32:39

their you know paid models I think the

play32:43

fundamental issues that you're

play32:45

delivering a huge amount of value you

play32:47

know displacing all the mental effort

play32:50

that would have been you know required

play32:53

to take the place of that AI um whether

play32:56

in your time or labor or what have you

play32:58

is enormous um and the same thing was

play33:01

true in search so I personally feel as

play33:04

long as there's huge value being

play33:06

generated we'll figure out the business

play33:12

models

play33:16

C

play33:18

model third party cookie you know I'm

play33:21

going

play33:22

to how naive I am about to detail I mean

play33:25

I vaguely am aware uh of that stuff but

play33:28

I don't um I know I can't think of how

play33:33

those things interact I'm s oh okay well

play33:36

you maybe you should answer the

play33:40

question um how many more do you want

play33:43

two okay two more questions where do you

play33:46

see Google search going where do I see

play33:48

Google search going well it's a super

play33:50

exciting time uh for search because your

play33:53

ability

play33:55

to answer questions will AI is just so

play33:57

much greater um I think it's the bigger

play34:04

opportunity is in

play34:06

situations where you are uh recall

play34:10

limited uh more so like you might ask a

play34:13

very specialized question or it's

play34:15

related to your own personal situation

play34:18

in a way that nobody out there you know

play34:20

on the internet has already written

play34:22

about you know for the questions that a

play34:24

million people have written about

play34:25

already and thought deep ple about it's

play34:28

probably not as big a deal but the

play34:30

things that are very specific to what

play34:32

you know you might care about right now

play34:34

in a particular way that's a huge

play34:36

opportunity and um you know you can

play34:40

imagine all kinds of products in your

play34:42

eyes and different ways to deliver that

play34:45

uh but basically AI is the nebor are

play34:47

just doing much better job in that case

play34:50

okay last question okay who's going to

play34:52

get the last question is it a good one

play34:54

who's got a good one in the back you

play34:55

have to be conf

play35:10

so for more

play35:13

what morality mortality

play35:17

[Laughter]

play35:25

oh

play35:27

look I I I'm probably not as well versed

play35:29

as all of you are to be honest but uh

play35:33

I've definitely

play35:34

seen the uh kind of the the molecular AI

play35:38

make huge amounts of

play35:40

progress um you could imagine that there

play35:43

would also be a lot of progress maybe

play35:45

haven't seen yet on the epidemiology

play35:47

side of things to just be able to get

play35:49

kind of I don't know more honest better

play35:52

controlled a kind of

play35:55

broader understanding what's happening

play35:57

to People's Health around the world um

play36:02

but yeah what can good answer on the

play36:05

last one um I don't know I don't have

play36:08

like a really brilliant immortality key

play36:10

byi just like that but uh you know it's

play36:14

the kind of field that for

play36:16

sure benefits from AI whether you're a

play36:20

researcher or like you know I want it to

play36:23

just summarize articles to me that one

play36:26

uh but in the future you know I would

play36:29

expect the AI would actually give you

play36:31

novel hypotheses to test uh it does that

play36:34

today with the alpha folds of the world

play36:36

but maybe in more complex systems than

play36:40

just

play36:41

molecules okay amazing thank you thank

play36:47

[Applause]

play36:51

you yeah I think uh really hum to have

play36:55

you here