The Future of Generative AI Agents with Joon Sung Park

Foundation Capital
20 Feb 202448:25

Summary

TLDRIn a discussion about AI agents, Jun Zen Park provides background on the evolution of agents, from early assistive agents like Clippy to modern conversational agents. He outlines two branches of agent development: tool-based agents meant to automate complex tasks, and simulation agents that mimic human behavior. Large language models enable more advanced, personalized agents, though interaction challenges remain around deploying agents for high-risk tasks. Park sees initial success for agents in soft-edge problem spaces like games and entertainment before expanding to other areas. Though applications like ChatGPT show promise, he questions if conversational agents are the ultimate killer app compared to historic examples like Microsoft Excel.

Takeaways

  • ๐Ÿ˜€ LLMs like GPT-3 made AI agents possible by providing the ability to predict reasonable next actions given a context
  • ๐Ÿ‘ฅ There are two main types of AI agents - tool-based agents to automate tasks, and simulation agents to model human behavior
  • ๐Ÿ’ก LLMs still need additional components like long-term memory and planning for full agent capabilities
  • ๐ŸŽฎ Games were an inspiration for early agent research aiming to create human-like NPCs
  • ๐Ÿšฆ Current LLM limitations around safety and fine-tuning may limit the range of possible agent behaviors
  • ๐ŸŽญ Simulation agents for 'soft edge' problems like games and entertainment may succeed sooner than tool agents
  • ๐Ÿ”ฎ Multimodal (text + image) agents are an exciting area for future research
  • โ“ It's unclear if ChatGPT represents the 'killer application' for LLMs we expected
  • ๐Ÿ“š Agent hype cycles have spiked and faded as expectations exceeded capabilities
  • ๐Ÿค” Carefully considering human-agent interaction and usage costs will be key to adoption

Q & A

  • What was the initial motivation for Jun to research generative agents?

    -Jun was motivated by the question of what new and unique interactions large language models like GPT-3 would enable. He wanted to explore if these models could be used to generate believable human behavior and agents when given a micro context.

  • How does Jun define 'tool-based' agents versus 'simulation' agents?

    -Tool-based agents are designed to automate complex tasks like buying plane tickets or ordering pizza. Simulation agents are used to populate game worlds or simulations, focusing more on replicating human behavior and relationships.

  • What capability did large language models add that enabled new progress in building agents?

    -Large language models provided the ability to predict reasonable next sequences given a micro context or moment. This could replace manually scripting all possible agent behaviors.

  • What does Jun see as a current limitation in using models like GPT-3 for simulation agents?

    -Models like GPT-3 have been specifically fine-tuned to remain safe and not surface unsafe content. This limits their ability to accurately reflect a full range of human experiences like conflict.

  • Where does Jun expect agent technologies to first succeed commercially in the next few years?

    -Jun expects agent technologies to first succeed commercially in 'soft edge' problem spaces over the next few years, like simulations and games. There is more tolerance for failure in these areas.

  • What does Jun see as a key open question around why previous periods of hype around agents failed?

    -Jun wonders if past agent hype cycles failed because not enough thought was given to interactions - how agents would actually be used and whether they solved needs users really had.

  • What future line of questioning around large language models is Jun interested in pursuing?

    -Jun wonders if ChatGPT represents the 'killer app' for large language models that people were waiting for. He thinks it's worth discussing whether ChatGPT is actually as transformational as expected.

  • How does Jun suggest thinking about future model architectures that could replace Transformers?

    -Jun suggests thinking about Transformer capabilities as an abstraction layer - focusing on the reasoning capacity it provides. The implementation could be replaced over 5-10 years while still building useful applications today.

  • Where does Jun look for inspiration on new research directions?

    -Jun looks to foundational insights from early Artificial Intelligence researchers that have stood the test of time. He believes great ideas are timeless even if hype cycles come and go.

  • What aspect of current agent capabilities is Jun most interested in improving further?

    -Jun is interested in enhancing accuracy to better reflect real human behavior and diversity. This could enable personalized and scalable simulations grounded in real communities.

Outlines

00:00

๐Ÿ‘จโ€๐Ÿ’ป Jun introduces himself and his work on generative agents

Jun provides background on himself, stating he is a PhD student working on human-computer interaction and natural language processing. He discusses his interest in exploring how large language models can enable new capabilities, which led him to work on a project called Generative Agents. This involves using language models to create AI agents that can populate and behave realistically within simulation worlds.

05:01

๐Ÿค” What motivated Jun's research direction and focus on agents

Jun explains his thought process in identifying a research direction. When large language models like GPT-3 emerged, it was unclear what novel capabilities they would truly enable. Simple tasks like classification weren't fundamentally new. Simulating human behavior via agents was ambitious, tied to longstanding AI goals, and could open new interaction possibilities.

10:01

๐Ÿ˜Ž The evolution of agents and how they split into two communities

Jun provides historical context on agents. There have been tool-based agents focused on automation vs simulation agents focused on modeling human communities. With large language models, these have splitted into two present-day communities - one using tools and one focused on personalized, multi-agent simulation.

15:01

๐ŸŒŸ How LLMs enabled a breakthrough in quality and scale for agents

Previously, generating sophisticated agent behaviors required manual scripting which didn't scale. Large language models provide the ability to predict reasonable next actions given a situation context. By adding memory and planning, this provides a scalable path to coherent, long-term agent behavior.

20:04

๐ŸŽฅ Multimodal Inputs (Image, Video) will make the next breakthrough

Currently agents operate via text from visual inputs translated to language descriptions. With multimodal models like DALL-E expanding to handle images, video, etc directly, future agents could take raw perceptual inputs and better simulate human perspective and behavior.

25:06

๐ŸŽฏOpportunity for more accurate agents that model actual communities

Jun sees opportunity to evolve agents to accurately reflect real human social dynamics instead of fictional scenarios. This is challenging today due to content moderation but may be feasible over time. Accurate community simulation could enable applications in markets, personalization, policy, and beyond.

30:08

๐Ÿ˜… Deployments today constrained to "soft failure" cases, long road ahead

Agents face adoption challenges when mistakes incur high costs, as in purchasing transactions. History shows hype cycles then fading interest in 6-12 months. The first deployments are likely in entertainment, gaming etc. Standards for auditability, controllability needed before handling complex real-world tasks.

35:11

๐Ÿ”ฎ Don't assume Agents will work now just because technology improved

Despite progress, it is worth questioning why past agent hype cycles failed. Just having greater technological capacity doesn't guarantee actual usefulness and adoption. Key questions remain around real user need, interaction model fit, and compelling benefit over cost.

40:13

โ“Is ChatGPT actually the "killer app" for LLMs we've been waiting for?

The breadth of ChatGPT adoption seems incredible. However, it is largely an interface wrapper on existing skills. We should question whether it truly constitutes the "killer app" that maximizes impact of LLMs. If not, what applications are still missing that could better deliver generalized value?

45:14

๐Ÿ“š Historical inspiration from Simon, Newell as timeless resources

Instead of recent papers, Jun is inspired by timeless foundational works, like Herbert Simon and Allen Newell. Their insights launched fields and won accolades. Unlike hype cycles, their textbooks showcase ideas that stood test of time and continue enabling impact decades later.

Mindmap

Keywords

๐Ÿ’กagents

The video focuses heavily on 'agents' - AI systems designed to execute tasks and interact with humans. Examples from the transcript show agents being used for automation, simulations, games and more. The evolution of agent design and capabilities is a central theme.

๐Ÿ’กlarge language models (LLMs)

LLMs like GPT-3 are presented as a key enabling technology for the current explosion in agent development. The models provide the natural language processing capacity to power human-like interactions.

๐Ÿ’กgenerative agents

The guest's research on 'generative agents' is highlighted - using LLMs to automatically generate simulations populated by human-like NPCs (non-player characters). This demonstrates a key new application area for LLMs.

๐Ÿ’กmulti-agent systems

The potential to combine multiple specialized agents together into collaborative 'multi-agent systems' is discussed. This could enable more complex automations.

๐Ÿ’กkiller application

The speakers explore whether chatbots like ChatGPT represent the 'killer application' for LLMs that will drive mass adoption. Or if there are even bigger opportunities still to be unlocked.

๐Ÿ’กlimitations

Current limitations around accuracy, safety and scalability of agents are covered. For example, the over-cautious nature of ChatGPT. And the challenges of deploying agents in high-risk real world scenarios.

๐Ÿ’กcommercial deployment

Discussion on the most viable initial areas for commercial agent deployment in enterprises. With a focus on use cases that allow for a higher tolerance of mistakes during the current limitations.

๐Ÿ’กfuture research

The guest mentions key avenues for future academic research on agents. Including increasing accuracy through personalization and scaling up the number of agents in simulations.

๐Ÿ’กinteraction challenges

A recurring theme is the historical failure of agents to deliver on hype because of unsolved interaction challenges. Despite advanced technology, if applications are not intuitive and valuable for end users they will not succeed.

๐Ÿ’กmultimodality

The potential to leverage multi-modal LLMs that can process images and video as well as text is discussed. This could significantly enrich the perception and situational awareness of agents.

Highlights

Agents have finally made their way into real enterprises with real use cases

Large language models provided the key capability to generate believable human behavior

Agents can be categorized as tool-based or simulation-based

Tool-based agents aim to automate complex tasks while simulation agents model communities

The agent architecture gives agents long-term memory and planning abilities

Multimodal capabilities like images will make agents more powerful

Accuracy limitations due to safety constraints may need to be addressed

Agents will likely succeed first in soft-edge problem spaces before hard-edge ones

The interaction challenges, not the technology, caused past agent hype cycles to fail

It's worth questioning if ChatGPT is the killer app we've been waiting for

The killer app should enable manipulating the key data type the technology generates

Learn from past insights that had impact and stood the test of time

Review recently published cutting-edge papers for the latest developments

Refer to pioneering works in AI and cognitive science for foundational ideas

Current hype cycles shouldn't discount timeless, foundational concepts

Focus on capacities and modalities over specific underlying technologies

Transcripts

play00:00

the number of users who use CH gbt is

play00:02

that's

play00:03

incredible but I think it's sort of

play00:05

worth asking ourselves is that sort of

play00:08

quote unquote the killer applications

play00:11

that we were waiting for chpt does feel

play00:14

like it's a fairly simple wrap around

play00:16

lar sling model because that's what the

play00:18

main is open AI has done fantastic

play00:21

things to make it safer and make it more

play00:22

useful F tuning I think what's really

play00:24

great but I think it's worth asking if

play00:27

that is actually the killer application

play00:30

why is it a killer application and the

play00:32

answer might actually come out that

play00:34

maybe it actually isn't the killer

play00:36

application that we were waiting for um

play00:39

in which case what is going to be the

play00:40

killer application that's really going

play00:42

to add value in a much more

play00:44

generalizable

play00:47

way welcome to AI in the real world I'm

play00:50

your host my name is Joan Chen and I am

play00:53

a general partner at Foundation capital

play00:55

I work closely with startups are

play00:56

reshaping business with AI in this

play00:59

series I'll be holding in-depth

play01:01

discussions with leading AI researchers

play01:03

we'll explore how St of-the-art AI

play01:05

models are being applied in real

play01:07

Enterprises today to kick things off I'm

play01:10

excited to speak with Jun Zen Park a phc

play01:12

student in computer science at Stanford

play01:14

June works at the intersection of human

play01:16

computer interaction and natural

play01:18

language processing he is best known for

play01:21

his research on AI agents we break down

play01:24

how AI is transforming agent design

play01:26

share advice for Builders working with

play01:28

these models and unpack why we haven't

play01:30

yet found the perfect killer app for AI

play01:33

agents here's our conversation how are

play01:36

you good what about you great seeing you

play01:38

again good to see you again it's been a

play01:39

while because the unconference was last

play01:42

I want to say May June May or June yeah

play01:47

last June May or June wow TimeWise and

play01:51

the world has changed I think that

play01:54

agents have finally made its way into

play01:58

real Enterprises with real news cases

play02:01

and it was not um back then it was a lot

play02:04

of like what what could this be right um

play02:07

right thanks to you and some of your

play02:08

work which is why I'm super excited to

play02:11

have this

play02:12

conversation together um especially

play02:15

right now since like uh Enterprises are

play02:19

um uh like are are sinking in a real way

play02:22

to adopt so uh so I thought who can can

play02:26

we chat with that would have like a

play02:28

really interesting perspective and

play02:30

that's why we reached out back to you so

play02:32

really appreciate the time of course

play02:35

thanks for having me do you mind maybe

play02:38

just to start like getting giving giving

play02:40

us a quick rundown of like what's

play02:43

happened maybe some of the background

play02:44

that you have uh building this

play02:47

technology yeah so let's see uh so do

play02:52

you want me to just sort of speak about

play02:54

sort of what has happened in the past

play02:55

six months where sort of what would be

play02:56

interested uh interested for you just

play02:59

just a brief overview of what youve

play03:00

worked on and also what's happened in

play03:02

the last uh six months to a year in

play03:05

terms of evolution yeah all right that

play03:07

sounds good right so um I guess I do a

play03:11

quick intro so I'm sort of I'm a PhD

play03:14

student here uh sort of working the area

play03:16

of HCI and and NP so as you know sort of

play03:20

the work that we've done I think the one

play03:22

that I'm sort of mainly known for is

play03:23

this paper code generative agents um and

play03:27

generative agents in particular was a

play03:29

project that to ask can we use our

play03:32

language models to create General agents

play03:34

that can populate a simulation world

play03:37

right so if you play something like Sim

play03:38

City or Sims uh can we actually create

play03:41

these NPC like characters that would

play03:42

actually flood into the City and

play03:44

actually live like humans and by

play03:47

definition it is sort of everything from

play03:49

how they would wake up in the morning

play03:52

talk to each other form routines and

play03:54

relationships all the way to creating

play03:57

basically communities and ersing social

play03:59

dynamic

play04:00

Dynamics and sort of my interest in this

play04:03

area really stems from this idea so this

play04:05

is sort of what people at the

play04:07

intersection of human computer

play04:08

interaction and natural language

play04:09

processing and machine learning like to

play04:10

ask which is we now have these really

play04:13

amazing models like large language

play04:15

models and Foundation models the

play04:17

question really becomes what are you

play04:19

going to do with them right these models

play04:21

are new and they're great and we think

play04:23

they have great capacity but are they

play04:26

really going to enable us to do

play04:28

something that's quite and unique and

play04:31

that has been sort of the focal point

play04:32

for a lot of the research that I do and

play04:35

ultimately the conversation that we got

play04:37

down to was this idea of well these

play04:40

models of train on Broad data like uh

play04:42

like the web Wikipedia and so forth so

play04:45

they can actually be used to generate a

play04:47

lot of believable human behavior when

play04:49

you're given a very micro context so can

play04:52

we actually piece this together to

play04:53

create human leg agents which is

play04:55

something that AI more broadly has

play04:58

envisioned since its founding day

play05:00

um and we decided that this is the time

play05:02

to do that and so that's how we got to

play05:05

where we are uh so that's the general of

play05:07

Agents uh and this is the paper that was

play05:10

published in April last year or we put

play05:12

it on archive in April and was

play05:14

officially published November which is

play05:17

which is crazy how much um uh the world

play05:20

has has developed I'm curious what

play05:22

initially

play05:23

motivated uh this topic for you I'm sure

play05:26

you had lots of different options in

play05:28

terms of what to research and study why

play05:31

did you decide to focus on

play05:33

this yeah so ultimately it really was

play05:37

the question of what will L language

play05:40

models these new models that are being

play05:43

trained really going to enable us to do

play05:45

and when I started my PhD was around

play05:47

like 2020 and that was when gpt3 was

play05:50

just about to come out during my first

play05:52

year we wrote a paper called uh

play05:54

Foundation

play05:55

models uh which sort of made this

play05:57

observation that there's going to be

play05:58

this new wave of models that's going to

play06:00

come out where we're not going to be

play06:02

training these models for a specific

play06:04

task but rather we'll be training for a

play06:06

modality right we're going to be

play06:08

training this language model that can

play06:10

process language and so

play06:12

forth

play06:14

and we thought that was going to be a

play06:16

big opportunity there in terms of what

play06:18

we can do with them but the question of

play06:20

what are we going to do with them was

play06:21

incredibly unclear so really our first

play06:24

instinct as sort of researchers in more

play06:27

machine learning in NLP community where

play06:29

we sort of were drawn to was this idea

play06:31

of can we do classifications or

play06:33

generations with these models and seeing

play06:36

that these models could do that was

play06:37

really exciting because we didn't train

play06:39

these models to do that but they could

play06:41

but more from the interaction

play06:43

perspective doing classification and

play06:45

simple generation was something that we

play06:47

already knew how to do so that didn't

play06:49

feel fundamentally new so really the

play06:52

question again became what are we going

play06:54

to do that's going to be truly new and

play06:57

transformative in the sense of

play06:59

interaction

play07:01

so that's uh what really Drew us to look

play07:03

for these kind of ideas um and again

play07:06

that's we thought simulating human

play07:09

behavior in general computational agents

play07:12

that felt like a big problem because in

play07:14

part because it's something that again

play07:16

our community had wanted for many

play07:18

decades uh it was sort of the idea that

play07:21

people in more the cognitive science

play07:24

field that will inspire the early AI

play07:26

research like alen new and Hart's these

play07:29

folks were asking and we were certainly

play07:33

inspired by those ideas and of course we

play07:36

thought it would be a lot of fun because

play07:37

we sort of grew up with Sims Pokemon and

play07:40

these kind of games in the 90s and early

play07:43

2000s and we were certainly inspired by

play07:45

those games as well I love I love those

play07:47

games as well and it's nice to see some

play07:49

of that um play out in the real world

play07:52

now I agree I think games are fun in the

play07:55

sense that um like you know I think they

play07:57

are inspirational in many ways because

play07:59

they do they very forward looking in

play08:00

many ways right because you can be a

play08:02

little bit more playful and I think

play08:04

research can be in many ways playful

play08:07

especially when you're trying to do

play08:08

really forward looking research so it

play08:10

certainly is a big inspiration and I was

play08:13

just going to sort of end that comment

play08:15

by saying that I think it's worth asking

play08:17

for us as a community what's going to be

play08:21

the new sort of quote unquote cure

play08:23

application of these

play08:25

models in the sense that um when we had

play08:28

personal computer in the early 80s and

play08:31

so forth the computers were very cool uh

play08:35

but what really made them into household

play08:39

applications were the existence of this

play08:41

what we would now consider as killer

play08:44

application of PCS like Microsoft Excel

play08:47

that really made tabular information uh

play08:50

usable and scalable I think we l

play08:53

language model Community are should also

play08:55

be looking for those kind of ideas as

play08:56

well because that's going to be

play08:58

ultimately what's going to really

play08:59

transform the user experience around

play09:01

these models and I think we're seeing

play09:03

some great uses uh usage of these models

play09:06

but I think there's a lot more to do

play09:08

going forward makes makes a lot of sense

play09:11

when you look at the what's happened

play09:13

since April right a lot of things have

play09:16

changed uh we have new LM capabilities

play09:20

we have a whole flurry of startups

play09:23

building in the space could you maybe

play09:25

summarize what what you've

play09:27

seen right

play09:30

so right so agan cty has been a big

play09:32

thing uh especially first the latter

play09:35

half of

play09:37

2023 this is how I'm seeing it um agent

play09:40

community in the sort of the way I view

play09:43

it has split into two communities I

play09:45

would argue now

play09:47

so maybe it might actually Mak a little

play09:50

more sense to really talk about like the

play09:52

history of Agents because Asian became a

play09:54

big thing last year but this is not a

play09:56

new idea in and out of itself right in

play09:59

this in even even in the commercial

play10:01

space we actually had agents like

play10:03

Microsoft clippy I'm not sure how many

play10:04

of us will actually remember that but

play10:06

there used to be these agents uh in this

play10:09

in sort of our industry and in research

play10:12

so this is certainly not a new idea uh

play10:15

so if you go all the way back um so we

play10:19

had agents like clippy and in many ways

play10:22

these agents especially in the

play10:23

reinforcement learning and machine

play10:25

Learning Community agents were these

play10:27

elements that basically

play10:29

could simulate human behavior I think

play10:31

that is ultimately sort of underlying

play10:33

thesis but many of the agents were given

play10:36

tools to automate certain tasks and the

play10:39

tasks it were meant to automate were

play10:41

tasks that are not simple right it's not

play10:44

something like you're running a for loop

play10:46

with your python code but it's a little

play10:47

bit more complex in there right it

play10:49

operates in much more embodied spaces or

play10:52

in spaces that we often operated right

play10:56

the web uh right can it the simplest

play11:00

example with these kind of tool-based

play11:01

agents are can it order me pizza can it

play11:04

buy plane tickets and those might sound

play11:07

simple but we we know from our

play11:09

experiences that even ordering pizza

play11:11

actually does require multiple steps

play11:13

right we need to travel to certain

play11:15

websites we need to look through the

play11:16

menus actually make the payment and de

play11:19

with sort of entering your address and

play11:21

so forth so that was one Gena of agents

play11:24

that already sort of existed for a long

play11:26

time or I would say all genas of Agents

play11:29

sort of existed but that was one draine

play11:30

that was highlighted in the past uh so

play11:33

you see things like clipp is also in

play11:35

that draine as well you're a Microsoft

play11:38

uh office user clippy would try to

play11:40

automate some tasks for you based on

play11:41

your prior interaction with the

play11:43

software another a set of Agents um was

play11:48

this idea of simulation agents or agents

play11:51

that were cre cl to clarify on that

play11:53

point those agents are single agents

play11:56

correct they can be single agents they

play11:58

were

play11:59

often implemented as single agents

play12:02

that's right I don't think by definition

play12:04

they actually had to be single agents so

play12:06

you actually try you're now seeing in at

play12:08

least in the research you're start to

play12:11

see glimpse of people trying to imagine

play12:14

what would it look like for these agents

play12:16

to be in a multi-agent setting so

play12:18

research paper that I remember coming

play12:20

out after uh gener of Agents was

play12:23

basically what if you have a company of

play12:26

Agents right there's going to be a CEO

play12:28

but there's also going to be designer

play12:30

agent who spe who works in some other

play12:33

aspects there's going to be editor in

play12:34

this company and those are still much

play12:37

within the literature of what I would

play12:40

call tool-based agents right they're

play12:42

trying to automate some complex tasks

play12:45

for the users and I think there's going

play12:47

to be a lot of sort of really big

play12:49

opportunities in the space that's

play12:50

something that people have been working

play12:51

on for a long time uh for for all the

play12:53

right reasons now another community that

play12:59

has form but to some extent actually has

play13:02

a slightly different route is agents

play13:05

that were created for

play13:08

simulations uh and these agents were

play13:11

certainly a part of games right in the

play13:13

past we had Sims but we also had these

play13:16

NPC characters that we could interact

play13:18

with now those NPCs and agents back then

play13:21

were very much like it was simpler

play13:23

agents that were either rule-based uh

play13:26

there were some reinforcement learning

play13:27

agents back then as well in

play13:29

space uh but another one that we could

play13:32

usually think about were agents that

play13:34

were used basically in social science

play13:37

economic agents or agents that would

play13:39

simulate our policy decision making and

play13:41

so forth uh and those agents were also a

play13:45

part of this

play13:46

literature and what we're seeing today

play13:49

is we're one recognizing that lar

play13:51

language model is simulating human

play13:54

behavior so it touches on all these

play13:55

agents that it can be a foundational

play13:57

sort of a architectural layer for

play13:59

creating all these different sorts of

play14:01

Agents but in terms of our initial

play14:03

application spaces we're seeing this

play14:05

split where there's one Community who's

play14:07

now deeply interested in agents using

play14:09

tools but another community that that is

play14:12

deeply interested in this idea of can we

play14:14

simulate and this is where I would say

play14:17

like

play14:17

multi-agents uh and as well as

play14:19

personalization is really starting to be

play14:21

highlighted in the simulation space

play14:23

because it's a little bit more directly

play14:26

incorporated into the idea of

play14:27

simulations who are we simulating for

play14:30

what are we simulating who are we

play14:32

simulating and by definition simulations

play14:34

often happen in this multi-agent space

play14:36

so those are the two communities that

play14:38

you're starting to see um so generative

play14:41

agent certainly stands on the far end of

play14:43

the simulation based agents whereas some

play14:45

other projects that were also really

play14:47

cool last year I think a lot of sort of

play14:49

uh open AI gpts I would say are another

play14:52

end of the simulation agents or another

play14:54

end of toolbase Agents so those are the

play14:57

axes that you're sort of seeing right

play14:58

now now I sort of End by saying my hunch

play15:01

actually is again though because they

play15:03

all start from the same technical thesis

play15:05

that we can simulate human behavior they

play15:07

will merge in the end I don't think they

play15:09

will be like completely separate thesis

play15:11

like five to 10 years down the line it's

play15:14

more going to be the question of where

play15:16

are we going to make our short-term bets

play15:18

and what's going to be an interesting

play15:20

and meaningful application space in the

play15:23

next two to five years so that's the

play15:25

field that I'm seeing and how it's

play15:26

developing right now before we maybe go

play15:29

into that could you maybe describe how

play15:32

um llm specifically has

play15:35

affected um the especially the latter

play15:38

cohort right what is the before and what

play15:40

is the after and what is the magnitude

play15:43

of improvement because of this uh

play15:46

technology that's now you know cheap

play15:48

enough to use right so L Lang motor is

play15:51

really what made this possible uh that

play15:54

is really the fundamental T stack that

play15:56

we needed uh in the past when you wanted

play15:59

to create and this goes for both types

play16:01

of Agents uh tool-based and simulations

play16:04

what you really needed

play16:07

was you basically needed rule-based

play16:09

agents that was the most common and Rule

play16:11

based agents are sort of a more

play16:13

sophisticated way of saying we're

play16:15

scripting all the behaviors so imagine

play16:18

you're building an NPC in for a game a

play16:21

human author would actually write every

play16:24

sentence that the agent would say to the

play16:25

user for instance it would author would

play16:29

actually describe in in either code or

play16:31

language if this happens you do this so

play16:35

you basically design all the possible

play16:37

behaviors now that is expensive and not

play16:40

scalable right um and that was the

play16:42

fundamental block that we had now

play16:45

toolbase agents had similar issues that

play16:49

in many of the context it had to operate

play16:51

it's not very generalizable tool so if

play16:54

you sort of see how clippy or even some

play16:58

of the agents that we're using today

play17:00

very simple types of agent actually are

play17:02

already embedded into our daily usage so

play17:04

you may have used Google spreadsheet or

play17:06

Google doc it would autocomplete in some

play17:10

very rudimentary way that actually could

play17:12

be considered in some ways an agent in

play17:14

this direction of tool-based agents and

play17:17

the rules they were using so far were

play17:19

very simple it's not exactly rule based

play17:21

but it is something that was very much

play17:23

hardcoded into the agent's behavior and

play17:25

there wasn't there was some learning

play17:27

going on but that were very root like

play17:29

very straightforward simple uh like

play17:31

statistics that we're using what L

play17:34

language model changes is l l model

play17:37

gives us a single ingredient which is

play17:40

given a micro context micro moment let's

play17:43

say I'm sitting in this room talking to

play17:48

Jan and about let's say generative

play17:50

agents or simulations and so forth given

play17:53

that micro moment description a language

play17:56

model is extremely good at predicting

play17:59

the next moment right so what is what is

play18:03

the what are the reasonable set of

play18:04

things that June might say in this

play18:07

particular conversation given what he

play18:09

knows it's very good at doing that um

play18:13

that on its own not is not a perfect

play18:16

agent uh or it's not the complete

play18:18

ingredient that you need to create these

play18:21

agents that are meant to live for many

play18:23

many years or decades but they are the

play18:27

right ingredient or building block that

play18:29

we needed because that can be used to

play18:31

replace what was in the past manual

play18:35

authoring in the past we had to manually

play18:37

author all the possible sequences given

play18:40

any micro moment but L's language model

play18:43

can come

play18:44

in so given that ingredient what we

play18:48

really could do is bake in long-term

play18:51

memory and some reflection module on top

play18:53

of it and planning module so given the

play18:56

micro ingredient plus an agent

play18:59

architecture that we give it on top of

play19:01

it these agents can basically now start

play19:04

to function as something that can

play19:06

operate in that in a world that's much

play19:09

like ours with a fairly decent degree of

play19:13

long-term coherence so that's where we

play19:15

are and that's really the difference it

play19:16

made and I'd say this is sort of a zero

play19:19

to one difference not a degree

play19:21

difference because before Lis language

play19:22

model this was not

play19:24

possible what else is um so we like

play19:28

large language models gave memory gave

play19:31

context uh gave interactions to these

play19:33

agents what else in a perfect world

play19:37

would these agents have in order to

play19:39

better mimic the real world like what's

play19:41

maybe in the next next step just out of

play19:43

curiosity uh right so to clarify L

play19:46

language model doesn't actually have so

play19:48

L language model provides one element

play19:50

it's the micro sort of a module for

play19:53

predicting the next sequence yeah uh the

play19:56

it is the agent architecture that

play19:58

actually ends up giving the memory and

play20:00

planning ability um but those two pair

play20:03

becomes a fantastic combination um now

play20:08

going forward what I do think is going

play20:10

to be interesting are so right now we're

play20:12

using L language model but we may have

play20:14

have all noticed that things like chpt

play20:17

can now not only deal with just language

play20:21

but also other modality like

play20:24

image I think that's going to be really

play20:27

interesting right

play20:28

so right now let's say if you and this

play20:32

is sort of based on my prior work like

play20:34

call generative agents and we had this

play20:36

game world like Sim and that we call a

play20:38

Smallville the way these agents

play20:41

perceived and operated in their world

play20:44

was basically by um by translating like

play20:47

our system translating the visual world

play20:50

into natural language so we would tell

play20:53

the agent you are in your apartment or

play20:56

you are in the kitchen

play20:58

talking to someone so we would actually

play21:01

take the visual world and use our system

play21:05

to translate the visual world into

play21:07

natural language and then feeding it to

play21:09

the agent architecture that would use

play21:11

Earth language model to process this but

play21:13

now with these models being able to deal

play21:17

with multimodal aspect we might actually

play21:20

be able to bypass that pH and go

play21:23

straight from here is the visual world

play21:25

or space that you're seeing right now

play21:28

that is your memory now act on it I

play21:32

think that's going to be potentially

play21:33

very powerful because in

play21:35

part image is much richer to some it

play21:39

conveys a lot more I I do come from

play21:42

natural language uh processing

play21:43

background at least that's my other half

play21:46

of my sort of academic background so I

play21:49

have bias to towards believing the

play21:51

natural language is profound and I think

play21:53

that we're going to be that will be the

play21:55

case going forward as well but image

play21:57

does offer something that just language

play22:00

alone does not so image is going to be a

play22:03

big thing now imagine in the future

play22:05

video is going to be a big thing as well

play22:07

then gradually the more these agents

play22:10

where basically increasingly get more

play22:12

powerful as this new modality gets piled

play22:14

on so that's something that we should be

play22:16

looking forward to that's that's great

play22:19

what are some on the downside what are

play22:21

some of the

play22:22

limitations um that you're seeing in

play22:25

terms of these um these agents

play22:27

especially General

play22:29

agents right um

play22:32

so to so there are limitations that I

play22:36

can mention just about sort of in sort

play22:38

of in the context of our work um and

play22:40

then I think there are going to be

play22:41

interesting limitations that are much

play22:43

more application specific so for

play22:46

generative agents today certainly the

play22:49

technical limitation right now might

play22:51

have to do with things like so you're

play22:53

using whether it's an open so right now

play22:56

we use open AI model open AI has

play22:59

actually done a lot of work to make the

play23:01

model safer and I think for open AI I

play23:04

think that was the right approach in the

play23:06

sense that what they really wanted to

play23:07

create was these chatbots or agents or

play23:11

chpt that was safe tool to use for most

play23:15

of people now if you want to run a

play23:18

simulation or create truly accurate and

play23:20

believable agents with something like

play23:22

chpt however that could become a

play23:26

limitation because what we really

play23:29

experience as humans uh is we fight we

play23:33

sometimes have conflicts we disagree

play23:34

with each other and that might not be

play23:37

something that something like Chachi B

play23:39

that's been fine tuned to not behave

play23:41

that way to remain safe it's something

play23:44

it might not be something that these

play23:45

models will try to surface and that

play23:48

could be a potential Block in creating

play23:52

more accurate more believable

play23:54

simulations or agents for that matter so

play23:57

that certainly is one limitation right

play24:00

now that we're seeing um an interesting

play24:03

way to tackle this I think going forward

play24:05

is to use open source models or other

play24:08

models that have less of this fine-tuned

play24:11

nature um but it's going to be highly

play24:14

dependent on the models that we'll be

play24:16

using for this so I think that's one

play24:18

thing to look forward to got it got it

play24:21

that's that's super that's super helpful

play24:24

maybe one last question on kind of the

play24:26

research side um when you think about

play24:29

future areas to explore for you

play24:32

specifically yeah um what are some of

play24:37

the more narrow topics that you're

play24:39

hoping to to dive deeper in given

play24:44

world so ultimately I think making the

play24:49

agents more accurate as rep for these

play24:52

agents to be more accurate reflection of

play24:54

who we are I think it's going to be a

play24:56

really interesting research and much

play24:59

like it's I think it's going to that's

play25:01

going to be an area that's going to have

play25:02

more of a research and broader impact um

play25:06

so right now you may have seen the sort

play25:09

of simulation demo the agents that live

play25:11

in that simulation are fictional that we

play25:14

just for instance we have an an agent

play25:17

named Isabella we told Isabella that she

play25:19

is a cafe owner and LGE language model

play25:23

basically makes up what a Persona that

play25:26

is reasonable given

play25:28

description but I think it's going to be

play25:30

far more interesting if we can make

play25:32

these simulations actually closely model

play25:36

our actual human communities so it's not

play25:39

just functional but actually has

play25:40

groundings that's going to open up from

play25:43

our perspective an entirely new set of

play25:46

application spaces as well as research

play25:48

impact Um this can be used for instance

play25:51

to accurately model or predict markets

play25:54

or it's going to be able to use to more

play25:58

closely personalized many of these

play26:00

agents for individual use cases so

play26:03

that's something that we're looking

play26:04

forward to in terms of sort of a

play26:05

particular topic that we're diving into

play26:08

that plus of course scaling of the

play26:09

agents I think that's another big one

play26:11

but those two that that makes sense you

play26:14

know one of the things that's missing in

play26:17

most

play26:18

AI uh Technologies is like really the

play26:22

emotional part of how humans feel right

play26:25

right like all of that data is largely

play26:28

not captured and therefore not part of

play26:31

any kind of models today uh language is

play26:34

one small output of what we have it's

play26:36

it's a very important output for sure

play26:39

right um but it's still one small output

play26:40

so I wonder how uh we might be able to

play26:43

incorporate some of the data around our

play26:47

emotions I agree and thoughts in the

play26:50

future um maybe let's move on to the

play26:53

applications today since you talked

play26:55

about some of the challenges for agent

play26:59

many organizations are thinking about

play27:02

how to use large language models today

play27:04

right there's a huge amount of

play27:05

aspirations a subset of them are also

play27:08

thinking about what are some of the

play27:09

agent technology applications that are

play27:12

viable within an Enterprise which has

play27:16

limitations around infrastructure around

play27:19

data silos security and all that

play27:22

stuff any particular areas where you've

play27:25

seen companies be successful at using

play27:28

these Technologies in

play27:30

production right so I think this is

play27:33

going to be incredibly like Case by case

play27:36

uh answer so I let me

play27:41

think or if not like any hypothesis as

play27:45

to where you might see the first

play27:48

commercial deployments at

play27:50

scale

play27:54

right so there is something that I have

play27:58

I this is there's a message that I have

play28:01

in trying to communicate I think in in

play28:03

different settings so this is not

play28:05

something I'm I sort of conve for the

play28:06

first

play28:07

time and and my opinion has been getting

play28:11

updated but I think fundamentally I

play28:13

think this is right so the way I've been

play28:15

describing it is in human computer

play28:18

interaction or in

play28:20

most task settings there are two types

play28:23

of problems that we deploy our machines

play28:25

or agents in one

play28:27

has very hard Edge problem spaces these

play28:30

are things like hey order me pizza or

play28:33

buy me a plane ticket these are tasks

play28:35

where there's a very concrete outcome

play28:38

and there often is a right or wrong

play28:40

answer that the agent has to achieve at

play28:43

least from the user's perspective there

play28:45

is something that will absolutely be yes

play28:48

or

play28:49

no and then there are problems PES where

play28:52

we have soft Edge problems these are

play28:55

problems where we can increase L he'll

play28:58

climb towards sort of being better but

play29:00

at certain level it starts to actually

play29:02

become useful so to make this intuition

play29:04

a little bit to make this a little bit

play29:06

more intuitive here for instance if I

play29:09

guess the worst possible case scenario

play29:11

is I ask the agent to buy me a plane

play29:13

ticket and it it bought me the wrong

play29:15

ticket that just goes to a different

play29:16

place then that's like a heavy

play29:19

no uh whereas let's say I ask the agent

play29:24

to sort of simulate the behavior that is

play29:28

sort of fun so that I when I'm in a game

play29:30

this is sort of entertaining and

play29:32

interesting that might be something that

play29:34

the Asian doesn't need to be quite

play29:35

perfect in but it can still get there

play29:37

quite quickly and then we can gradually

play29:40

improve those two are the spaces that we

play29:43

can sort of in terms of when we consider

play29:46

where to deploy or how this will

play29:48

actually make its first impact we might

play29:50

actually be looking at those problem

play29:51

spaces first and these are the classes

play29:54

that I'm seeing if I were to make my bet

play29:57

AG agents will likely succeed first in

play30:00

the soft Edge problems basis and will

play30:03

gradually inch into making it work in

play30:05

the hardage problems space this has been

play30:08

sort of an intuition with agent research

play30:10

Community for some time so when clippy

play30:14

for instance

play30:15

failed our intuition there at least from

play30:18

research perspective wasn't that these

play30:21

agents failed because we didn't have the

play30:23

technology there certainly these were

play30:25

deployed and there were some confidence

play30:27

around the technology but the the

play30:29

problem there actually was with

play30:31

interaction that when agents are

play30:34

deployed in hardage problems basis it's

play30:36

often deployed in states where it

play30:39

actually has to have a fairly long chain

play30:42

of steps and fairly High the risk were

play30:46

fairly High when it fails the cost of

play30:48

correcting its error is actually quite

play30:50

High Cost of auditing its error is

play30:52

actually quite

play30:53

high so when these agents are deployed

play30:56

in hardage problems SP bis it has to

play30:58

reckon with the fact that it will

play30:59

undoubtedly make

play31:01

mistakes and it when it does make

play31:03

mistake it has to be increasingly

play31:05

auditable and controllable by the users

play31:08

so that the cost of correcting its error

play31:11

is not high enough that from the user's

play31:14

perspective the cost benefit anal

play31:16

analysis basically has to make sense and

play31:19

that's been a fundamental challenge with

play31:21

agents that's why in every era we see

play31:25

the interest around agents Spike for a

play31:27

while and then it quickly subsides after

play31:29

maybe a half a year or two year now

play31:32

there's there's a real question now

play31:35

though that given the large language

play31:36

model and the progress we saw that this

play31:38

might not be the case this time or at

play31:39

some point we might be able to make this

play31:41

work but for now so I'm closely

play31:45

monitoring this as well and I think we

play31:47

all should I don't think we should just

play31:48

go and say because it didn't work before

play31:50

it's not going to work this time but my

play31:54

hunch is that we will likely see very

play31:57

similar pattern Arise at least for the

play32:00

first of future and we haven't quite

play32:03

dealt with the interaction problems with

play32:05

those types of Agents so I think it's

play32:08

much safer to assume that it is going to

play32:11

be the softage problems basis and that's

play32:13

why in some in many of the aspects

play32:15

that's why we our team was also

play32:17

interested in this idea of simulation

play32:19

because simulation is sort of the prime

play32:22

example of soft problems spasis where

play32:24

the simulation has to be good enough for

play32:27

to be start being useful that's also why

play32:30

I think a lot of really early promising

play32:32

AI startups that's going to go in the

play32:34

agent space are places that does NPCs

play32:37

for games because those are very safe

play32:40

softage problem spaces where the agents

play32:42

can fail but that's okay and gradually

play32:45

we'll sort of go to the other area as

play32:47

well but I think that's where the impact

play32:49

is going to start in the next couple of

play32:52

years I I I also think just seeing um

play32:56

the startups in the the space in sectors

play33:00

and functions that allow for failure

play33:04

like you said include things like

play33:08

marketing where if you Market

play33:11

incorrectly it's not that big of a deal

play33:14

if you uh write the wrong

play33:18

code that's probably going to be a

play33:20

bigger deal if you pick the wrong

play33:22

security features that's a huge deal

play33:25

right U if you pick the wrong things for

play33:27

healthcare that's you know that's that's

play33:29

an even bigger deal so there are degrees

play33:31

of Tolerance fault tolerance within the

play33:34

Enterprise um that's one thing and the

play33:37

second on the consumer side especially

play33:40

if the agents are just assisting

play33:42

consumers by not executing on

play33:46

anything um that could probably also

play33:48

work for example you know there's a

play33:51

company called rewind which is using

play33:52

some of the agent Technologies I believe

play33:55

um and they're getting a bunch of

play33:57

consumer demand but but what consumers

play34:00

are doing is just searching for a

play34:04

behavior that they have had

play34:06

before um and this this this product is

play34:09

helping them do that versus do anything

play34:12

you know um real world really

play34:16

interesting uh but that's that's super

play34:19

the way that you frame it is very um uh

play34:22

is very

play34:23

useful what about just from an

play34:26

architecture

play34:27

standpoint uh large language models

play34:31

enabled by Transformer

play34:33

architectures this is a whole different

play34:35

direction we're already seeing companies

play34:38

that are saying hey Transformers are are

play34:41

not the most efficient you know

play34:43

inference costs are very high uh let's

play34:46

look at this the next thing have you

play34:48

spent much time thinking about that um

play34:51

and if so any impact to the work that

play34:54

you're

play34:55

doing right so certainly um next sort of

play35:00

model that we're going to be banking on

play35:02

I think that always is an important

play35:04

topic and that's something that I think

play35:06

we as a community always has to sort of

play35:08

monitor because I think you're right the

play35:10

Transformer is not going to be the end

play35:12

model that would be hopefully I mean KN

play35:15

going what the Hope here is that we

play35:17

wouldn't be using Transformer 10 years

play35:18

down the

play35:20

line uh but one way that we do view this

play35:23

is this is very much like a programmer a

play35:26

programmer way of looking at this we

play35:28

view this in abstractions right so what

play35:31

Transformer has gotten us right now is

play35:33

this amazing capacity for reasoning and

play35:36

processing information and generating

play35:39

information so it might be the case that

play35:42

in the future that task will be done by

play35:44

even better models and hopefully that's

play35:47

going to be the case um but for the sake

play35:50

of building applications it is true that

play35:52

we can sort of view this as a layer of

play35:54

abstraction that there might be some

play35:56

other of Technology that's going to be

play35:58

powering it in the future but really

play36:00

what we're focusing on is the capacity

play36:03

and the modality what kind of reasoning

play36:06

using what modality can these technology

play36:08

that exist today do and we're going to

play36:10

be building on top of it so I think

play36:13

that's sort of uh our way of looking at

play36:15

it in sort of I say medium term like in

play36:18

the next three to five years now if

play36:21

you're look because right now there are

play36:24

some promising architectures that's

play36:25

going on that's sort of been created at

play36:28

the Forefront of I more in the machine

play36:30

learning and natural language processing

play36:32

communities that I personally getting a

play36:35

little bit excited about um but at the

play36:38

moment those are still in the very much

play36:40

in the research Pro uh phase and can you

play36:43

share some examples of that yeah so I

play36:46

think like there's one model that

play36:48

recently came out like Mamba by some

play36:49

folks uh those are from Stanford uh

play36:52

folks so I think the author is now as

play36:55

CMU and Princeton uh sort of all with

play36:57

sort of this community so that's one

play37:00

example of sort of a potentially

play37:01

promising or interesting model and

play37:03

that's a one that I recently heard about

play37:04

that I think is interesting to be

play37:06

looking

play37:07

at but these models for them to be

play37:10

deployed at scale in a commercial way if

play37:14

we decide to basically go with certain

play37:17

model that's getting created today it

play37:19

will give us

play37:21

maybe two to five year timeline before

play37:24

they can really take off because

play37:26

Transformer is

play37:28

not it is a relatively modern model but

play37:32

it really how you look at sort of the

play37:34

timeline this transform sevenish years

play37:37

sevenish years so I think hopefully we

play37:40

if we find something like this time

play37:41

maybe it's going to go much faster but

play37:43

it still took about sevenish years for

play37:45

chat PT to really come out so it's not

play37:48

immediate uh whereas a lot of

play37:50

interactions that we can build I think

play37:51

there's a lot that we can do like today

play37:53

to create really cool experiences so I

play37:56

think that's how we're looking at this

play37:57

there is a medium-term this is where we

play37:59

are focused on the level of extraction

play38:01

that this is the capacity that we'll

play38:03

have and then maybe in the down the line

play38:05

five to 10 year term we can really be

play38:07

looking forward to some new models

play38:09

that's going to make an

play38:10

impact that's great that's great um cool

play38:14

very cool um maybe more generally if you

play38:17

just zoom out for for a moment when you

play38:21

look at the ecosystem today uh what are

play38:24

some of the problems that you want to

play38:26

see solve we talked about multimodal a

play38:28

little bit we talked about new models

play38:30

right after Transformers that might come

play38:33

out what are some of the problems that

play38:35

you are most excited about someone

play38:38

solving not necessarily uh you

play38:41

personally but someone solving I sort of

play38:43

have two in mind um and it's a little

play38:48

bit less of a this is a specific problem

play38:50

that I want someone to solve but it's

play38:52

more sort of questions that I have that

play38:55

I think more of us should be thinking

play38:57

about and to some extent and this this

play39:00

happens a lot with sort of the way I do

play39:02

my research as well where I get inspired

play39:05

by big problems or foundational problems

play39:08

that we had in previous decades because

play39:11

often times there's a lot of insights

play39:13

that we can learn from the past as we

play39:16

sort of build on the

play39:17

future

play39:19

one certainly I'm embedded into this

play39:22

agent

play39:24

space one is in the past agent had its

play39:28

hype cycle basically but it

play39:32

failed um that the hype cycle lasted for

play39:35

a couple of years and then people very

play39:37

quickly lost interest basically because

play39:39

it didn't quite deliver the promises

play39:42

that it

play39:43

had I think it's worth asking ourselves

play39:47

why that was the case I think the

play39:50

opportunity this time is real but I also

play39:53

think the opportunity in the past was

play39:54

also real to some aspect as well so just

play39:58

because the opportunity is real and

play40:00

language model is really cool doesn't

play40:01

necessarily guarantee us at least from

play40:03

my perspective that we're going to that

play40:05

agent will finally be a thing that

play40:07

everyone will use down the f I think

play40:10

there is a future where that's that will

play40:12

happen at some point I think it might

play40:14

even happen this cycle but I think it's

play40:16

really worth asking as a community why

play40:20

did it fail in the past so that we don't

play40:21

repeat those mistakes one sort of main

play40:24

thing I'm sort of curious about that I

play40:25

don't think a lot of are thinking about

play40:27

is actually not the technology part but

play40:30

the

play40:31

interaction how are these agents going

play40:33

to be used in what way because

play40:36

ultimately that's where it really

play40:37

delivers value to the end users and

play40:41

that's where agents in the past have

play40:42

failed that was really cool technology

play40:45

but we didn't seriously ask ourselves is

play40:47

this something that people really need

play40:49

and does the cost benefit analysis of

play40:51

using these agents and learning how to

play40:53

use them well really makes sense for the

play40:55

broader user user base so that's one and

play40:59

other one is sort of

play41:02

my it's a little bit of a hot take but

play41:04

it's also a shorter take which is we

play41:07

have large language models and I think

play41:08

these have made a huge impact already

play41:11

right the number of users who use CH

play41:13

jbts that's

play41:15

incredible but I think it's sort of

play41:17

worth asking

play41:18

ourselves is that sort of quote unquote

play41:21

the killer applications that we were

play41:23

waiting for because in many ways

play41:27

chbt or maybe it is and I I think if it

play41:31

is I think somebody should articulate

play41:33

this but chpt does feel like it's a

play41:36

fairly simple wrap around L language

play41:38

model because that's what itain is and

play41:41

you know open AI has done fantastic

play41:43

things to make it safer and make it more

play41:45

useful by tuning I think what really

play41:47

great but I think it's worth asking if

play41:50

that is actually the killer application

play41:52

why is it a killer application and the

play41:55

answer might actually come out that

play41:56

maybe it actually isn't the killer

play41:58

application that we were waiting for um

play42:01

in which case what is going to be the

play42:03

killer application that's really going

play42:05

to add value in a much more

play42:07

generalizable way um that's a very

play42:10

abstract question it's for for now it's

play42:13

for me it's just a hunch that I think

play42:15

there's something to be asked about

play42:16

there and if I'm wrong like I would also

play42:19

love to hear again somebody really say

play42:21

we already have this clear application

play42:23

maybe it's co-pilot track GPT and here's

play42:25

why

play42:26

but for now this is a question that I'm

play42:29

still asking

play42:30

myself that makes sense um and thank you

play42:33

thank you for sharing that what are some

play42:35

of your favorite AI apps today that you

play42:39

use I love chbt I use it every day I

play42:43

chpt did make a difference in my

play42:45

workflow uh so I so as sort of a

play42:48

researcher and one of the main things I

play42:50

do is I program every day uh or at least

play42:53

most days or I write or write papers so

play42:56

I do one of those two things chpt is

play42:59

fantastic at both uh so it's you know

play43:03

it's as all programmers sort of know you

play43:05

know we sometimes don't bother

play43:06

remembering all the different functions

play43:08

or

play43:09

documentation it's very good at

play43:11

generating a lot of the code when I have

play43:13

an idea really impressive it's also

play43:16

quite a good editor so if I make grammar

play43:19

error in my sort of sentences chpt will

play43:22

usually catch them for me right it's

play43:24

simple and easy thing but it's good

play43:26

enough now that it's actually making a

play43:28

difference in the

play43:29

workflow um so I say trpd for sure by

play43:32

extension I think co-pilot will make a

play43:35

difference um so it's sort of worth

play43:37

asking maybe going back to the question

play43:39

around like killer replication what is

play43:41

the definition of killer application I

play43:43

think it does some people Define it as

play43:46

application that has more users and the

play43:48

fact of that I think always has to be

play43:50

the case that no killer application has

play43:52

no user killer application by default

play43:54

means the application that will have the

play43:56

most number of

play43:58

users but I think there's more

play44:01

theoretical definition to what a killer

play44:03

application is that implies a lot of

play44:05

users or the most number of users but

play44:08

for instance if we look back to the

play44:10

prior era of PC a killer application

play44:12

that I mentioned was something like

play44:14

Microsoft's Excel or this tabular data

play44:18

format the thing that would man that

play44:19

would let us manipulate the tabular data

play44:22

so really the definition in a more

play44:24

theoretical sense of K application here

play44:27

is there's new technology stack that is

play44:29

been developed there's a new file type

play44:31

that is getting generated then the

play44:33

killer application is the one that would

play44:35

let us manipulate the file application

play44:37

file type that's one theoretical

play44:40

definition that one could give at least

play44:42

that's sort of the definition that I've

play44:44

been toying with I think it's an

play44:45

interesting one I don't think it's the

play44:46

only one but those are sort of the ways

play44:48

that I'm looking at this that that that

play44:51

makes a lot of sense I also use CH GPT

play44:53

every single day uh it's been um it's

play44:55

been very helpful everything from coming

play44:57

up with menu names to you know rewriting

play45:02

uh emails that don't sound as As Nice um

play45:05

and I've tried a little bit to you know

play45:07

give it files and images oh I actually

play45:09

help my mother create a background she's

play45:12

a dancer so she was performing and

play45:14

wanted a very specific background for

play45:15

her dance and you know I created that

play45:17

for her using chat gbt so all sorts of

play45:20

utility there um but I love the way that

play45:22

you framed um the last potential

play45:24

application maybe just one last question

play45:27

from from my end any um any resources or

play45:31

books that you you love that is uh on um

play45:36

this

play45:38

topic on this

play45:40

topic

play45:46

right I I do think and this is this is

play45:49

often the case with many many of The

play45:51

Cutting Edge like spaces I think a lot

play45:53

of the papers that are coming out that

play45:55

are gaining a lot of attention I think

play45:56

those are sort of wor checking out as

play45:58

sort of the resources it's not exactly

play45:59

like here's one book that we can all

play46:01

look at uh uh but things are moving fast

play46:04

enough that I think um I think those are

play46:07

sort of interesting resources are just

play46:09

things that are getting created today uh

play46:11

and they're that quations uh so those I

play46:13

sort of mentioned as sort of a generic

play46:16

answer I do think I think this has been

play46:19

a sort of running theme in some of the

play46:20

things I mentioned today I get inspired

play46:24

by insight

play46:28

that basically had impact and stood the

play46:30

test of

play46:31

time right and the reason why that is

play46:34

the case is because I personally think

play46:37

all the great all great ideas are sort

play46:39

of Timeless that because current hyp

play46:43

cycle is over doesn't mean they're less

play46:44

interesting or less meaningful for sure

play46:46

that foundational ideas that will

play46:48

continue to have impact um so when I

play46:52

look for resources I actually look back

play46:54

to books from truly the prior

play46:57

Generations um so some of the works that

play47:00

I often go back to are works by Herbert

play47:02

Simon Ali new those are founders of AI

play47:05

and this many of this Fields who would

play47:07

later go on to win the Turing award the

play47:09

Nobel Prize and so forth and those Works

play47:13

uh early cognitive psychologists and

play47:15

scientists inspired my worka uh and

play47:17

their textbook uh those people actually

play47:20

have written books because they were

play47:22

much more established than sort of The

play47:23

Cutting Edge spaces today uh so I go

play47:26

back to those as sort of my personal

play47:28

resources for getting

play47:30

ideas that's great thank you so much

play47:32

this was super super helpful to me

play47:34

personally because what we do as

play47:36

investors is we try to understand the

play47:40

impacts of technology and start to

play47:43

invest in companies when it

play47:46

becomes at the beginning of when it

play47:48

becomes commercially viable right so to

play47:50

your point around what are the problem

play47:53

spaces what are the applications in

play47:54

which this can be applied in a cost

play47:56

effective and secure way that where the

play48:00

end user is willing to

play48:03

interact uh and dra and and get value

play48:06

that's when we start to come in uh and

play48:09

invest in these companies which will

play48:11

hopefully uh be much bigger companies in

play48:13

the future so really appreciate this

play48:15

chat yeah it was

play48:19

[Music]

play48:24

fun