AI Pioneer Shows The Power of AI AGENTS - "The Future Is Agentic"
Summary
TLDR在本次演讲中,Dr. Andrew Ning分享了他对代理(agents)的乐观看法,特别是像GPT 3.5这样的模型如何通过代理工作流程提升到GPT 4的水平。他强调了多代理协作和迭代的重要性,并通过案例研究展示了代理工作流程的显著效果。他还讨论了代理设计模式,如反思、工具使用、规划和多代理协作,以及这些模式如何推动人工智能的未来。最后,他提出了对快速生成令牌和代理推理设计模式的期待,认为这些将为实现更广泛的AI任务打开大门。
Takeaways
- 🧠 Dr. Andrew Ning 是一位计算机科学家,曾是Google Brain的联合创始人和首席科学家,目前在人工智能领域具有重要影响力。
- 🌟 Sequoia 是硅谷最著名的风险投资公司之一,其投资的公司总市值占纳斯达克市场的25%以上。
- 📈 通过使用代理工作流,即使是GPT 3.5这样的模型也能实现接近GPT 4的性能水平。
- 🔄 代理工作流的核心优势在于迭代,通过多个具有不同角色和能力的代理不断改进任务结果。
- 🤖 代理可以采用不同的设计模式,包括反思、使用工具、规划和多代理协作,以提高问题解决的效率。
- 🔧 反思工具使大型语言模型能够自我评估并改进其输出,从而提高性能。
- 🛠️ 工具使用允许代理使用预定义的工具,如API调用和外部库,以增强其功能。
- 📝 规划使代理能够慢下来,计划步骤,通过逐步推理来提高输出质量。
- 🤼♂️ 多代理协作通过不同代理之间的合作,模拟人类团队合作解决问题的方式。
- 🚀 快速生成令牌(tokens)对于代理工作流至关重要,因为它允许代理快速迭代和改进。
- 🌐 随着模型变得更加商品化,未来可能会看到更多关于代理和推理速度的改进,这将推动人工智能的发展。
Q & A
Dr. Andrew Ning 是谁?
-Dr. Andrew Ning 是一位计算机科学家,曾是 Google Brain 的联合创始人和首席科学家,也是在线教育平台 Coursera 的联合创始人。他在人工智能领域有着举足轻重的影响力。
Sequoia 是一家怎样的公司?
-Sequoia 是一家传奇的硅谷风险投资公司,其投资组合中的公司占纳斯达克总市值的25%以上,投资过的知名公司包括 Apple、Instagram、Zoom 等。
什么是非代理性工作流程?
-非代理性工作流程是指用户向 AI 模型提出问题,模型生成答案,类似于人一次性写作一篇文章,不使用退格键进行修改。
代理性工作流程有什么特点?
-代理性工作流程是一种迭代的过程,涉及多个具有不同角色、背景、人格和工具的代理,它们共同协作,通过多次迭代来完成任务。
Dr. Andrew Ning 为什么认为代理性工作流程如此强大?
-Dr. Andrew Ning 认为代理性工作流程的强大之处在于可以结合多个代理的力量,每个代理都有不同的专长,通过迭代和协作来达到最佳的结果,这与人类的工作方式相似。
在编程基准测试中,使用零次射击提示的 GPT 3.5 和 GPT 4 的正确率是多少?
-在使用零次射击提示时,GPT 3.5 的正确率是 48%,而 GPT 4 的正确率是 67%。
Dr. Andrew Ning 介绍了哪些代理性工作流程的设计模式?
-Dr. Andrew Ning 介绍了四种代理性工作流程的设计模式:反思(reflection)、工具使用(tool use)、规划(planning)和多代理协作(multi-agent collaboration)。
反思在代理性工作流程中是如何应用的?
-反思是指让大型语言模型对其输出进行反思,找出改进的方法并返回改进后的结果。这种方法可以使模型的性能得到显著提升。
工具使用在代理性工作流程中的作用是什么?
-工具使用允许给大型语言模型提供额外的工具,如自定义编码工具、API 调用或复杂的数学库,从而扩展模型的功能和应用范围。
多代理协作如何提高工作流程的效率?
-多代理协作通过让不同的代理扮演不同的角色,如编码者、批评者、设计师等,共同协作完成任务。这种协作可以产生更好的结果,并且能够模拟人类的协作方式。
Dr. Andrew Ning 对于未来 AI 发展的趋势有何看法?
-Dr. Andrew Ning 认为代理性工作流程将是未来 AI 发展的重要趋势,他预计 AI 能够完成的任务范围将会因为代理性工作流程而大幅扩展。同时,他也看好推理速度的提升和模型的进一步商品化。
Outlines
🤖 人工智能代理的未来展望
本段落介绍了Dr. Andrew Ning在Sequoia的演讲,他对于人工智能代理的未来非常乐观。他提到像GPT 3.5这样的模型可以驱动代理进行推理,达到GPT 4的水平。他还分享了许多关于人工智能的有趣见解。Andrew Ning是计算机科学家,曾是Google Brain的联合创始人和首席科学家,也是在线教育平台Coursera的联合创始人。他的演讲强调了人工智能代理的重要性和未来潜力。
📈 人工智能代理的工作流程与案例研究
这一段讨论了非代理性工作流程与代理性工作流程的对比。通过一个编写文章的例子,说明了代理性工作流程的迭代性和协作性,以及如何通过多个代理(如写作代理、审阅代理、拼写检查代理等)共同工作,不断迭代改进结果。此外,通过一个编程基准测试的例子,展示了使用GPT 3.5和代理性工作流程相比零拍GPT 4能获得更好的结果,强调了代理性工作流程的强大能力。
🛠️ 人工智能代理的设计模式
本段介绍了几种人工智能代理的设计模式,包括反思工具的使用、工具使用、规划和多代理协作。反思是指让大型语言模型审视并改进其输出。工具使用意味着可以给模型提供工具,如网络爬虫工具或数学库。规划是指让模型进行更慢的思考和步骤规划。多代理协作是指不同代理之间的合作,如自动生成代码和审查代码的代理。这些设计模式展示了人工智能代理的多样性和灵活性。
🌐 人工智能代理的未来发展
这一段讨论了人工智能代理的未来发展,包括对更快的推理速度和更高效的代理解算模型的期待。提到了使用代理工作流程可以显著提高生产力,并且随着更好的代理模型和工具的出现,代理的'finicky'方面将大大减少。此外,还提到了多代理系统和辩论的潜力,以及如何通过不同模型的协作获得更好的性能。
🚀 人工智能代理的快速迭代与性能提升
最后一段强调了快速迭代和改进在人工智能代理中的重要性。提到了快速生成令牌的能力对于代理工作流程的重要性,以及如何通过快速迭代来提高模型的性能。此外,还提到了对下一代模型如GPT-5和Clay 4的期待,以及这些模型如何帮助我们更接近通用人工智能(AGI)的目标。最后,演讲者对人工智能代理的未来表示了极大的乐观,并鼓励观众对这一领域保持关注。
Mindmap
Keywords
💡人工智能代理
💡迭代工作流程
💡工具使用
💡多代理协作
💡反思
💡计划
💡Dr. Andrew Ning
💡Sequoia
💡GPT 3.5和GPT 4
💡云计算
💡AGI(通用人工智能)
Highlights
Dr. Andrew Ning is incredibly bullish on agents and their potential in AI development.
Agents can reason at levels comparable to GPT 4 when powered by models like GPT 3.5.
Dr. Ning is a renowned computer scientist, co-founder of Google Brain, and founder of Corsera.
Sequoia, the venue of the talk, has an impressive portfolio including 25% of NASDAQ's total value.
Non-agentic workflows are likened to asking a person to write an essay without ever using backspace.
Agentic workflows involve multiple agents with different roles and tools, iterating on tasks for better outcomes.
Using agentic workflows, GPT 3.5 can outperform GPT 4 in certain tasks.
Agentic workflows can lead to remarkable improvements in coding benchmarks.
Reflection is a tool that allows large language models to self-improve their outputs.
Tool use allows LMs to utilize pre-existing code and libraries, expanding their capabilities.
Planning and multi-agent collaboration are emerging technologies with great potential.
Multi-agent systems can simulate different roles and collaborate on complex tasks.
Agent workflows can produce better results than zero-shot prompting or single model use.
The future of AI may involve a shift in expectations for immediate responses to prompts.
Fast token generation is crucial for iterative agentic workflows.
Agentic reasoning design patterns are expected to greatly expand the set of tasks AI can perform.
The path to AGI is seen as a journey, with agent workflows contributing to incremental progress.
Transcripts
Dr Andrew ning just did a talk at
Sequoia and is all about agents and he
is incredibly bullish on agents he said
things like GPT 3.5 powering agents can
actually reason to the level of GPT 4
and a lot of other really interesting
tidbits so we're going to watch his talk
together and I'm going to walk you
through step by step what he's saying
and why it's so important I am
incredibly bullish on agents myself
that's why I make so many videos about
them and I truly believe the future of
artificial intelligence is going to be a
gentic so first who is Dr Andrew ning he
is a computer scientist he was the
co-founder and head of Google brain the
former Chief scientist of Buu and a
leading mind in artificial intelligence
he went to UC Berkeley MIT and Carnegie
melon so smart smart dude and he
co-founded this company corsera where
you can learn a ton about computer
science about math a bunch of different
topics absolutely free and so what he's
doing is truly incredible and so when he
talks about AI you should listen so
let's get to this talk this is at seoa
and if you're not familiar with seoa
they are one of the most legendary
Silicon Valley venture capital firms
ever now here's an interesting stat
about seoa that just shows how
incredible they are at picking
technological winners their portfolio of
companies represents more than 25% of
Today's total value of the the NASDAQ so
the total value of all the companies
that are listed on the NASDAQ 25% of
that market capitalization are companies
that are owned or have been owned or
invested in by seoa incredible stat
let's look at some of their companies
Reddit instacart door Dash Airbnb a
little company called Apple block
snowflake vanta Zoom stripe WhatsApp
OCTA Instagram this list is absolutely
absurd all right another of the preface
let me get into the Talk itself so a
agents you know today the way most of us
use l Shish models is like this with a
non- agentic workflow where you type a
prompt and generates an answer and
that's a bit like if you ask a person to
write an essay on a topic and I say
please sit down to the keyboard and just
type the essay from start to finish
without ever using backspace um and
despite how hard this is L's do it
remarkably well in contrast with an
agentic workflow this is what it may
look like have an AI have an LM say
write an essay outline do you need to do
any web research if so let's do that
then write the first draft and then read
your own first draft and think about
what parts need revision and then revise
your draft and you go on and on and so
this workflow is much more iterative
where you may have the L do some
thinking um and then revise this article
and then do some more thinking and
iterate this
through a number of times so I want to
pause it there and talk about this
because this is the best explanation for
why agents are so powerful I've heard a
lot of people say well agents are just
llms right and yeah technically that's
true but the power of an agentic
workflow is the fact that you can have
multiple agents all with different roles
different backgrounds different personas
different tools working together and
iterating that's the important word
iterating on a task so in this example
he said okay write an essay and yeah an
llm can do that and usually it's pretty
darn good but now let's say you have one
agent who is the writer another agent
who is the reviewer another for the
spell checker another for the grammar
checker another for the fact Checker and
they're all working together and they
iterate over and over again passing the
essay back and forth making sure that it
finally ends up to be the best possible
outcome and so this is how humans work
humans as he said do not just do
everything in one take without thinking
through and planning we plan we iterate
and then we find the best solution so
let's keep listening what not many
people appreciate is this delivers
remarkably better results um I've
actually really surprised myself working
these agent workflows how well how well
they work other let's do one case study
at my team analyzed some data using a
coding Benchmark called the human eval
Benchmark released by open a few years
ago um but this says coding problems
like given the nonent list of integers
return the sum of all the odd elements
are an even positions and it turns out
the answer is you co snipper like that
so today lot of us will use zero shot
prompting meaning we tell the AI write
the code and have it run on the first
spot like who codes like that no human
codes like that just type out the code
and run it maybe you do I can't do that
um so it turns out that if you use GPT
3.5 uh zero shot prompting it gets it
48% right uh gbd4 way better 67 7% right
but if you take an agentic workflow and
wrap it around GPT 3.5 say it actually
does better than even
gbd4 um and if you were to wrap this
type of workflow around gbd4 you know it
it it also um does very well all right
let's pause here and think about what he
just said over here we have the zero
shot which basically means you're simply
telling the large language model do this
thing not giving it any example not
giving it any chance to think or to
iterate or any fancy prompting just do
this thing and it got the human evalve
Benchmark 48% correct then GPT 4 67%
which is you know a huge Improvement and
we're going to continue to see
Improvement when GPT 5 comes out and so
on however look at this GPT 3.5 wrapped
in an agentic workflow any of these all
perform better than the zero shot GPT 4
using only GPT 3.5 and this lb BD plus
reflection it's actually nearly 100%
it's over 95% then of course if we wrap
GPT 4 in the agentic workflow metag GPT
for example we all know about it
performs incredibly well across the
board and agent coder kind of at the top
here so it's really just showing the
power of agentic workflows and you
notice that GB 3.5 with an agentic
workflow actually outperforms
gp4 um and I think this has and this
means that this has signant consequences
I think how we all approach building
applications so agents is the term has
been tossed around a lot there's a lot
of consultant reports how about agents
the future of AI blah blah blah I want
to be a bit concrete and share of you um
the broad design patterns I'm seeing in
agents it's a very messy chaotic space
tons of research tons of Open Source
there's a lot going on but I try to
categorize um bit more concretely what's
going on agents reflection is a tool
that I think many of us are just use it
just works uh to use I think it's more
widely appreciated but actually works
pretty well I think of these as pretty
robust Technologies when I all right
let's stop there and talk about what
these things are so reflection is as
obvious as it sounds you are literally
saying to the large language model
reflect on the output you just gave me
find a way to improve it then return
another result or just return the
improvements so very straightforward and
it seems so obvious but this actually
causes large language models to perform
a lot better and then we have tool use
and we learned all about tool use with
projects like autogen and crew AI tool
use just means that you can give them
tools to use you can custom code tools
it's like function calling so you could
say Okay I want a web scraping tool and
I want an SEC lookup tool so you can get
stock information about ticker symbols
you can even plug in complex math
libraries to it I mean the possibilities
are literally endless so you can give a
bunch of tools that the large language
model didn't previously have you just
describe what the tool does and the
large language model can actually choose
when to use the tool it's really cool
use them I can you know almost always
get them to work well um planning and
multi-agent collaboration I think is
more emerging when I use them sometimes
my mind is blown for how well they work
but at least at this moment in time I
don't feel like I can always get them to
work reliably so let me walk through
these full design Pat
all right so he's going to walk through
it but I just want to touch on what
planning and multi-agent collaboration
is so planning we're basically saying
giving the large language model the
ability to think more slowly to plan
steps and that's usually by the way why
in all of my llm tests I say explain
your reasoning step by step because that
kind of forces them to plan and to think
through each step which usually produces
better results and then multi-agent
collaboration that is autogen and crew
AI that is a very emergent technology
techology I am extremely bullish on it
it is sometimes difficult to get the
agents to behave like you need them to
but with enough QA and enough testing
and iteration you usually can and the
results are phenomenal and not only do
you get the benefit of having the large
language model essentially reflect with
different personalities or different
roles but you can actually have
different models powering different
agents and so you're getting the benefit
of the reflection based on the quality
of each model so you're basically
getting really different opinions as
these agents are working together so
let's keep listening and if some of you
go back and yourself will ask your
engineers to use these I think you get a
productivity boost quite quickly so
reflection here's an example let's say I
ask a system please write Cod for me for
a given task then we have a coder agent
just an LM that you prompt to write code
to say you def do Tas write a function
like that um an example of
self-reflection would be if you then
prompt the LM with something like this
here's code intended for a toss and just
give it back the exact same code that
they just generated and then say check
the code carefully for correctness sound
efficiency good construction CRI just
write a prompt like that it turns out
the same L that you prompted to write
the code may be able to spot problems
like this bug in line five and fix it by
blah blah blah and if you now take his
own feedback and give it to it and
reprompt it it may come up with a
version two of the code that could well
work better than the first version not
guaranteed but it works you know often
enough but this to be worth trying for a
law of appli so what you usually see me
doing in my llm test videos is for
example let's say I say write the Game
snake in Python and it gives me the game
Snake it's that is zero shot I'm just
saying write it all out in one go then I
take it I put it in my VSS code I play
it I get the error or I look for any
bugs and then I paste that back in to
the large language model to fix now
that's essentially me acting as an agent
and what we can do is use an agent to
automate me so basically look at the
code look for any potential errors and
even agents that can run the code get
the error and pass it back into the
large language model now it's completely
automated coding to foreshadow to use if
you let it run unit tests if it fails a
unit test then why do you fail the unit
test have that conversation and be able
to figure out failed the unit test so
you should try changing something and
come up with V3 by the way for those of
you that want to learn more about these
Technologies I'm very excited about them
for each of the four sections I have a
little recommended reading section in
the bottom that you know hopefully gives
more references and again just the
foreshadow of multi-agent systems I've
described as a single coder agent that
you prompt to have it you know have this
conversation with itself um one Natural
Evolution of this idea is instead of a
single code agent you can have two
agents where one is a code agent and the
second is a critic agent and these could
be the same base LM model but they you
prompt in different ways where you say
one your exper coder right code the
other one say your expert code review as
to review this code and this type of
workflow is actually pretty easy to
implement I think such a very general
purpose technology for a lot of
workflows this will give you a
significant boost in in the performance
of LMS um the second design pattern is
to use many of you will already have
seen you know lmb systems uh uh using
tools on the left is a screenshot from
um co-pilot on the right is something
that I kind of extracted from uh gbd4
but you know LM today if you ask it
what's the best coffee maker can do web
search for some problems LMS will
generate code and run codes um and it
turns out that there are a lot of
different tools that many different
people are using for analysis for
gathering information for taking action
personal productivity um it turns out a
lot of the early work and to use turned
out to be in the computer vision
Community because before large language
models LMS you know they couldn't do
anything with images so the only option
was that the LM generate a function call
that could manipulate an image like
generate an image or do object detection
or whatever so if you actually look at
literature it's been interesting how
much of the work um in two years seems
like it originated from Vision because
Elms would blind to images before you
know GPD 4V and and and lava and so on
um so that's to use in it all right so
tool use incredibly incredibly important
because you're basically giving the
large language model code to use it is
hardcoded code so you always know the
result it's not another large language
model that might produce something a
little different each time this is
hardcoded and always is going to produce
the same output so these tools are very
valuable and the cool thing about tools
is we don't have to rewrite them right
we don't have to write them from scratch
these are tools that programmers already
test app to use in their code so whether
it's external libraries API calls all of
these things can now be used by large
language models and that is really
exciting we're not going to have to
rewrite all of this tooling and then
planning you know for those of you that
have not yet played a lot with planning
algorithms I I feel like a lot of people
talk about the chat GPT moment where
you're wow never seen anything like this
I think if not use planning alums many
people will have a kind of a AI agent
wow I couldn't imag imagine the AI agent
doing this so I've run live demos where
something failed and the AI agent
rerouted around the failure I've
actually had quite a few of them like
wow you can't believe my AI system just
did that autonomously but um one example
that I adapted from hugging GPT paper
you know you say this general image
where the girls read where girl and by
the way I made a video about hugging GPT
it is an amazing paper I'll link that in
the description below I was reading a
book and it post the same as a boy in
the image example le. jpack and please
subcribe the new imagy re voice so give
an example like this um today we have ai
agents who can kind of decide first
thing I need to do is determine the post
of the boy um then you know find the
right model maybe on hugging face to
extract the post then next need to find
a post image model to synthesize a
picture of a of a girl of as following
the instructions then use uh image to
text and then finally use text to speech
and today we actually have agents that
I don't want to say they work reliably
you know they're kind of finicky they
don't always work but when it works is
actually pretty amazing but with agentic
Loop sometimes you can recover from
earlier failures as well so yeah and
that's a really important Point agents
are a little bit finicky but since you
can iterate and the Agents can usually
recover from their issues that makes
them a lot more powerful and as we
continue to evolve agents as we get
better agentic models better tooling
better Frameworks like crew aai and
autogen all of these kind of finicky
aspects of agents are going to start to
get reduced tremendously I find myself
already using research agents in some of
my work well one a piece of research but
I don't feel like you know Googling
myself and spend long time I should send
to the research agent come back in a few
minutes and see what it's come up with
and and it it sometimes works sometimes
doesn't right but that's already a part
of my personal
workflow the final design pattern multi-
Asian collaboration ation this is one of
those funny things but uh um it works
much better than you might think uh uh
but on the left is a screenshot from a
paper called um chat Dev I made a video
about this it'll be in the description
below as well uh which is completely
open which actually open source many of
you saw the you know flashy social media
announcement of demo of a Devon uh uh
Chad Dev is open source it runs on my
laptop and what Chad Dev does is example
of a multi-agent system where you prompt
one LM to sometimes act like the CEO of
a software engine company sometimes act
a designer sometime a product manager
sometimes ACC a tester and this flock of
agents that you buil by prompting an LM
to tell them you're now coo you're now
software engineer they collaborate have
an extended conversation so that if you
tell it please develop a game develop a
GOI game they'll actually spend you know
a few minutes writing code testing it
iterating and then generate a like
surprisingly complex programs doesn't
always work I've used it sometimes it
doesn't work sometimes is amazing but
this technology is really um getting
better and and just one of design
pattern it turns out that multi-agent
debate where you have different agents
you know for example could be have ch
GPT and Gemini debate each other that
actually results in better performance
as well all right so he said the
important part right there when you have
different agents and each of them are
are powered by different models maybe
even fine-tuned models fine-tuned
specifically for their task and their
role you get really good performance and
that is exactly what a project like crew
AI like autogen is made for so Gabby
multiple simulated air agents work
together has been a powerful design
pattern as well um so just to summarize
I think these are the these are the the
the uh patterns I've seen and I think
that if we were to um use these uh uh
patterns you know in our work a lot of
us can get a prity boost quite quickly
and I think that um agentic reasoning
design patterns are going to be
important uh this is my small slide I
expect that the set of task AI could do
will expand dramatically this year uh
because of agentic workflows and one
thing that it's actually difficult
people to get used to is when we prompt
an LM we want to response right away um
in fact a decade ago when was you know
having discussions around at at at
Google on um called a big box search
type in Long prompt one of the reasons
you know I failed to push successfully
for that was because when you do a web
search you one have responds back in
half a second right that's just human
nature we like that instant gra instant
feedback but for a lot of the agent
workflows um I think we'll need to learn
to dedicate the toss and AI agent and
patiently wait minutes maybe even hours
uh to for response but just like us I've
seen a lot of novice managers delegate
something to someone and then check in
five minutes later right and that's not
productive um I think we need to it be
difficult we need to do that with some
of our AI agents as well all right so
this is actually a point which I want to
pose a different way of thinking about
it think about grock grock grq you get
500 700 850 tokens per second with grock
with their architecture and all of a
sudden the agents which you know you
usually expect them to take a few
minutes to do a semi complex task all
the way up to 10 15 20 minutes depending
on what the task is a lot of the time in
that task completion is the inference
running that is assuming you're getting
you know 10 15 20 tokens per second with
open AI but if you're able to get 800
tokens per second it's essentially
instant and a lot of people when they
first saw grock they thought well what's
the point of 800 tokens per second
because humans can't read that fast this
is the best use case for that agents
using hyper inference speed and reading
each other's responses is the best way
to leverage that really fast inference
speed humans don't actually need to read
it so this is a perfect example so if
all of a sudden that part of your agent
workflow is extremely fast and then
let's say we get an embeddings model to
be that fast all of a sudden the slowest
part of the entire agent workflow is
going to be searching the web or hitting
a third party API it's no longer going
to be the inference and the embeddings
and that is really exciting let's keep
watching the end and then one other
important Trend fast token generation is
important because with these agentic
workflows we're iterating over and over
so the elm is generating tokens for the
to read and I think that um generating
more tokens really quickly from even a
slightly lower quality LM might give
good results compared to slower tokens
from a betm maybe it's a little bit
controversial because it may let you go
around this Loop a lot more times kind
of like the results I showed with gpdc
and an agent architecture on the first
slide um and cand I'm really looking
forward to Cloud 5 and Cloud 4 and gb5
and Gemini 2.0 and all these other one4
models that many building and part of me
feels like if you're looking forward to
running your thing on gb5 zero shot you
know you may be to get closer to that
level of performance on some
applications than you might think with
agent reasoning um but on an early model
I think I I I I think this is an
important Trend uh uh and honestly the
path to AGI feels like a journey rather
than a destination but I think this typ
of agent workflows could help us take a
small step forward on this very long
journey thank you okay so he said a lot
of important things at the end there one
thing he said is if you're already
looking forward to GPT 5 clae 4 the
basically the next generation of The
Cutting Edge models you might be able to
achieve
and what's the cost of all these tokens
and all of that I think is going to get
sorted out as models become more and
more commoditized so I'm super excited
about agents I'm super excited about
inference speed improvements and I hope
you liked Andrew ning's talk if you
liked this video please consider giving
a like And subscribe and I'll see you in
the next one
浏览更多相关视频
Networking for GenAI Training and Inference Clusters | Jongsoo Park & Petr Lapukhov
GPT-4o AI Agents: Easily Create Medical Research Agents (Praison AI)
AI Leader Reveals The Future of AI AGENTS (LangChain CEO)
【生成式AI導論 2024】第9講:以大型語言模型打造的AI Agent (14:50 教你怎麼打造芙莉蓮一級魔法使考試中出現的泥人哥列姆)
What's next for AI agentic workflows ft. Andrew Ng of AI Fund
2 Ex-AI CEOs Debate the Future of AI w/ Emad Mostaque & Nat Friedman | EP #98
5.0 / 5 (0 votes)