#黃仁勳 驚喜為 #美超微 #梁見後 #COMPUTEX 主題演講站台|完整精華

天下雜誌 video
5 Jun 202421:35

Summary

TLDR在这段视频中,Nvidia的CEO Jensen Huang分享了他对人工智能和加速计算的远见。他强调了数据量指数级增长与CPU扩展速度放缓之间的矛盾,以及加速计算如何帮助释放数据中心的浪费能量,并加速每个应用。Jensen还介绍了生成式AI(generative AI)的概念,它将影响全球所有数据中心,推动其现代化。此外,他还提到了液冷技术在降低功耗方面的作用,以及如何通过提高吞吐量、利用率和启动速度来增加收入。最后,他强调了安全性、技术进步和良好政策的重要性,以确保AI的健康发展。

Takeaways

  • 🧠 AI正在改变最小化计算,因为数据量呈指数级增长,而CPU的扩展速度已经放缓多年。
  • 🌿 绿色计算(Green Computing)即能效计算,是Nvidia所关注的,旨在减少数据中心的能源浪费。
  • 🚀 加速计算(Accelerated Computing)时代已经到来,它将释放数据中心的潜能,提高应用和数据中心的效率。
  • 🤖 生成式AI(Generative AI)是一种新兴的计算方式,它涉及文本、图像、视频的生成,将影响全球所有的数据中心。
  • 💰 数据中心需要现代化改造以适应生成式AI,这将带来巨大的需求和市场潜力。
  • 🔧 Nvidia和Supermicro合作,提供多种新产品和解决方案,以支持加速计算和数据中心的现代化。
  • 💡 液冷技术(DLC)正在被用于降低数据中心的功耗,提高AI芯片的制造能力。
  • 🔗 高速互连技术(如NVLink)使得GPU和CPU之间能够高效地传输数据,这对于AI训练和推理至关重要。
  • 🛠️ Nvidia的软件堆栈,包括CUDA和网络计算,为高性能计算提供了强大的支持。
  • 🔐 安全技术的发展同样重要,需要为AI制定安全政策和实践,以确保其可靠性和安全性。

Q & A

  • Nvidia的CEO Jensen在演讲中提到了哪些关于AI和计算的变革?

    -Jensen在演讲中提到了加速计算(accelerated computing)和绿色计算(Green Computing)的到来,强调了由于数据量指数级增长和CPU扩展放缓,加速计算能够释放数据中心中浪费的能源和成本。他还提到了生成式AI(generative AI)的崛起,这将影响全球所有的数据中心。

  • Jensen提到了加速计算能带来哪些好处?

    -加速计算能够节省数据中心中浪费的能源和成本,提高应用和数据中心的效率。通过加速计算,可以释放之前浪费的能源,用于新的目的,比如加速每个应用和数据中心。

  • 什么是生成式AI,它将如何影响数据中心?

    -生成式AI是一种能够生成文本、图像、视频等内容的人工智能技术。Jensen认为,生成式AI的转变将影响世界上每一个数据中心,因为到2030年,全球数据的价值可能达到3万亿美元,这要求我们现代化所有的数据中心,以适应这些先进的系统。

  • Jensen如何描述Super Micro和Nvidia的合作关系?

    -Jensen将Super Micro团队描述为准备好接受订单的合作伙伴,他们准备好提供所需的产品和服务,以帮助现代化数据中心。他还开玩笑说自己是Super Micro的最佳销售人员,并且他们之间有良好的合作关系。

  • 在演讲中提到的数据中心现代化涉及到哪些方面?

    -数据中心现代化涉及到采用加速计算和生成式AI等先进技术,以及采用液冷系统(DLC)来降低功耗,提高性能和成本效率。此外,还需要软件的兼容性和整个系统的整合,以确保快速启动、高利用率和高吞吐量。

  • 为什么Jensen强调了数据中心的能源效率?

    -Jensen强调能源效率是因为在加速计算中,能源效率直接转化为更高的性能。通过节省能源,可以将这些能量用于更多的计算任务,从而提高数据中心的整体性能和产出。

  • Nvidia在数据中心现代化中扮演了什么角色?

    -Nvidia在数据中心现代化中扮演了关键角色,提供了加速计算平台和生成式AI技术。Nvidia的GPU和其他技术是构建现代AI工厂的基础,这些AI工厂能够直接产生收入。

  • Jensen提到了哪些Nvidia的新产品或技术?

    -Jensen提到了Nvidia的多个新产品和技术,包括新的GPU服务器、液冷系统(DLC)以及与Super Micro合作的各种配置,如Blackwell GPU和Grace CPU的结合使用。

  • 在演讲中,Jensen如何描述AI的未来和其在不同行业中的应用?

    -Jensen认为AI的未来非常重要,他提到生成式AI能够嵌入智能,并且可以在许多不同的行业中使用。他强调了AI的实用性,以及如何通过不同的技术来提高AI的安全性和功能性。

  • Jensen在演讲中提到了哪些关于AI安全性的观点?

    -Jensen提到为了确保AI的安全性,我们需要开发出多种技术,比如AI监控AI、人员监控AI以及建立各种安全护栏。他强调了良好的政策、实践和工业标准的重要性,以确保AI的安全和可靠。

  • Jensen如何描述Super Micro在数据中心现代化中的角色?

    -Jensen将Super Micro描述为一个准备好立即服务的合作伙伴,他们能够提供所需的硬件和工程支持,以帮助客户快速启动和运行现代化的数据中心。

  • 在演讲中,Jensen提到了哪些关于数据中心的经济效益?

    -Jensen提到了数据中心的经济效益与吞吐量、利用率和启动时间直接相关。他强调了通过提高这些指标,可以增加数据中心的收入,这也是为什么他们称其为AI工厂的原因。

  • Jensen在演讲中提到了哪些关于软件在高性能计算中的作用?

    -Jensen强调了软件在高性能计算中的重要性,他提到了Cuda、网络计算以及分布式计算的软件堆栈,这些都是构建在Nvidia硬件之上的。他还提到了软件的兼容性,这是高性能计算中非常关键的部分。

Outlines

00:00

🤖 AI与加速计算的新时代

本段落介绍了Nvidia CEO Jensen Huang的演讲,他强调了AI如何改变最小化计算,并提出了加速计算和绿色计算的概念。加速计算由于数据量指数级增长和CPU扩展放缓,变得尤为重要。Jensen提到了数据中心的能源和成本浪费问题,并强调通过加速计算可以释放这些资源,用于新的目的。此外,他还提出了生成性AI的概念,这是一种能够生成文本、图像、视频等的AI技术,它将影响全球所有的数据中心,并需要在未来几年内进行现代化改造。

05:01

🚀 数据中心的液体冷却技术革新

在第二段中,讨论了数据中心的液体冷却技术(DLC),这是一种降低功耗并提高AI芯片制造能力的技术。Jensen Huang提到了Supermicro公司每月能够生产1000个液体冷却系统,这将是一个巨大的公司规模。他还提到了GPU服务器的复杂性,强调了除了GPU之外,还有许多其他组件和技术。此外,Jensen还提到了通过液体冷却系统节省下来的能源可以用于未来的计算,这将有助于创建新的商品——计算吞吐量,这将直接影响数据中心的收入。

10:02

🏭 AI工厂的兴起与经济效益

第三段讲述了AI工厂的概念,这些工厂不仅仅是数据存储或文件检索的地方,而是直接产生收入的设施。Jensen Huang强调了AI工厂的启动时间、吞吐量和利用率对收入的影响,以及如何通过整合系统到机架规模来提高这些指标。他还提到了Supermicro公司如何准备好为这些AI工厂提供支持,并且所有这些系统都经过了Nvidia软件的认证。此外,Jensen还讨论了不同类型的配置,以适应不同的使用案例和数据中心需求。

15:03

🌐 网络作为计算结构的重要性

在第四段中,Jensen Huang讨论了网络作为计算结构的重要性,特别是对于分布式计算。他提到了Cuda和DOA这两个软件堆栈,它们使得工作负载能够高效地在网络中分布。此外,他还提到了400和800吉比特每秒的高速互连,以及即将到来的1600吉比特每秒的技术。Jensen强调了在训练阶段和推理阶段,不同类型的CPU如何被用于AI,以及如何通过节能的互连提高性能。

20:05

🛡️ AI的安全性与未来发展

最后一段中,Jensen Huang讨论了AI的安全性和未来发展。他比喻了飞机的自动驾驶技术,强调了为了确保AI的安全性,需要发明许多技术、实践和政策。Jensen提到了AI监控、人为监督以及安全防护措施的重要性。他强调了为了使AI变得“难以置信地好”,需要在技术、安全性、政策和实践方面不断进步。最后,他以幽默的方式结束了演讲,强调了购买更多产品可以提高安全性的观点。

Mindmap

Keywords

💡人工智能

人工智能(AI)是指由人制造出来的能够执行复杂任务的系统,这些任务通常需要人类智能才能完成。在视频中,人工智能被提到是改变数据中心的关键技术之一,并且与加速计算和绿色计算紧密相关。例如,Nvidia的CEO Jensen提到了AI在推动数据中心发展中的作用。

💡加速计算

加速计算是指使用专门的硬件和软件来提高数据处理速度,尤其是在处理大量数据时。视频中提到,随着数据量的指数级增长,CPU的扩展速度已经放缓,因此加速计算变得尤为重要,能够释放数据中心中浪费的能源和成本。

💡绿色计算

绿色计算通常指的是能效高的计算,旨在减少能源消耗和环境影响。在视频中,Nvidia强调了绿色计算的重要性,特别是在加速数据中心的能效和性能方面,以及如何通过绿色计算实现成本和能源的节约。

💡生成式AI

生成式AI是一种人工智能技术,它能够生成新的数据实例,如文本、图像或视频。视频中提到,生成式AI将影响世界上每一个数据中心,并且是推动数据中心现代化的关键因素之一。

💡数据中心现代化

数据中心现代化是指更新和升级数据中心的基础设施,以提高效率、性能和可持续性。视频中提到了全球数据中心的价值,并强调了到2030年,需要对这些数据中心进行现代化改造的紧迫性。

💡液冷系统

液冷系统是一种冷却技术,它使用液体(如水或特殊冷却液)来吸收和传输热量,以降低设备的运行温度。视频中提到了Nvidia正在出货的液冷系统,以降低功耗并提高AI芯片的制造能力。

💡GPU

GPU(图形处理单元)是一种专门设计用于处理图形和图像计算的微处理器。在视频中,GPU被提及为构建复杂系统的核心,这些系统被用来执行高级计算任务,如AI训练和推理。

💡软件栈

软件栈是指一系列相互兼容的软件产品,它们共同工作以提供特定的功能。视频中提到了几个关键的软件栈,包括CUDA和用于网络的软件,这些都是高性能计算的基础。

💡Cuda

CUDA(Compute Unified Device Architecture)是Nvidia推出的一个并行计算平台和编程模型,它允许开发者使用Nvidia的GPU进行通用计算。视频中提到CUDA作为构建软件栈的基础之一。

💡以太网

以太网是一种局域网技术,用于计算机之间的数据传输。在视频中,以太网被提到作为现代计算网络的一部分,不再是仅仅用于发送电子邮件,而是成为了支持分布式计算的计算网络。

💡AI安全

AI安全是指在开发和部署人工智能系统时采取的安全措施,以确保系统的可靠性和安全性。视频中提到了AI安全的重要性,并强调了需要为AI制定安全技术和政策。

Highlights

英伟达CEO Jensen Huang分享了他对AI改变世界的伟大愿景。

加速计算和绿色计算同时到来,标志着数据中心的能源效率和成本效率的显著提升。

数据量呈指数级增长,而CPU扩展速度放缓,导致数据中心存在大量能源和成本浪费。

加速计算可以释放数据中心的潜能,为新目的使用之前浪费的能源。

英伟达宣布了220种新产品,展示了其在加速计算领域的创新能力。

生成性AI(Generative AI)的出现,预示着数据中心的转型。

到2030年,全球数据中心的价值可能达到3万亿美元,需要现代化改造以适应生成性AI。

超微公司(Super Micro)准备接受订单,提供现代化的数据中心解决方案。

英伟达正在出货数据中心液冷技术,以降低功耗并提高AI芯片的制造效率。

超微公司每月出货1000个液冷系统,展现了其在数据中心冷却技术方面的领先地位。

数据中心的现代化改造对于提高能源效率和降低成本至关重要。

英伟达的GPU不仅是芯片,整个系统比单纯的GPU要复杂得多。

超微公司的工程设计使得安装和启动时间极快,提高了数据中心的利用率和吞吐量。

英伟达的软件许可确保了所有产品在数据中心的兼容性和即插即用。

数据中心被视为AI工厂,直接产生收入,强调了可靠性、吞吐量和启动时间的重要性。

英伟达的CUDA和其他软件堆栈为高性能计算提供了基础。

网络现在是计算的织物,而不仅仅是发送电子邮件的工具。

英伟达的Blackwell GPU和Grace CPU通过高速互连连接,优化了训练和推理的性能。

AI技术的发展需要同时关注安全性和效率,确保AI的可靠性和实用性。

英伟达致力于推进AI技术的发展,同时注重安全技术和政策的制定。

Transcripts

play00:00

I know only

play00:01

some fortunately we are very lucky again

play00:05

to invite the AI

play00:07

genius our common

play00:13

friend our common friend is very busy

play00:15

huh Invidia found CEO Jensen to share

play00:20

his great vision with us

play00:23

[Applause]

play00:28

[Music]

play00:29

[Applause]

play00:29

[Music]

play00:33

thank

play00:33

you hi

play00:38

everybody now

play00:41

what that AI is changing minium because

play00:47

of you what's new

play00:51

today I have to admit just now when I

play00:55

was coming to your keynote in the car I

play00:58

fell asleep

play01:01

and so right now right now I'm a little

play01:04

bit groggy so if I say nonsense things

play01:08

please I let me apologize first no well

play01:12

let's see um uh Charles we've gone back

play01:15

a very long ways yeah and and um uh what

play01:20

are we doing oh I needed some water I

play01:23

need to spe up okay right my energy

play01:28

yeah they said I was on this side and

play01:30

you keep going on my

play01:33

side this is what happens when we don't

play01:36

practice you don't need to and you are

play01:39

no time you you don't need and so so um

play01:43

I uh what what were we saying um this is

play01:46

a very important time because we have a

play01:47

new agent Computing coming there are two

play01:50

things that are happening at the same

play01:51

time the first is accelerated Computing

play01:54

accelerated Computing has arrived at a

play01:57

time

play01:58

oh Green Computing

play02:01

yeah Green computer yeah okay

play02:11

Computing I think I think when you say

play02:14

Green Computing you mean energy

play02:15

efficient Computing right yes Nvidia is

play02:18

energy efficient Computing yes we have S

play02:21

we follow you all

play02:23

right look Green Computing and Green

play02:27

Computing all right so so um uh

play02:30

accelerated computing's time has come

play02:32

because for a very long time the amount

play02:35

of data processing has been increasing

play02:38

exponentially yeah and yet CPU scaling

play02:41

has slowed for many many years so we've

play02:44

been we have now an enormous amount of

play02:47

waste wasted energy and wasted cost

play02:50

trapped inside the data centers so when

play02:53

we accelerate the data centers the

play02:56

savings

play02:57

incredible because it has been sold long

play03:00

of waste

play03:02

trapped and so now we can release the

play03:05

waste and use that energy for a new

play03:10

purpose number one accelerate every

play03:12

application accelerate every data center

play03:15

these

play03:16

amazing servers here right so many new

play03:19

products so many new products you have

play03:21

220 new products unbelievable did he

play03:24

tell you that already no I W very high

play03:30

I came to announce super micros products

play03:33

and so that's the first thing the second

play03:35

thing is because the Energy Efficiency

play03:38

and the performance efficiency and the

play03:39

cost efficiency is so incredibly great

play03:42

with accelerated Computing a new way of

play03:44

doing Computing has emerged and it's

play03:46

called generative AI generative AI is an

play03:50

incredible thing people say generative

play03:53

AI inference it's related not the same

play03:56

inference

play03:58

recognizing C dog speech inference

play04:03

generation text Generation image

play04:06

Generation video generation that's what

play04:09

we call a generative AI the pressure of

play04:12

generative AI to not the pressure but

play04:15

the the transition to generative AI will

play04:18

affect every single data center in the

play04:19

world we have a trillion dat a trillion

play04:22

dollars worth of data centers in the

play04:23

world that's established $3 trillion

play04:26

probably by 2030 in another 6 years we

play04:29

have to to modernize all of them with

play04:32

these amazing systems yeah that's the

play04:35

reason why the demand is so great

play04:36

because all of these data centers has to

play04:38

be modernized and Charles and the super

play04:42

micro team is ready to take your

play04:47

order Json I'm your I'm your best sales

play04:51

guy thank you I work on commission no

play04:57

commission we buy more cheaper from you

play04:59

don't buy more

play05:00

[Laughter]

play05:05

chips so

play05:08

that's Jon sh Michael is now shipping

play05:12

data center uh liqu cooling DLC R inum

play05:16

production now to lower the power

play05:18

consumption so you can manufacture more

play05:21

AI chip yeah yeah thousand of how here

play05:25

you see

play05:31

[Applause]

play05:39

[Laughter]

play05:43

I have many American colleagues they

play05:45

don't understand my Chinese I have many

play05:47

Chinese colleagues they don't understand

play05:49

my

play05:52

Chinese hi

play05:56

y we are shipping up to 1,000 R per

play06:01

month now 1,000 R like it is multiply by

play06:05

ASP yeah you're going to be a gigantic

play06:08

company yeah thank

play06:11

you that's why I need a more

play06:14

CH did you guys all do the

play06:17

math Millions

play06:19

times thousands time 52 no no no you

play06:24

charging me $2 million more than $2

play06:26

million for d

play06:31

[Laughter]

play06:37

are we allowed to do this on TV are we

play06:40

on

play06:41

TV I I guess the well is

play06:49

this so we are shipping about 1,000

play06:53

that's incredible now this this uh

play06:55

600,000 Parts this is probably more than

play06:57

600,000 parts how many pounds oh I don't

play07:02

know can I move three I think it's 3,000

play07:06

lb more than 3,000 lb

play07:09

yeah it's incredible so yeah our goal

play07:13

this year is to ship more than 10,000

play07:17

record you know the Charles this is the

play07:19

thing that's really amazing uh people

play07:21

think that we're building

play07:22

gpus you know GP is a

play07:25

chip there are 72 chips in here and then

play07:28

there are six

play07:30

600,000 other

play07:32

parts

play07:33

it's 72 chips probably weighs one

play07:37

pound this is 3, 2,999 other

play07:42

pounds so the amount of Technology

play07:44

that's inside one of these RS is really

play07:46

quite extraordinary this is a technology

play07:48

Marvel the most most most complex most

play07:52

advanced computer the world's ever made

play07:54

yeah exactly the p in the world now yeah

play07:57

absolutely incredible and the software

play07:59

that it takes to run this is

play08:01

unbelievable yeah unbelievable isn't

play08:03

that right and so I think that that

play08:06

people now are starting to realize that

play08:08

when we say GPU server of course the

play08:11

brain is the GPU yeah but the system is

play08:14

much much more complex than that and

play08:15

super micro does amazing engineering

play08:18

thank

play08:20

[Laughter]

play08:27

you huh what I

play08:37

okay then we there's some Americans this

play08:40

year we are going to ship hopefully make

play08:44

when we're together sometimes we speak

play08:46

Taiwanese sometimes we speak Mandarin

play08:48

and then when we disagree we speak

play08:50

[Laughter]

play08:53

English we try to make a thlc mar share

play08:56

from 1% to 15%

play09:00

this year wow Save lot of power for your

play09:02

TB yeah yeah the Energy Efficiency is so

play09:05

much better the cost to the data center

play09:07

is cheaper cheaper that's right people

play09:09

don't realize this liquid cooled systems

play09:12

eliminates an enormous amount of cost in

play09:14

the data center yeah so that you can use

play09:17

that waste capture that waste and put it

play09:20

into Computing in the future in the

play09:23

future Computing throughput is

play09:26

revenues because it's token generation

play09:30

and token generation is dollars per

play09:34

million tokens just like

play09:37

energy dollars per kilowatt hour we have

play09:42

now invented a new commodity this is a

play09:44

very important idea for all of you this

play09:46

is a new commodity it has value and the

play09:50

faster you can generate it the higher

play09:53

throughput the greater utilization the

play09:56

higher your revenues it is absolutely

play09:59

true and it's directly measurable that's

play10:01

why this is a factory not a data center

play10:04

that's why this is a factory not a file

play10:06

server it's not a retrieval of files

play10:09

it's not used for exchanging emails this

play10:11

is directly generating revenues for

play10:14

factories that's why we call it AI

play10:16

factories and so

play10:18

powerful and only s million

play10:23

[Laughter]

play10:27

dollars a

play10:30

such okay so $3 million and you can

play10:34

generate who knows how much revenue per

play10:37

year right uh 3 million 1,000 and every

play10:42

year have how many

play10:44

months

play10:46

12 the the return on the return on large

play10:50

language model generation token

play10:52

generation is going to be very very good

play10:54

yeah be huge and the reason for that is

play10:56

because the token embeds intelligence

play10:58

yeah and the int could be used in so

play11:00

many different Industries and so the

play11:02

future is very important it's time to

play11:08

Startup yeah time to

play11:10

Startup throughput yeah

play11:15

utilization all matter so

play11:18

reliability has Revenue implication

play11:20

throughput has Revenue implication

play11:23

startup has Revenue implication yeah

play11:26

that's why it's so important that we

play11:27

integrate the whole s whole system into

play11:29

a rack scale get all the software

play11:32

working connected to all the all the

play11:34

networking so that and we build all of

play11:37

our own data centers we build our own

play11:39

supercomputers so that we know when you

play11:42

install this when you install super

play11:44

micro in your factories the startup time

play11:47

will be extremely fast your utilization

play11:50

will be extremely high and your

play11:52

throughput will be extremely high

play11:54

because your revenues depends on it

play11:57

Factory output is measured by all of

play12:00

those factors very complicated yeah and

play12:03

all of those R are Invidia sofware

play12:07

license all certified so the sound of

play12:10

that parking the cable and they can run

play12:13

and it runs that's right and all of the

play12:15

Nvidia Nims all of the large language

play12:17

models it just runs on all these systems

play12:19

yeah

play12:29

[Laughter]

play12:32

[Applause]

play12:37

we are shipping thousand R

play12:41

very

play12:52

yes very

play12:54

beautiful Charles Charles said that this

play12:57

is

play12:58

everything everything in here is NVIDIA

play13:01

for all the American citizens

play13:04

there

play13:06

from to H AI everything all Nvidia sare

play13:11

all all Green Computing all Green

play13:14

Computing all green computer all all

play13:17

support that's

play13:19

fantastic

play13:20

good let go through something detail

play13:29

okay okay

play13:32

okay H1 H2 B1 for you cooling wow

play13:39

shipping in B wow and this one your p200

play13:44

uhhuh fully ready beautiful for your

play13:46

chip beautiful beautiful this will be

play13:50

how many time faster than this so we

play13:53

have we have we have uh uh for Blackwell

play13:57

Blackwell has air cool

play14:00

liquor

play14:01

cooled

play14:03

x86

play14:05

Grace MV link 8 MV link 2 MV link 36 MV

play14:11

link 72 yeah so many different

play14:14

configurations yeah so that depending on

play14:16

the type of type of utilization type of

play14:19

use case you have the type of data

play14:21

center that you have uh Charles is ready

play14:23

to serve you immediately right

play14:25

immediately doesn't need to acheve yeah

play14:27

one hand we got to acheve second hand we

play14:30

Shi to C W thank goodness we only need

play14:33

two hands in two weeks in two

play14:39

weeks that's incredible and all of it

play14:41

software compatible this is really this

play14:44

is really the amazing thing certifi

play14:45

literally everything here is software

play14:47

compatible one% yeah and software as we

play14:50

know is the most complex part of high

play14:52

performance Computing yeah thank you for

play14:54

those great offering they are all ready

play14:57

to service our customer there are three

play14:59

very important software Stacks that we

play15:01

have in our company that everything is

play15:03

built on top of the first of course is

play15:05

Cuda very famous the second for all of

play15:07

the networking because networking is

play15:09

just not networking networking today

play15:12

networking today is a Computing

play15:15

fabric networking today is a Computing

play15:17

fabric not just for sending email to

play15:19

each

play15:20

other

play15:21

4 Mez a gigahertz megahertz this is not

play15:27

1980s

play15:35

be

play15:36

Mez

play15:38

kilohertz gahz gigahertz yes 400

play15:43

gigabits per second 800 gigabits per

play15:45

second and and then of course Next

play15:46

Generation coming 1600 but the important

play15:48

thing is all of the software that we

play15:51

have that runs on the networking for

play15:54

distributed computing is on top of two

play15:58

software Stacks one is called DOA for

play16:00

the nick nickel for the fabric yeah and

play16:04

it enables us to distribute the workload

play16:07

across the network very very efficiently

play16:10

because ethernet was was not designed

play16:12

for hyperform computing you make our job

play16:14

easier but still very py because you

play16:16

have so many

play16:19

great my job is to help give you

play16:23

[Laughter]

play16:27

job we

play16:29

and because because you do such a good

play16:31

job it becomes gives me job oh don't

play16:34

forget that your another

play16:40

baby yeah yeah yeah

play16:50

yeah inside

play16:53

here this this is an incredible

play16:57

incredible system in fact in

play17:00

fact in fact um these chips are all

play17:02

connected together using high-speed

play17:05

interconnect the world's fastest CIS the

play17:07

CIS is incredibly fast and very energy

play17:11

efficient and so we can connect this

play17:14

great CPU to dual Blackwell gpus and

play17:20

that's very important because in the

play17:22

training stage the memory system of

play17:26

Grace could be used for checkpoint

play17:27

restart checkpoint and restarting is

play17:30

very important for high utilization and

play17:32

high uptime and so checkpoint restart uh

play17:35

could be stored in the system memory

play17:37

that system memory is very low energy

play17:39

very low power and the link between

play17:42

Blackwell and Grace is very very high

play17:44

second during inference time as you know

play17:47

there's a concept called

play17:49

prompts context in context training

play17:54

prompting that prompt memory that

play17:56

context memory is right here this is the

play17:58

memory memory the thinking memory the

play18:00

working memory of AI and so this memory

play18:03

needs to be very high performance very

play18:05

low energy and so during training we

play18:07

have good use for gray gray CPU during

play18:10

inference we have excellent use for gray

play18:12

CPU and the interconnect is very very

play18:15

high speed very low power F optimiz and

play18:17

so the re the benefit is because we

play18:20

compress so many in one system yeah if

play18:23

we

play18:24

save 20 watts 50 Watts on the

play18:27

interconnect you multiply by the whole

play18:30

rack then we can take the energy and use

play18:32

it for computing y so Energy Efficiency

play18:36

translates to higher performance to

play18:39

that's right Green

play18:42

[Applause]

play18:45

Computing

play18:52

huh I am a super micro employee

play18:59

super micro

play19:06

employee where AI Control

play19:12

us of course not um we we have to we

play19:16

have to uh the most important thing of

play19:19

course at the moment is we have to make

play19:21

AI work

play19:23

well right now ai is of course uh

play19:26

working extremely well and in many

play19:28

applications AI has become good enough

play19:31

to good enough to become useful it has

play19:35

achieved the plateau of good enough very

play19:37

useful however we want it to be

play19:40

incredibly good we want it to be very

play19:42

functional everything from Guard railing

play19:45

for uh fine-tuning skill learning there

play19:48

are many different things that we still

play19:50

have to improve okay so we know that AI

play19:52

is AI still has long ways to go that's

play19:55

job number one is Advance the technology

play19:57

at the same time we have to advanced

play19:59

Safety technology as you know uh our the

play20:02

planes that we all flew on to come here

play20:05

has autopilot and autopilot is automatic

play20:08

technology in order for planes to be

play20:11

safe a great deal of Technology had to

play20:13

be invented to keep the plane safe yeah

play20:16

also practices to monitor the planes air

play20:19

traffic control other planes monitor the

play20:22

planes Pilots monitoring each other many

play20:24

different ways to keep uh AI uh keep

play20:28

autopilot safe in the future we'll do

play20:30

the same thing with AI there will be AIS

play20:32

that watch AIS there are people that

play20:33

watch AIS there's gu right guard rails

play20:36

that keep AI guard rail and so there's

play20:38

going to be a whole lot of different

play20:39

Technologies we need to create for

play20:41

safety technology for safety and then

play20:43

third of course we need to have good

play20:45

policies for safety good practices and

play20:47

good policies for safety talking about

play20:50

it is very important so that we can all

play20:52

remind each other that we have to do

play20:54

good science good engineering good

play20:57

business practice good policy practice

play20:59

good industrial practice all of those

play21:01

things has to advance so perfect

play21:03

strategy so the conclusion is one the

play21:07

more you buy the more

play21:10

safe the more you buy the more you safe

play21:12

the more you buy the more you safe

play21:15

yeah thank you Jas thank you so much

play21:18

good job thank you

play21:20

everybody thank you okay thank you thank

play21:25

you thank you thank you all right have a

play21:28

great

play21:29

thank you

play21:32

[Music]

Rate This

5.0 / 5 (0 votes)

相关标签
黄仁勋英伟达加速计算生成式AI数据中心能源效率计算性能液冷系统技术创新未来计算