Nvidia 2024 AI Event: Everything Revealed in 16 Minutes
Summary
TLDRスクリプトの魅力的な要約で、ユーザーの興味を引き付ける短くyet正確な概要を提供します。
Takeaways
- 🌐 ブラックウェルは新しいプラットフォームで、GPUを製造しているという一般的な認識と異なります。
- 🔄 ブラックウェルチップは280億トランジスタを持っており、2つのダイが密接して一つのチップのように機能します。
- 🚀 10TBのデータが1秒間に伝送され、2つのブラックウェルチップの側は自分がどちらの側にあるか認識しません。
- 💻 Hopperに互換性のあるブラックウェルのバージョンがあり、既存のHopperのインフラストラクチャに簡単に取り入れることができます。
- 🌟 ブラックウェルはメモリーの統合性があり、計算が非常に効率的に行われます。
- 🛠️ 新しいチップMVY Link Switchは500億トランジスタを持っており、Hopperのサイズに近いものです。
- 🔗 MVY Link Switchは4つのMVリンクを持ち、1.8TBの速度で他のGPUと通信できます。
- 🤖 ジェネラティブAI時代のためのプロセッサが開発され、コンテンツトークン生成という重要な機能が含まれています。
- 📈 計算が進化する速度は非常に速いにも関わらず、まだ十分に迅速ではないため、さらなる Chips の開発が行われています。
- 🔄 パートナーシップはAIの時代に向けて製品やサービスを加速するための重要な要素です。
- 🌐 ブラックウェルは世界中の多数の企業と協力し、AIファクトリーの概念を導入し、産業に革新をもたらす予定です。
Q & A
Blackwellはどのようなプラットフォームですか?
-Blackwellは、GPUを制作しているという誤解を招く名前のプラットフォームです。実際には、Blackwellは28億トランジスタを持つ非常に高性能なチップで、新しいタイプのコンピューティングデバイスを提供します。
Blackwellチップの特徴は何ですか?
-Blackwellチップは、2つのダイ(dies)が密接に接続され、10TBのデータが1秒間にやり取りされることで、メモリの局所性やキャッシュの問題が発生しないという特徴があります。これは、2つの側面が自分がどの側面にあるかわからないほど、一つの巨大なチップのように動作します。
Blackwellチップはどのようなシステムに適用されますか?
-Blackwellチップは、2つのタイプのシステムに適用されます。一つは、Hopperとフォームフィット機能互換性のあるもので、HopperをスライドしてBlackwellをプッシュインします。もう一つは、メモリーコヒーレントであり、複数のGPUが同時にフルスピードで通信できるように設計されたシステムです。
MVYリンクスイッチチップは何特筆すべき特性がありますか?
-MVYリンクスイッチチップは、Hopperのほぼ同じサイズで、50億トランジスタを持っており、4つのMVリンクを持ち、それぞれ1.8TBの速度でデータを転送できるという特筆すべき特性があります。また、コンピューティングも含まれており、各GPUが他のGPUと同時に通信できるシステムを構築することができます。
NvidiaはAI時代にどのようなプロセッサを作成しましたか?
-Nvidiaは、AI時代に生成のための新しいプロセッサを作成しました。このプロセッサの最も重要な部分の1つは、コンテンツトークン生成であり、彼らはこれをFP4フォーマットと呼んでいます。
Nvidiaが提供するAIファブリックとは何ですか?
-NvidiaのAIファブリックは、AIを構築するための3つの柱から成り立っています。これには、Nvidia Inference Microservice(NIMS)、Nemo Microservice、DGX Cloudが含まれます。これらのサービスは、AIの学習、微調整、パフォーマンス評価、そしてAPIを使用した簡単なデータの準備とキュレーションを可能にします。
Nvidiaはどのようなパートナーシップを発表しましたか?
-Nvidiaは、SAP、cohesity、snowflake、NetApp、Dell、AWS、Google、Oracle、Microsoftなど、世界中の偉大な企業とのパートナーシップを発表しました。これらのパートナーシップは、AIファブリックの構築、データ処理、ロボット学習、デジタルツインの作成など、多岐にわたるイニシアチブに関連しています。
NvidiaのOmniverseとは何ですか?
-NvidiaのOmniverseは、ロボットやAIエージェントが学習し、複雑な工業空間での予測不可能な出来事をナビゲートするためのデジタルな世界を代表するシミュレーションエンジンです。Omniverseは、設計、エンジニアリング、マーケティングなど、企業の異なる部門が同じデータの真実に基づいて作業できるようにするプラットフォームです。
NvidiaのIsaac Simは何ですか?
-NvidiaのIsaac Simは、ロボット学習アプリケーションです。このアプリケーションを使用すると、グループモデルをトレーニングし、ロボットは人間の示範から学び、日常のタスクを支援するように調整できます。
NvidiaのJetson Thorロボットチップはどのようなものです?
-NvidiaのJetson Thorは、AIパワードのロボットCHIPS(CARs)のためのロボットチップです。Isaac Lab、Osmo、Grootを使用することで、このチップは来世代のAIパワードのロボットの基盤を提供します。
NvidiaのProject Grootは何ですか?
-NvidiaのProject Grootは、人型ロボット学習のための一般的な基礎モデルです。このグループモデルは、マルチモーダルな指示と過去の相互作用を入力として取り込み、ロボットが実行する次のアクションを生成します。
Outlines
🚀 革新的なBlackwellチップとMVYリンクスイッチの紹介
この段落では、Blackwellという新しいプラットフォームが紹介されています。Blackwellは28億トランジスタを持つチップで、2つのダイが新しい方法で接続されており、10TBのデータが1秒間にやり取りされています。Blackwellチップは2つのシステムに適用され、1つはHopperに滑り込み、Blackwellをプッシュインできる形で、効率性が非常に高いため、導入が課題となります。また、MVYリンクスイッチという新しいチップも紹介されており、50億トランジスタでHopperのサイズに近いものです。このスイッチは1.8TBの速度で4つのMVリンクを持ち、全てのGPUが同時にフルスピードで通信できるシステムを構築することができます。
🤖 AI時代のためのNvidiaのパートナーシップとサービス
この段落では、NvidiaがAI時代に対応するための様々なパートナーシップとサービスについて説明しています。Google、AWS、Oracle、Microsoftなどの大企業と協力しており、AIサービスやデータ処理、ロボット技術を加速しています。また、NvidiaがAIファウンダリーとして機能し、企業のAIニーズに応えることが強調されています。Nems、Nivo、dgx Cloudという3つの柱が紹介され、SAPやcohesity、snowflake、net appppなどの企業と協力してAIエージェントを構築しています。
🌐 デジタルツインとOmniverseによるロボットの学習と生産性向上
この段落では、デジタルツインとOmniverseというNvidiaのシミュレーションエンジンによるロボットの学習と生産性向上について説明されています。Omniverseはロボットが学習するための仮想世界であり、ovxというコンピュータがAzure Cloudでホストされています。産業界ではデジタルツインから始まり、AIエージェントがロボット労働者やインフラストラクチャを助け、複雑な工業空間での予測不可能なイベントをナビゲートします。Nvidia Project Grootという一般的な目的のヒューマノイドロボット学習モデルも紹介され、Isaac Labのアプリケーションを使用してOmniverse上でトレーニングされています。
🎉 Blackwellチップの発表とNvidiaの将来のビジョン
最後の段落では、Blackwellチップの発表がハイライトされ、このチップがNvidiaのビジョンにおけるGPUの姿を示していると強調されています。Blackwellは驚くべきシステム設計であり、MVYリンクスイッチやネットワークシステムが含まれています。この発表はNvidiaの将来のビジョンを示すものであり、コンピュータグラフィックス、物理学、人工知能の交差点であるNvidiaの魂を表しています。
Mindmap
Keywords
💡Blackwell
💡Hopper
💡transistors
💡memory coherent
💡MVY link switch
💡dgx system
💡AI Foundry
💡Omniverse
💡Jetson Thor
💡Nemo microservice
💡dgx Cloud
Highlights
Arrival at a developers conference with a focus on science, algorithms, and computer architecture.
Introduction of Blackwell, a platform that changes the traditional concept of GPUs.
Hopper, a revolutionary chip with 28 billion transistors, setting a new standard in computing.
The unique design of Blackwell allows two dies to function as one, with 10 terabytes of data transfer per second.
Form fit function compatibility with Hopper, making Blackwell a seamless integration.
The challenge of ramping up to the efficiency of Blackwell due to its innovative design.
The creation of a processor for the generative AI era, emphasizing content token generation with FP4 format.
The incredible advancement in computing, yet still striving for greater speed.
Introduction of the MVY link switch chip with 50 billion transistors and a capacity of 1.8 terabytes per second.
A system that allows every GPU to communicate at full speed simultaneously, revolutionizing AI systems.
The partnership with major companies like AWS, Google, and Microsoft in accelerating AI services and systems.
Nvidia's role as an AI Foundry, providing comprehensive solutions for AI development.
Collaboration with SAP to build service now assist virtual assistance using Nvidia Nemo and dgx Cloud.
The development of Cohesity's generative AI agent with Nvidia AI Foundry, leveraging vast data resources.
Snowflake's partnership with Nvidia AI Foundry to build co-pilots, enhancing their cloud data services.
NetApp's utilization of Nvidia AI Foundry to build chatbots and co-pilots, improving data storage solutions.
Dell's role in building AI factories for enterprises, partnering with Nvidia for large-scale systems.
Omniverse as a digital twin simulation engine for AI agents and robots, hosted in the Azure Cloud.
Nvidia Project Groot, a foundation model for humanoid robot learning, showcasing the intersection of computer graphics, physics, and AI.
The Jetson Thor robotics chips, designed for the future of AI-powered robotics, demonstrated by Disney's BDX robots.
Transcripts
I hope you realize this is not a
concert you have
arrived at a developers
conference there will be a lot of
science
described algorithms computer
architecture mathematics Blackwell is
not a chip Blackwell is the name of a
platform uh people think we make
gpus and and we do but gpus don't look
the way they used to this is hopper
Hopper changed the
world this is
Blackwell it's okay
Hopper 28 billion transistors and so so
you could see you I can see there there
a small line between two dyes this is
the first time two dieses have abutted
like this together in such a way that
the two CH the two dies think it's one
chip there's 10 terabytes of data
between it 10 terabytes per second so
that these two these two sides of the
Blackwell Chip have no clue which side
they're on there's no memory locality
issues no cach issues it's just one
giant chip and it goes into two types of
systems the first
one is form fit function compatible to
Hopper and so you slide a hopper and you
push in Blackwell that's the reason why
one of the challenges of ramping is
going to be so efficient there are
installations of Hoppers all over the
world and they could be they could be
you know the same infrastructure same
design the power the electricity The
Thermals the software identical push it
right back and so this is a hopper
version for the current hgx
configuration and this is what the other
the second Hopper looks like this now
this is a prototype board this is a
fully functioning board and I just be
careful here this right here is I don't
know10
billion the second one's
five it gets cheaper after that so any
customer in the audience it's okay the
gray CPU has a super fast chipto chip
link what's amazing is this computer is
the first of its kind where this much
computation first of all fits into this
small of a place second it's memory
coherent they feel like they're just one
big happy family working on one
application together we created a
processor for the generative AI era and
one of the most important important
parts of it is content token generation
we call it this format is fp4 the rate
at which we're advancing Computing is
insane and it's still not fast enough so
we built another
chip this chip is just an incredible
chip we call it the mvy link switch it's
50 billion transistors it's almost the
size of Hopper all by itself this switch
ship has four MV links in
it each 1.8 terabytes per
second
and and it has computation in it as I
mentioned what is this chip
for if we were to build such a chip we
can have every single GPU talk to every
other GPU at full speed at the same time
you can build a system that looks like
this
now this system this
system is kind of
insane this is one dgx this is what a
dgx looks like now just so you know
there only a couple two three exop flops
machines on the planet as we speak and
so this is an exif flops AI system in
one single rack I want to thank I want
to thank some partners that that are
joining us in this uh aw is gearing up
for Blackwell they're uh they're going
to build the first uh GPU with secure AI
they're uh building out a 222 exif flops
system we Cuda accelerating Sage maker
AI we Cuda accelerating Bedrock AI uh
Amazon robotics is working with us uh
using Nvidia Omniverse and Isaac Sim AWS
Health has Nvidia Health Integrated into
it so AWS has has really leaned into
accelerated Computing uh Google is
gearing up for Blackwell gcp already has
A1 100s h100s t4s l4s a whole Fleet of
Nvidia Cuda gpus and they recently
announced the Gemma model that runs
across all of it uh we're work working
to optimize uh and accelerate every
aspect of gcp we're accelerating data
proc which for data processing the data
processing engine Jacks xlaa vertex Ai
and mujo for robotics so we're working
with uh Google and gcp across whole
bunch of initiatives uh Oracle is
gearing up for blackw Oracle is a great
partner of ours for Nvidia dgx cloud and
we're also working together to
accelerate something that's really
important to a lot of companies Oracle
database Microsoft is accelerating and
Microsoft is gearing up for Blackwell
Microsoft Nvidia has a wide- ranging
partnership we're accelerating could
accelerating all kinds of services when
you when you chat obviously and uh AI
services that are in Microsoft Azure uh
it's very very very likely nvidia's in
the back uh doing the inference and the
token generation uh we built they built
the largest Nvidia infiniband super
computer basically a digital twin of
ours or a physical twin of ours we're
bringing the Nvidia ecosystem to Azure
Nvidia DJ's Cloud to Azure uh Nvidia
Omniverse is now hosted in Azure Nvidia
Healthcare is in Azure and all of it is
deeply integrated and deeply connected
with Microsoft fabric a NM it's a
pre-trained model so it's pretty clever
and it is packaged and optimized to run
across nvidia's install base which is
very very large what's inside it is
incredible you have all these
pre-trained stateof the open source
models they could be open source they
could be from one of our partners it
could be created by us like Nvidia
moment it is packaged up with all of its
dependencies so Cuda the right version
cdnn the right version tensor RT llm
Distributing across the multiple gpus
tried and inference server all
completely packaged together it's
optimized depending on whether you have
a single GPU multi- GPU or multi- node
of gpus it's optimized for that and it's
connected up with apis that are simple
to use these packages incredible bodies
of software will be optimized and
packaged and we'll put it on a
website and you can download it you
could take it with you you could run it
in any Cloud you could run it in your
own data Center you can run in
workstations if it fit and all you have
to do is come to ai. nvidia.com we call
it Nvidia inference microservice but
inside the company we all call it Nims
we have a service called Nemo
microservice that helps you curate the
data preparing the data so that you
could teach this on board this AI you
fine-tune them and then you guardrail it
you can even evaluate the answer
evaluate its performance against um
other other examples and so we are
effectively an AI Foundry we will do for
you and the industry on AI what tsmc
does for us building chips and so we go
to it with our go to tsmc with our big
Ideas they manufacture and we take it
with us and so exactly the same thing
here AI Foundry and the three pillars
are the NIMS Nemo microservice and dgx
Cloud we're announcing that Nvidia AI
Foundry is working with some of the
world's great companies sap generates
87% of the world's global Commerce
basically the world runs on sap we run
on sap Nvidia and sap are building sap
Jewel co-pilots uh using Nvidia Nemo and
dgx Cloud uh service now they run 80 85%
of the world's Fortune 500 companies run
their people and customer service
operations on service now and they're
using Nvidia AI Foundry to build service
now uh assist virtual
assistance cohesity backs up the world's
data their sitting on a gold mine of
data hundreds of exobytes of data over
10,000 companies Nvidia AI Foundry is
working with them helping them build
their Gia generative AI agent snowflake
is a company that stores the world's uh
digital Warehouse in the cloud and
serves over three billion queries a day
for 10,000 Enterprise customers
snowflake is working with Nvidia AI
Foundry to build co-pilots with Nvidia
Nemo and Nims net apppp nearly half of
the files in the world are stored on
Prem on net app Nvidia AI Foundry is
helping them uh build chat Bots and
co-pilots like those Vector databases
and retrievers with enidan Nemo and
Nims and we have a great partnership
with Dell everybody who everybody who is
building these chatbots and generative
AI when you're ready to run it you're
going to need an AI Factory
and nobody is better at Building
endtoend Systems of very large scale for
the Enterprise than Dell and so anybody
any company every company will need to
build AI factories and it turns out that
Michael is here he's happy to take your
order we need a simulation
engine that represents the world
digitally for the robot so that the
robot has a gym to go learn how to be a
robot we call that
virtual world Omniverse and the computer
that runs Omniverse is called ovx and
ovx the computer itself is hosted in the
Azure Cloud the future of heavy
Industries starts as a digital twin the
AI agents helping robots workers and
infrastructure navigate unpredictable
events in complex industrial spaces will
be built and evaluated first in
sophisticated digital twins once you
connect everything together it's insane
how much productivity you can get and
it's just really really wonderful all of
a sudden everybody's operating on the
same ground
truth you don't have to exchange data
and convert data make mistakes everybody
is working on the same ground truth from
the design Department to the art
Department the architecture Department
all the way to the engineering and even
the marketing department today we're
announcing that Omniverse
Cloud streams to The Vision Pro and
it is very very
strange that you walk around virtual
doors when I was getting out of that
car and everybody does it it is really
really quite amazing Vision Pro
connected to Omniverse portals you into
Omniverse and because all of these cat
tools and all these different design
tools are now integrated and connected
to Omniverse
you can have this type of workflow
really
incredible this is Nvidia Project
Groot a general purpose Foundation model
for humanoid robot
learning the group model takes
multimodal instructions and past
interactions as input and produces the
next action for the robot to
execute we developed Isaac lab a robot
learning application to train Gro on
Omniverse Isaac
Sim and we scale out with osmo a new
compute orchestration service that
coordinates workflows across djx systems
for training and ovx systems for
simulation the group model will enable a
robot to learn from a handful of human
demonstrations so it can help with
everyday
tasks and emulate human movement just by
observing us all this incredible
intelligence is powered by the new
Jetson Thor robotics chips designed for
Gro built for the future with Isaac lab
osmo and Groot we're providing the
building blocks for the next generation
of AI powered
[Applause]
[Music]
robotics
about the same
size the soul of
Nvidia the intersection of computer
Graphics physics artificial intelligence
it all came to bear at this moment the
name of that project general robotics
003 I know super
good
super
good well I think we have some special
guests do
[Music]
we hey
guys so I understand you guys are
powered by
Jetson they're powered by
Jetson little Jetson robotics computer
inside they learn to walk in Isaac
Sim ladies and gentlemen this this is
orange and this is the famous green they
are the bdx robots of
Disney amazing Disney
research come on you guys let's wrap up
let's
go five things where you going
I sit right
here Don't Be Afraid come here green
hurry
up what are you
saying no it's not time to
eat it's not time to
eat I'll give I'll give you a snack in a
moment let me finish up real quick
come on green hurry up stop wasting
time this is what we announce to you
today this is Blackwell this is the
plat amazing amazing processors MV link
switches networking systems and the
system design is a miracle this is
Blackwell and this to me is what a GPU
looks like in my mind
Browse More Related Video
Accenture CEO on earnings beat and revenue cut, spending and generative AI
コンテンツには5種類ある3 実際にわっきーのコンテンツをその視点で見る
This Midjourney Update is WILD + Pika's New AI Video Feature
ReAct Agent (Part 1, Introduction to Agents)
The Problem with Human Specialness in the Age of AI | Scott Aaronson | TEDxPaloAlto
A Comprehensive Cookbook for Claude 3
5.0 / 5 (0 votes)