Is the AI Boom Real?

Asianometry
23 Feb 202413:43

Summary

TLDRThe author traveled to Japan and the U.S. to learn about the AI chip industry. Much buzz surrounds Sam Altman raising huge funds for a new AI chip venture. The concept of 'scaling laws' suggests that with more data and compute, AI models like GPT continue improving. This drives massive investment. Nvidia faces competition but responds aggressively. Giants like Microsoft vertically integrate to cut costs. OpenAI's ChatGPT generated $2B revenue, but it's unclear if consumer demand can sustain the AI boom. The industry hopes Microsoft's Copilot subscription catches on. If not, some assumptions need rethinking. AI's main impact could just be improving ad targeting for giants like Google and Meta.

Takeaways

  • 😲 The chip venture Sam Altman is working on could involve up to $1-7 trillion in investments over several years, largely for data centers and manufacturing
  • 📈OpenAI's 'scaling laws' show that combining more data and compute leads to better AI model performance
  • 👀Nvidia is aggressively updating its AI chips to stay ahead of new competitors
  • 🚨Tech giants like Microsoft and Google want to cut costs by making their own custom AI chips
  • 🤔It's unclear if demand from mainstream consumers will sustain the massive investments expected in AI
  • 🙌ChatGPT is one of the fastest growing consumer tech products ever at over $2 billion annual revenue
  • 🔎All eyes are on Microsoft's Copilot subscription to indicate enterprise demand for AI
  • 💰AI promises automation that could greatly boost advertising revenue
  • 😕But AI may just make rich companies even richer, not necessarily transform industries
  • ✈The author plans more trips to learn about and discuss the AI landscape

Q & A

  • What was the purpose of the author's recent travel to Japan and the United States?

    -The purpose was part sight-seeing and part learning tour to understand more about the AI and AI chip boom landscape outside of Taiwan.

  • What figure is Sam Altman said to be seeking for investments related to AI chips?

    -Initially it was reported to be billions, but more recently a figure of $1-7 trillion was reported, likely referring to total investments over multiple years related to manufacturing, real estate, power, etc.

  • What concept is driving the massive investments forecasted in AI chips?

    -The 'scaling laws' concept that combining more data and compute leads to better AI model performance. This concept helped drive advances like GPT-3 and GPT-4.

  • How is Nvidia responding to increased competition in AI chips?

    -By aggressively increasing the pace of updates to their AI accelerator lineup, releasing new products as often as every 6 months.

  • How might vertical integration by tech giants like Microsoft impact Nvidia?

    -In the medium to long term, as tech giants develop more advanced custom AI chips, it could reduce Nvidia's market share. Nvidia aims to stay ahead by leading on performance and cost.

  • What evidence is there so far that AI products are driving real consumer demand?

    -There are some signs of promise like ChatGPT's reported $2 billion annualized revenue run rate, but it's still early. Microsoft's Copilot subscription will be an important test case.

  • Where else might AI have a major financial impact?

    -In advertising and automation, where AI could greatly boost revenues and efficiencies for major platforms like Google and Meta.

  • What does the author see as a disappointing possibility for AI?

    -That its biggest impact is automating and optimizing existing industries rather than enabling wholly new products and companies, making the rich richer rather than producing new wealth.

  • What approach does the author recommend for embedding AI capabilities?

    -Recommend embedding AI into existing successful products to maximize its impact, using AI as an enabling technology rather than a standalone consumer product.

  • What are the author's plans going forward regarding AI?

    -The author plans to make more frequent trips to the US and other regions to have interesting conversations with people in the AI space and keep abreast of the latest developments.

Outlines

00:00

🤔 The Trillion Dollar Question

Paragraph 1 discusses the recent news around Sam Altman raising massive investments for an AI chip venture, with figures ranging from billions to trillions of dollars over several years. It questions the feasibility of such a high investment level and compares it to the entire semiconductor industry revenue. The key theme is the uncertainty around the scale of investment needed to realize Altman's vision.

05:06

📈 Scaling Laws - The New Moore's Law?

Paragraph 2 introduces the concept of scaling laws that have driven advancements in AI like GPT. It notes the parallels between scaling laws and Moore's Law in driving industry roadmaps and investments. However, it also highlights arguments against the perpetual viability of scaling laws.

10:09

💸 Is the AI Boom Financially Sustainable?

Paragraph 3 questions whether the current AI boom, despite rapid progress, is financially sustainable in the long run. It examines early revenue data for ChatGPT and other potential indicator products. The conclusion is cautious optimism, noting AI's potential as an enabling technology, but uncertainty around transformative mainstream adoption.

Mindmap

Keywords

💡Scaling Laws

The Scaling Laws refer to the concept from a 2020 OpenAI paper that combining more data and compute leads to better AI model performance. This idea helped drive the development of large language models like GPT-3 and is a core rationale behind the massive investments being made in AI.

💡ChatGPT

ChatGPT is an AI chatbot launched by OpenAI in late 2022. It uses a large language model and has rapidly gained popularity. The Financial Times reported it already has a $2 billion annual revenue run rate, making it one of the fastest growing consumer tech products ever.

💡Nvidia

Nvidia is the leading supplier of AI/GPU chips. It faces competition from other chipmakers and tech giants developing custom AI chips. But Nvidia is responding aggressively with rapid new chip iterations, reminiscent of the GPU wars era.

💡Moore's Law

Moore's Law refers to the long-term trend of computing power doubling every 1-2 years. This drove massive investment in semiconductors to fuel industries like PCs and smartphones. The similarities between Moore's Law and Scaling Laws suggest AI could also fuel huge economic growth.

💡Microsoft

Microsoft is emerging as the tech giant most aggressively pursuing vertical integration in AI, with custom chips and data center gear. This could cut costs but also suggests AI industry growth is slowing if giants now focus on pricing.

💡ASICs

ASICs refers to Application-Specific Integrated Circuits - customized chips designed by large tech companies to save costs over general purpose chips like GPUs. The push towards ASICs highlights the tech giants' incentive to cut suppliers like Nvidia out.

💡Advertising

Some argue AI's economic impact could be most realized by supercharging digital advertising, which already generates tens of billions in revenue. AI-powered ads may be the "killer app" enriching the giants further.

💡Vertical Integration

Vertical integration refers to tech giants like Microsoft controlling more of the AI tech stack in-house, e.g. custom chips, data center gear. This saves costs but may signal slowing industry growth.

💡Consumer Demand

Despite rapid AI progress, consumer demand remains unclear beyond ChatGPT subscriptions. Historical examples suggest AI investment likely requires public enthusiasm to be sustainable long-term.

💡Killer App

It's debated if AI itself is the next "killer app" fueling broad economic growth, or merely an enabling technology making existing apps and giants richer. Lack of clear consumer use cases beyond ChatGPT feeds this skepticism.

Highlights

Sam Altman and his team have talked to TSMC, people at Abu Dhabi, and the US Government about a trillion dollar chip venture

The trillion dollar figure may be a negotiating tactic to set high expectations for future talks

The scaling laws concept is driving massive investments - more data and compute leads to better AI model performance

The scaling laws could have a similar rallying effect in AI to what Moore's Law had on the semiconductor industry

Nvidia is aggressively updating their AI accelerators to fend off startup competitors

Nvidia ships GPU designs before fully testing them using advanced modeling tools

Tech giants like Google and Microsoft pose the biggest threat to Nvidia through custom ASICs

Microsoft is pushing hard on vertical integration for inference chips and networking

It's unclear if the AI boom is financially sustainable without compelling consumer products

ChatGPT grew quickly, but few products have development and operating costs like AI

All eyes are on Microsoft's Copilot subscription to indicate consumer demand

The killer app for AI could just be improving ad targeting and sales

If AI mainly helps the wealthy get richer, it may not justify massive investments

AI could have more impact by supercharging existing products versus creating new ones

Ads enhanced by AI automation generate tens of billions in revenue already

Transcripts

play00:02

For the past three weeks I have been  traveling through Japan and the United States.

play00:07

This trip has been part sight-seeing  and part learning tour with the goal  

play00:11

of understanding more about the AI and  AI chip boom landscape outside of Taiwan.

play00:17

Now that I am finished with the big conversations,  I wanted to sit down and share a few thoughts.

play00:23

## The Trillion

play00:23

There’s been a lot of news about a  chip venture thing that Sam Altman  

play00:26

of OpenAI is working on for the AI industry.

play00:30

The news came slowly. First there was  the report - released at around the  

play00:34

time he was temporarily ousted from  OpenAI - that he was raising money  

play00:38

from Middle East investors for some chip  venture. Back then it was just "billions".

play00:44

Since then, you have also gotten news that  Sam Altman and his team have talked to TSMC,  

play00:49

people at Abu Dhabi, and the US Government.

play00:52

Just as I was leaving Tokyo for San  Francisco, the WSJ broke the report with  

play00:56

that eye catching $1-7 trillion figure that  has got everyone buzzing. It was one of the  

play01:02

first things that I started asking people  about when I arrived in Silicon Valley.

play01:07

Many people agreed with me that the  number is a bit too much. Perhaps  

play01:10

it can be a negotiating tactic to set  expectations for future talks. Once you  

play01:15

start talking trillions then hundreds of  billions no longer feels as substantial.

play01:21

Then you have the Information adding a bit  of context to Altman's remarks, saying:

play01:25

> But in reality, Altman privately has told  people that figure represents the sum total of  

play01:30

investments that participants ... would need  to make, in everything from real estate and  

play01:35

power for data centers to the manufacturing  of the chips, over some period of years.

play01:41

The ecosystem concept makes a lot more sense.  Total semiconductor sales in 2023 were about $520  

play01:47

billion. Total capital expenditures - according  to the trade group SEMI - were about $140 billion.

play01:55

But even if we were to assume that this  trillion is spread out over five years  

play01:59

and diluted with real estate expenses and power,  

play02:02

that is still another COVID-like step  function upwards in capital expenditure.

play02:08

The semiconductor industry is a conservative  one. Many of the people there are old-heads  

play02:12

who have seen many a cycle of booms and busts  downstream. Sam isn’t really the first guy to  

play02:17

knock on a foundry’s door saying that his  use case is going to change the world.

play02:23

## Scaling Laws

play02:24

The concept driving this investment is  something called the "scaling laws".

play02:27

The name dates to a 2020 paper  posted by OpenAI titled "Scaling  

play02:32

laws for neural language models". I am  not going to go over the fine details,  

play02:36

but the gist is that if we combine more data and  compute, we get better results (i.e. less loss).

play02:42

This was one of the core ideas that  helped OpenAI make the GPT-series a  

play02:46

reality. Ilya Sutskever, one  of the company's cofounders,  

play02:50

mentioned something like this in an appearance  on the No Priors podcast back in November:

play02:56

> I was very fortunate in that I was able to  realize that the reason neural networks of  

play03:00

the time weren't good is because they are too  small. So like if you tried to solve a vision  

play03:05

task with a neural network with a thousand  neurons, what can it do? It can't do anything.

play03:10

> It doesn't matter how good your  learning is and anything else.  

play03:13

But if you have a much larger network  then it can do something unprecedented.

play03:18

GPT-3 is big. GPT-4 was even bigger, performing  far better. And GPT-5, whenever it comes out, will  

play03:25

be even bigger. So far there are no indications  that the scaling laws have broken apart.

play03:34

I find the parallels with the semiconductor  

play03:36

industry's Moore's Law very interesting.  The two "laws" do not say the same thing,  

play03:41

but they might have similar effects  on their respective industries.

play03:45

In the 1980s and 1990s, Moore's Law became  the rallying cry of the entire semiconductor  

play03:50

industry - the metronome keeping time for every  company from Santa Clara to Tokyo to Hsinchu.

play03:56

There is a chance that the Scaling Laws can make  a similar impact on the AI industry. A simple set  

play04:02

of easily understandable rallying cries that  drive R&D roadmaps for whatever years to come.

play04:09

There are a variety of arguments against  the Scaling Laws. For instance, people have  

play04:13

commented that we have basically pulled all the  existing data across the entirety of the Internet.

play04:18

But there are ways around such things, if the  money is there to make it happen. Physics problems  

play04:22

have cropped up ahead of Moore's Law, and the  industry derived new engineering solutions:  

play04:27

the High-K Metal Gate, FinFET, DUV Lithography,  Dry Etch, Ion Implantation, SEM, and so on.

play04:35

The bigger question is whether  the money is there to continue  

play04:38

driving this investment. More on that later.

play04:41

For a way deeper review of the  technical issues behind scaling,  

play04:45

I recommend Dwarkesh Patel's post - "Will scaling  work?" His podcast is quite excellent too.

play04:51

## AI Chips & Nvidia

play04:51

Where there are profits, there are competitors  coming out with ideas to take them.

play04:55

There has been a lot of ink spilled on Nvidia  and their competition - particularly the  

play05:00

multi-pronged attack on Nvidia's AI accelerator  profits. I don't want to add too much to this.

play05:06

I do question whether the Nvidia fortress is going  

play05:09

to be as assailable as it might at first  seem. Jensen is responding to the threat  

play05:14

by aggressively ramping up the annual  updates to their accelerator lineup.

play05:19

For more information on this, I refer you to  

play05:21

the report from SemiAnalysis. It goes  into much more detail. Dylan's great.

play05:27

But this relentless pace reminds me of the days  of old Nvidia during the Graphics Cards Wars,  

play05:31

when Nvidia released a new product  every six months to the market. It  

play05:35

worked for them then. Why not try  it again now for the AI Chip Wars?

play05:40

Nvidia can maintain this rollout  speed because - as Jensen implied  

play05:43

during his appearance at the Acquired  podcast - they ship before they test.

play05:48

They use the latest computer software  design and emulation tools to model  

play05:51

and ship new GPU designs to the market  without first fabbing a physical chip:

play05:56

> The reason why we needed that emulator is  because if you figure out how much money that  

play06:01

we have, if we taped out a chip and we got  it back from the fab and we started working  

play06:05

on our software, by the time that we found  all the bugs because we did the software,  

play06:10

then we taped out the chip again. We  would’ve been out of business already.

play06:14

The Nvidia teams will try to tape  out "perfect chips" as Huang said,  

play06:18

but this inevitably will cause problems  for customers. For instance, deployments  

play06:22

with drivers and other affiliated software  that won't work all that well at first.

play06:28

But the Nvidia brand can take a hit like that,  

play06:30

whereas it is unlikely that the AI chip  startups will be able to. It takes several  

play06:35

years to build a team and then get that  team good enough to ship a working product.

play06:40

This high speed of iteration will be rough  on customers. Many have no choice but to buy  

play06:46

what is available, but it stings to spend  tens of thousands of dollars on a machine,  

play06:50

only to have a vastly better one  coming out just a short time later.

play06:55

## Giants and Verticalization Nvidia is less concerned about the startups and

play06:57

even the established silicon players  than they are about the tech giants.

play07:01

The tech giants - Microsoft, Google, and the like  - are the ones driving the current investment  

play07:06

spend in AI today. They are also the ones with the  most incentive to cut Nvidia out of their margin.

play07:13

They would do this using custom-designed  chips or ASICs. Large companies have  

play07:17

sought these since the good old days of  the computer - commissioning an ASIC to  

play07:21

replace an entire board of discrete  electronic components to save money.

play07:25

An example of this would be the Apple IIe,  

play07:28

a third-generation version of the  Apple II PC powered by a custom ASIC.

play07:34

A more modern example would be the Google TPU,  which right now is in its fourth iteration.

play07:39

It makes up a big part of Google's compute  advantage over OpenAI and the other AI labs.

play07:45

Microsoft seems to be the giant pushing hardest  on vertical integration. CEO Satya Nadella said  

play07:51

in the Q2 2024 earnings call that most of  the usage in Azure AI services is coming from  

play07:57

inference rather than training. Those activities  are easier to push to a custom-designed ASIC.

play08:04

In addition, Microsoft has been working  on vertically integrating other parts  

play08:07

of the database stack. For instance,  the Information's recent report that  

play08:11

they are working on a network card  to shuttle data between servers.

play08:16

This type of vertical integration implies either  that the scale of the AI database stack is so  

play08:20

large and costly that every penny counts/will  count. Or that the growth in the industry has  

play08:26

petered out, leaving players to compete on price.  These scenarios both feel weird to consider.

play08:33

What does this vertical integration trend  mean for Nvidia? It will take some time  

play08:37

for the other tech giants to ramp up their AI  chip designs. Apparently the first Google TPU  

play08:42

was a very bare-bones product. So in the  short term, things will be as they are.

play08:48

But as those chips get better in the medium  to long term, it makes sense that Nvidia  

play08:52

push as hard as they can to always win  the performance-cost crown. Who knows.

play08:56

## Financially Sustainable?

play08:58

I want to move away from cost. I think the cost  benefits of vertical integration are significant  

play09:03

enough to matter. Now, I want to move to  the other side of the financial equation.

play09:09

One of the really big questions that I wanted to  answer coming to the United States was to learn  

play09:13

more about whether or not this OpenAI  boom is Real. Real, with a capital R.

play09:19

There has been so much progress in the AI  industry in the past year. For instance, the  

play09:23

OpenAI text to video Sora model. The improvement  in the models' output quality has been impressive.

play09:30

But even if the technology is amazing, it does  not seem like a hit product by itself. We need  

play09:36

to embed them into compelling products. Are  there indications of such products breaking  

play09:41

through into the public? Who is making  money from actual consumers with them?

play09:46

The massive investments required by Moore's  Law and the semiconductor industry were  

play09:51

driven by real demand from various downstream  industries. First the military, then consumer  

play09:56

electronics like radios and calculators, then  the PC, smartphones, and then cloud computing.

play10:04

These were all things that  ordinary people wanted. To me,  

play10:08

it feels unlikely that the truly  large investments in AI will happen  

play10:13

unless those ordinary consumers start  buying these services in a major way.

play10:18

The Financial Times did report that ChatGPT hit a  

play10:21

$2 billion run-rate revenue in December  2023. That is basically the only solid  

play10:27

piece of news indicating that people are  paying real money for an LLM service.

play10:32

Now, we should note that it is still early in the  

play10:35

generative AI boom. ChatGPT  is a little over a year old.

play10:39

But I also want to note that the  iPhone was first released in June 2007.

play10:44

About two years later in 2009,  

play10:46

Apple sold 20 million iPhones for like  $13 billion in revenue. Growth continued,  

play10:51

with 40 million iPhones sold the year after that,  and then 72 million and 125 million after that.

play10:59

Okay fine, the iPhone is the  greatest consumer technology  

play11:02

product that history has ever seen. It  is a difficult example to live up to.

play11:07

ChatGPT is still one of the fastest  growing consumer tech products in history,  

play11:11

depending on what you think about  Threads. And people are still  

play11:15

using it a lot - Altman recently said that  ChatGPT generates 100 billion words a day.

play11:21

And Facebook took a while to produce revenue  - at first investing to build up the audience.  

play11:26

Perhaps ChatGPT is doing the same Aggregator  approach, gathering users to monetize later.  

play11:32

But few products cost as much to use. AI  probably has shorter a leash than we think.

play11:39

Other than ChatGPT, the one product  that everyone in the industry seems  

play11:42

to have their eye on is the Microsoft  Copilot subscription service. We want  

play11:46

to see if this $20 a month offering  catches on with people and enterprises.

play11:52

If it does, then we are really off to the races  ... the sky really is the limit ... we are so back  

play11:57

... and so on, insert bombastic metaphor here. If  not, then we have to adjust some mental models.

play12:04

## Conclusion

play12:04

There is one last thing that I  should mention. Maybe where the  

play12:07

LLM revolution will be most Real is when it  is embedded into the products we know today.

play12:13

For instance, a few people I spoke  to insisted that I am overlooking the  

play12:17

impact that AI automation will have  in supercharging advertising sales.

play12:22

Like Google's Performance Max ads, which  use AI to automate the creation, deployment,  

play12:27

and targeting of new ads. Ads like Performance Max  apparently generate tens of billions of dollars.

play12:33

Or how when Apple disabled cookie tracking in  2021 with the App Tracking Transparency thing.  

play12:39

Meta/Facebook at first said that this would  cost the company over $10 billion in sales.

play12:46

But over time, Meta somehow managed to claw  back its targeting accuracies using AI,  

play12:51

collecting data from its Conversions API.

play12:54

So maybe the killer app for AI is just more ads.

play12:59

But Ben Thompson and others have been saying for a  

play13:01

while that AI is an enabling technology -  a technology for making the rich richer,  

play13:05

rather than making a whole new class of rich  people like the PC or electrification was.

play13:11

So for me, if this is all that it is,  

play13:14

then it is a bit disappointing. But it  makes the AI boom nevertheless real.

play13:20

Anyway, these are my reflections from my trip to  the United States. I hope to make more frequent  

play13:24

trips in the future and have more interesting  conversations with people in the space.

play13:29

If you are interested in having a chat,  

play13:30

shoot me an email. Maybe I will be coming to  town again and can have that chat in person.

Rate This

5.0 / 5 (0 votes)

Benötigen Sie eine Zusammenfassung auf Englisch?