New STUNNING Research Reveals AI In 2030...

TheAIGRID
2 Sept 202423:19

Summary

TLDREpoch AI's report predicts significant advancements in AI, suggesting models like GPT-5 could generate over $2 billion in revenue within a year of release. The report highlights the potential for AI to automate a substantial portion of the $60 trillion global economy, with companies investing trillions in AI development. It also discusses the possibility of training runs increasing by 5,000 times by 2030, the exploration of synthetic data to overcome data scarcity, and the race for gigawatt-scale data centers. The video emphasizes the conservative yet impressive estimates for AI's future, indicating an 'incredible time' ahead in the field.

Takeaways

  • 🔮 Epoch AI's report suggests that AI advancements could lead to models like GPT 5 generating over $2 billion in revenue within their first year of release.
  • 🌟 The potential economic impact of AI is vast, with the possibility of automating a small portion of the $60 trillion annual economic output.
  • 🚀 AI models are expected to develop agentic capabilities, allowing them to operate more independently and integrate seamlessly into existing workflows.
  • 📈 The report forecasts a significant scale-up in AI models, with future models by 2030 potentially being 20,000 times more capable than current ones.
  • 💹 There's a growing belief that investing trillions of dollars in AI development could be economically justified due to the enormous potential payoffs.
  • 💼 Wall Street's skepticism about AI's profitability is contrasted with the long-term vision of AI's capability to revolutionize industries and generate substantial revenue.
  • 💡 The report highlights the importance of AI's ability to automate tasks, with predictions that 100% of tasks could be automated by 2043.
  • 🌐 Companies like Meta and Amazon are investing heavily in power and data centers, indicating a commitment to the future of AI and the infrastructure needed to support it.
  • 🔋 The future of AI training involves larger models, with projections of training runs by 2030 being 5,000 times larger than those of llama 3.1, although power demand may not scale as much.
  • 📊 The report discusses the potential for synthetic data to address data scarcity issues, with reinforcement learning being a method to improve the quality of AI-generated data.

Q & A

  • What is the main focus of Epoch AI's research initiative?

    -Epoch AI's research initiative is focused on investigating trends in machine learning and forecasting the development of artificial intelligence.

  • What does the report by Epoch AI predict about the future of AI?

    -The report predicts significant advancements in AI functionality, with newer models like GPT 5 potentially generating over $2 billion in revenue within their first year of release.

  • How does the report suggest AI models will integrate into existing workflows?

    -The report suggests that AI models will seamlessly integrate into existing workflows, manipulate browser windows or virtual machines, and operate independently in the background.

  • What is the potential economic impact of AI models according to the report?

    -The report suggests that if an AI model can automate a small portion of the $60 trillion economic output, generating $20 billion in economic value is plausible.

  • What does the report say about the scale-up from GPT 4 to GPT 6 models?

    -The report indicates that GPT 4 to GPT 6 could represent a 10,000 times scale-up, and future models by 2030 could potentially be a 20,000 times scale-up.

  • How does the report address the concerns of Wall Street regarding AI investments?

    -The report justifies the massive investments in AI by highlighting the potential for unprecedented economic growth and the possibility of capturing a fraction of the $60 trillion global labor compensation.

  • What is the significance of the 10x output prediction mentioned in the report?

    -The 10x output prediction signifies that if AI automation replaces almost all human labor, economic growth could accelerate by tenfold or more over a few decades, increasing economic output significantly.

  • How does the report envision the future of training runs for AI models?

    -The report envisions training runs for AI models to be longer, potentially lasting up to six months, and to be 5,000 times larger than those of llama 3.1 by 2030.

  • What are the potential constraints on AI development according to the report?

    -The report identifies power and chip availability as the most binding constraints on AI development, with data scarcity being the most uncertain bottleneck.

  • How does the report address the issue of synthetic data and model collapse?

    -The report discusses the use of reinforcement to improve the quality of AI-generated data, which can prevent model collapse and even lead to perfect performance in some cases.

Outlines

00:00

🤖 Epoch AI's Predictions on AI's Future Economic Impact

Epoch AI, a research initiative, has released a report detailing the future of AI and its potential economic impact. The report suggests that AI advancements, particularly in models beyond GPT-4, could lead to significant economic returns. It highlights the possibility of newer models like GPT-5 generating over $2 billion in revenue within their first year of release. The report also discusses the potential for AI to automate a portion of the $60 trillion annual economic output, emphasizing the integration of AI into existing workflows and the development of agentic capabilities, allowing AI systems to operate more independently.

05:01

💹 AI's Economic Justification and Wall Street's Skepticism

The script delves into the economic justification for investing in AI, suggesting that even a fraction of the $60 trillion global labor compensation could be captured by AI, making substantial investments economically viable. It contrasts this with Wall Street's skepticism, highlighting an article questioning AI's profitability. The speaker argues that while companies are investing heavily in AI, the potential for future returns is enormous, and Wall Street may not fully appreciate the long-term value of these investments. The script also mentions predictions of AI automating all tasks by 2043, indicating a significant shift in economic dynamics.

10:03

🚀 AI's Computational Growth and Energy Demand

This section discusses the anticipated growth in AI model training, with projections that by 2030, training runs could be 5,000 times larger than those of Llama 3.1. It addresses concerns about power demand, suggesting that while training runs may become longer to spread out energy needs, advancements in algorithms and training techniques could mitigate the need for excessively long training periods. Companies like Meta and Amazon are investing in large-scale energy sources to support future AI development, indicating a commitment to overcoming potential power constraints.

15:06

🌐 The Future of AI Training and Data Constraints

The script explores the future of AI training, considering the constraints of data availability and the potential for synthetic data to extend training capabilities. It discusses the concept of a 'data wall' and how multimodal data and synthetic data generation could help overcome this limitation. The speaker references a study that suggests reinforcement methods can improve the quality of AI-generated data, preventing model collapse. The section also includes a discussion of the largest feasible training runs given current constraints, with projections for significant increases in compute power by 2030.

20:06

📈 AI Scaling and Its Impact on Future Economic Output

The final paragraph summarizes the report's findings on AI scaling, emphasizing that by 2030, we could see AI models that are 10,000 times larger in scale than current models. It suggests that this scaling, combined with increased investment and competition among tech giants, will lead to an explosion in AI capabilities. The speaker concludes by highlighting the conservative estimates of researchers and the potential for AI to revolutionize various industries, leaving viewers with a sense of the immense possibilities that AI advancements could bring to the economy and society.

Mindmap

Keywords

💡AI Hype

AI Hype refers to the widespread enthusiasm and expectations surrounding the capabilities and potential of artificial intelligence. In the video, the speaker discusses how the Epoch AI report suggests that the current AI hype is not unfounded, as AI advancements could lead to significant economic returns and transformative changes in various industries.

💡GPT Models

GPT (Generative Pre-trained Transformer) models are a series of language models developed by OpenAI. The video mentions GPT 4 to GPT 6, indicating the progression and potential capabilities of these models. The speaker highlights predictions that newer models like GPT 5 could generate substantial revenue, showcasing the economic potential of AI advancements.

💡Economic Output

Economic output refers to the total value of goods and services produced by an economy over a specific period. The script discusses how AI models could automate tasks within the $60 trillion global economic output, suggesting that even capturing a small fraction of this value through AI automation could be highly profitable.

💡Agentic Capability

Agentic capability in the context of AI refers to the ability of AI systems to operate independently, without the need for constant human intervention. The video script mentions this concept when discussing how future AI models will be able to integrate into existing workflows and operate in the background, indicating a significant shift towards more autonomous AI systems.

💡Algorithmic Improvements

Algorithmic improvements refer to enhancements made to the underlying algorithms that power AI systems. The video emphasizes the importance of such improvements in conjunction with scaling up AI models like GPT 4 to GPT 6, suggesting that these improvements will lead to more efficient and capable AI systems.

💡Economic Value

Economic value, in the context of the video, pertains to the monetary worth that AI can generate through automation and efficiency gains. The speaker cites the report's prediction that AI models could generate billions in economic value, highlighting the potential for AI to contribute significantly to economic growth.

💡Investment in AI

Investment in AI refers to the capital allocated towards the development and deployment of AI technologies. The video discusses how trillions of dollars could be invested in AI development, reflecting the belief in the transformative potential of AI and the desire to capture a portion of the global economic output.

💡Training Runs

Training runs in AI are the processes through which AI models are trained on large datasets to learn and improve their performance. The script mentions that future training runs for AI models are expected to be significantly larger and longer, indicating the increasing complexity and scale of AI model development.

💡Data Centers

Data centers are facilities used to house and manage large amounts of data and the infrastructure needed to run that data. The video mentions that companies are investing in large-scale data centers to support the power needs for training AI models, reflecting the growing infrastructure requirements for AI development.

💡Compute

In the context of AI, compute refers to the computational resources and processing power required to train and run AI models. The video discusses the projected increase in compute capabilities needed to train larger AI models by 2030, suggesting that significant advancements in hardware and processing power will be necessary to support AI growth.

Highlights

Epoch AI's report predicts significant advancements in AI, suggesting that AI hype is far from over.

The report forecasts that AI models like GPT 5 could generate over $2 billion in revenue within their first year of release.

AI models are expected to integrate seamlessly into existing workflows, manipulate browser windows, and operate independently.

The economic output potential of AI is highlighted, with the possibility of automating a small portion of the $60 trillion annual economic output.

The report discusses the potential for AI models to operate with agentic capability, reducing human reliance.

GPT 4 to GPT 6 models are predicted to represent a 10,000 times scale-up in AI capabilities by 2030.

Investment in AI is compared to the global labor compensation of $60 trillion, indicating the vast economic potential of AI automation.

The report addresses the skepticism on Wall Street about AI's profitability, suggesting a misunderstanding of AI's long-term potential.

Epoch AI predicts that AI could automate 100% of tasks by 2043, indicating a massive shift in the economic landscape.

The report suggests that AI development could lead to a tenfold increase in economic growth over a few decades.

Investors may redirect significant capital from traditional sectors into AI development due to the potential for unprecedented economic growth.

The report provides a conservative estimate that by 2030, training runs for AI models could be 5,000 times larger than those of llama 3.1.

Companies like Meta and Amazon are investing in large-scale energy sources to support future AI training needs.

The report discusses the potential for gigawatt-scale data centers to become feasible by 2030, supporting large AI training runs.

The report addresses the 'data wall' concern, suggesting that synthetic data generation could mitigate data scarcity constraints.

A method called reinforcement is presented as a way to improve the quality of AI-generated data, preventing model collapse.

The report concludes that by the end of the decade, we could train AI models that are 10,000 times larger, indicating a significant leap in AI capabilities.

Transcripts

play00:00

Epoch AI is a research initiative

play00:02

focused on investigating Trends in

play00:04

machine learning and forecasting the

play00:06

development of artificial intelligence

play00:09

now they've recently released a report

play00:12

on the future of AI and some of their

play00:14

predictions are probably the most

play00:16

accurate and it's rather surprising

play00:19

considering what most people are saying

play00:21

so essentially in this video I'll dive

play00:23

into their findings and show you why the

play00:26

AI hype is truly far from over and I'll

play00:29

show you the actual conservative

play00:31

estimates that show we in for a pretty

play00:33

wild ride over the next 6 years up until

play00:37

at least 20130 so one of the craziest

play00:40

things that I saw from the report and

play00:43

I've just you know picked up a few

play00:44

things because the entire report was I

play00:46

think around 60 or so Pages I'm not

play00:48

exactly sure how many pages but it was

play00:51

rather extensive so I decided to just

play00:53

show you guys a few Snippets from that

play00:55

report now one of the things that was

play00:57

there was that they talk about how the

play01:00

potential for sufficiently large

play01:02

economic returns that could emerge from

play01:05

scaling Beyond GPT 4 to a GPT 6

play01:08

equivalent model coupled with

play01:10

substantial algorithmic improvements and

play01:13

post-training improvements it says okay

play01:15

and this is the bit that I've

play01:17

highlighted that this evidence might

play01:19

manifest as newer models like GPT 5

play01:22

generating over $2 billion in Revenue

play01:26

within their first year of release now

play01:29

that is absolutely incredible but I

play01:32

think later on in the article they talk

play01:34

about how the entire economic output is

play01:37

around 60 trillion per year and they're

play01:40

basically stating that look if an AI

play01:42

model is able to automate a small

play01:44

portion of that it being able to get $20

play01:47

billion of economic value is not that

play01:51

hard when you actually think about the

play01:53

amount of value the economy produces so

play01:55

what you can see here is that they're

play01:56

talking about significant advancement in

play01:58

AI functionality allowing for models to

play02:01

seamlessly integrate into existing

play02:03

workflows manipulate browser windows or

play02:05

virtual machines and operate

play02:07

independently in the background so

play02:08

basically what they're talking about

play02:10

here is that you know allowing models to

play02:12

seamlessly integrate into existing

play02:14

workflows manipulate browser windows or

play02:16

virtual machines and operate

play02:18

independently in the background what

play02:20

they're referring to here is agentic

play02:22

capability so operating independently is

play02:25

where we have these systems that you

play02:27

know don't longer require humans as much

play02:30

now currently if we want AI systems to

play02:32

perform well at nearly any task what we

play02:35

have to do is we have to prompt that AI

play02:37

model so we open up a chat we say hey

play02:40

can you do this can you do that and then

play02:42

of course we have to you know refine The

play02:44

Prompt and get the AI system to do a lot

play02:47

of different things now in the future

play02:49

these things are going to be operating

play02:50

independently in the background which

play02:53

means that there's going to be quite a

play02:55

lot more scale didn't mean to draw a box

play02:56

there but this is going to be one of the

play02:58

biggest things now the thing about this

play03:00

is that if you saw another video that I

play03:02

spoke about you know the trends in

play03:04

machine learning and how we're going to

play03:05

evolve for future models a GPT 46 to GPT

play03:10

6 level equivalent model coupled with of

play03:12

course as they say substantial

play03:14

algorithmic improvements and post

play03:16

trining that is going to be absolutely

play03:18

incredible because when I looked at

play03:20

another part they basically talked about

play03:22

GPT 4 to GPT 6 could be a 10,000 times X

play03:26

scaleup or future models by 2030 could

play03:29

be entirely a 20,000 times scale up so

play03:33

it's going to be super intriguing to see

play03:36

how models scale up from gp4 to GPT 6

play03:39

because there's going to be likely two

play03:41

giant training runs there's going to be

play03:43

substantial algorithmic improvements and

play03:45

considering the fact that GPT 5 is

play03:47

likely to be released later this year or

play03:50

early next year it's going to be

play03:52

interesting to see exactly what those

play03:54

improvements are with every iterative

play03:56

cycle so this being $20 billion of

play04:00

economic revenue or economic value is

play04:02

going to be absolutely incredible but

play04:04

the point is is that it should show you

play04:06

what is going to come in the future and

play04:08

if GPT 5 could generate $20 billion in

play04:10

Revenue within its first year of release

play04:13

I'm wondering what future models are

play04:14

going to be able to do at that time now

play04:16

you can see right here like I said

play04:17

before this is where we talk about the

play04:20

$60 trillion economy and it says here

play04:23

that the potential payoff for AI that

play04:25

can automate a substantial portion of

play04:26

economic task is enormous it's plausible

play04:29

that an economy would invest trillions

play04:31

of dollars basically stating that of

play04:33

course you know it's plausible that the

play04:36

economy would invest trillions of

play04:38

dollars building up that stock of

play04:40

computer related Capital including data

play04:42

senders semiconductor fabrication plants

play04:45

and lithography machines and it says of

play04:47

course here the part I highlighted to

play04:50

understand the scale of this potential

play04:53

investment consider that Global labor

play04:56

compensation is approximately $60

play04:58

trillion per year basically stating that

play05:00

this is how much we pay people to do

play05:02

tasks that move the economy and even

play05:05

without factoring accelerated economic

play05:07

growth from AI automation if it becomes

play05:10

feasible to develop AI capable of

play05:13

effectively substituting for human labor

play05:15

investing trillions of dollars to

play05:17

capture even a fraction of the $60

play05:20

trillion flow would be economically

play05:22

Justified basically stating that look

play05:24

like I said before $60 trillion okay is

play05:27

a lot of money and if we get even a

play05:30

slice of that like even if you get $1

play05:31

trillion like think about these

play05:33

companies and what they're trying to do

play05:35

like this is why a lot of people can't

play05:37

understand why these companies are

play05:39

spending millions and millions of

play05:41

dollars on AI like there was an article

play05:44

recently where it's talking about okay

play05:46

you know AI they're spending millions

play05:48

and millions of dollars on these

play05:49

training runs on these researchers but

play05:52

Wall Street just can't understand the

play05:54

long-term picture cuz Wall Street

play05:56

they're thinking about you know cash

play05:57

flow thinking about all these metrics

play06:00

stock valuations but I'm going to show

play06:01

you guys this article right now you can

play06:03

see here it says has the AI Bubble Burst

play06:05

Wall Street wonders if artificial

play06:07

intelligence will ever make money and

play06:10

you can see that you know there has been

play06:11

one question in the minds of Wall Street

play06:13

TCH earning season when will anyone

play06:15

start actually making money from

play06:17

artificial intelligence and in the 18

play06:20

months that's kicks off the arms race

play06:21

they've promised that this is poised to

play06:23

re revolutionize every single industry

play06:25

but like I said before of course they're

play06:27

spending billions of dollars on data

play06:29

senders and you know semiconductors

play06:30

needed to run the AI models but like I

play06:32

said you know these guys on Wall Street

play06:34

are not thinking about you know

play06:35

completely 2030 when things start to get

play06:38

a little bit more crazy I like to think

play06:40

of it like this where AI right now yes

play06:42

it's having a chat gbt moment but once

play06:44

you know a lot more capabilities are on

play06:45

the line these AI companies are going to

play06:47

become so much more valuable like the

play06:49

money that they're going to make is just

play06:50

going to go up and up and up and up like

play06:52

that I think it's really going to be

play06:53

like that of course it's probably going

play06:54

to be a level level off but we're

play06:56

definitely still on that sigmoid curve

play06:58

where there's going to be huge G

play07:00

towards the end and I think that you

play07:02

know many of these um you know companies

play07:04

just can't seem to Fathom that in the

play07:06

future okay they're predicting that you

play07:08

know even this company okay this

play07:09

research organization they are

play07:11

predicting that I think 100% of tasks

play07:14

get automated by something like 2043 and

play07:16

I mean you have to think about it okay

play07:18

if the global economy is going to be

play07:20

outputting $60 trillion per year I'm not

play07:22

sure how much okay you know GL Global

play07:25

label compensation is going to grow or

play07:26

decrease by but you have to think about

play07:28

it you know these top companies they're

play07:30

going to be getting a lot of that value

play07:31

now none of these companies make

play07:33

trillions of dollars per year but you

play07:34

could argue in the future that with AI

play07:36

and automation that this is going to be

play07:38

something for for the first time it's

play07:40

probably going to happen so I do think

play07:42

that those companies their valuations

play07:44

are going to be you know astronomical in

play07:45

the future this isn't like a stock you

play07:47

know Point video but here the

play07:49

researchers are saying that look

play07:50

investing trillions of dollars to

play07:52

capture even a fraction of the flow is

play07:54

economically Justified which is what a

play07:56

lot of people can't seem to think which

play07:58

means that like when you think about you

play08:00

know the future 2030 2040 what the years

play08:04

are going to look like it truly is you

play08:05

know something that's going to blow my

play08:07

mind now so now this is where we talk

play08:09

about 10x output so it says here that

play08:11

standard economic models predict that if

play08:13

AI automation reaches a point where it

play08:16

replaces almost all human labor economic

play08:19

growth could accelerate by tenfold or

play08:22

more over just a few decades this

play08:24

accelerated growth could increase

play08:26

economic output by several orders of

play08:28

magnitude and given this potential

play08:30

achieving complete or near complete

play08:32

automation earlier could be worth a

play08:34

substantial portion of global output and

play08:36

recognizing this imense value investors

play08:38

May redirect significant portions of

play08:41

their capital from traditional sectors

play08:43

into AI development and essential

play08:45

infrastructure such as the energy

play08:47

production the distribution and the

play08:49

semiconductor fabrication plants and

play08:51

data centers and it says that this

play08:53

potential for

play08:54

unprecedented economic growth could

play08:57

drive trillions of dollars in investment

play08:59

in AI development now if you remember

play09:01

previously earlier this year where a

play09:03

certain someone a certain Sam Alman was

play09:06

talking about how much money he is going

play09:08

to be spending on AI and some of the

play09:11

future valuations that he was talking

play09:12

about you can see here it says Sam mman

play09:15

has a mindboggling price tag according

play09:17

to the Wall Street Journal somewhere

play09:19

between 5 and 7 trillion and you can see

play09:22

here that you know pretty much everyone

play09:23

is clowning him it says such numbers are

play09:26

Preposterous the fact that they're being

play09:27

talked about with anything approaching a

play09:29

straight face is indicative okay of a

play09:32

degree to which the broader AI discourse

play09:34

has become unmowed from reality however

play09:37

we're seeing that these guys that do

play09:39

research and they try to truly

play09:41

understand with conservative outputs

play09:44

okay where the AI growth is actually

play09:46

going to be and remember this isn't some

play09:47

lab that's doing like a clickbait

play09:49

article they're literally just

play09:50

publishing their research for anyone to

play09:52

view and they're just tweeting it out

play09:54

it's not like this hypey hypy thing but

play09:56

what we're seeing here is that they're

play09:58

also stating that you know trillions of

play10:00

dollars being invested in here is not

play10:02

that crazy but you can see here but you

play10:04

can see here that because Sam Alman has

play10:06

been seeking trillions of dollars to

play10:08

reshape the business and chips of AI

play10:09

many people are say think that this is

play10:11

insane this is incredible look at it

play10:13

guys look I mean look at the the

play10:14

research guys like this is something

play10:16

that they're saying that look okay when

play10:17

you start to see okay how much AI you

play10:21

know is going to be automating the

play10:22

economy and how much you know economic

play10:25

value AI is going to eat up like putting

play10:27

trillions of dollars into that doesn't

play10:29

seem that crazy when you you know

play10:31

factorize that it says you know

play10:33

recognizing this immense value investors

play10:35

May redirect significant portions of

play10:37

their Capital into traditional sectors

play10:39

of AI development so when Sam Alman was

play10:41

talking about trillions of dollars he

play10:43

wasn't just completely you know going

play10:45

off the rails in terms of AI hype this

play10:47

is something that certain research

play10:48

organizations are already starting to

play10:51

talk about now this is where we talk

play10:53

about some of the compute for larger

play10:55

models you can see here that it says

play10:57

Frontier training runs by 203 are

play11:00

projected to be 5,000 times larger than

play11:03

llama 3.1 and it says however we don't

play11:05

expect power demand to scale as much and

play11:08

this is for several different reasons

play11:09

but a 5,000 times larger training run

play11:13

than llama 3.1 in the next 6 years it

play11:16

seems crazy but I mean you can just

play11:18

imagine okay and this is actually a

play11:20

conservative estimate because they do

play11:23

have you know values that are on the

play11:24

high end but in this writing they've

play11:26

actually put the conservative estimate

play11:28

because like I said before it's not like

play11:30

this hype you know journalistic article

play11:32

it's actually just people that are doing

play11:34

research based on what they see based on

play11:36

the data that they're looking at so I

play11:38

mean when you actually take that into

play11:39

account it seems that the future is

play11:42

going to be absolutely incredible now

play11:43

you can see right here it says that they

play11:44

also expect training runs to be longer

play11:47

okay and it says since 2010 the length

play11:49

of training runs has increased by 20%

play11:52

per year among notable models since we

play11:55

expect power constraints to become more

play11:57

pressing training run durations could

play11:59

lengthen to spread out energy needs over

play12:02

time of course they're talking about

play12:03

many different things but basically

play12:04

they're stating that training runs could

play12:06

take around a year um or around you know

play12:09

a few hundred days so they do state that

play12:11

look it's going to be unlikely that

play12:13

training runs will exceed a year as Labs

play12:16

will wish to adopt better algorithms and

play12:18

training techniques on the order of time

play12:20

scale which these will provide

play12:22

substantial performance gain so

play12:23

basically saying that look no point

play12:25

training it for an entire year because

play12:27

by the time you finish training it

play12:28

there's going to be algorithmic

play12:29

improvements that you're going to need

play12:30

to go ahead and retrain the model you

play12:33

know completely once again so it's going

play12:34

to be completely intriguing to see what

play12:37

these future models are and how they're

play12:39

going to be trained but you can see

play12:40

right here that llama 3.1 was trained

play12:42

over 72 days just over 3 months but gp4

play12:45

was trained over 100 days which is

play12:47

actually 3 months no this one is 2

play12:49

months and this one is actually 3 months

play12:51

the point is is that it's going to be

play12:52

interesting to see how these training

play12:54

techniques differ now one thing that we

play12:56

are seeing is that companies are

play12:58

starting to absolutely buy into this we

play13:01

can see that meta bought the rights to a

play13:04

power output of 350 megawatt solar farm

play13:07

in Missouri and a 300 megawatt solar

play13:10

farm in Arizona and Amazon owns a data

play13:14

center campus in Pennsylvania with a

play13:17

contract for 960 megawatt for the

play13:19

adjoining 2.5 GW nuclear plant so you

play13:23

can see that Amazon is really really

play13:25

pushing the envelope when it comes to

play13:27

the amount of power that they're going

play13:29

to need because they are really going

play13:31

all in on this stuff and you can see

play13:34

here that it says that the primary

play13:35

motivation behind these deals is to save

play13:38

on grid connection costs and guarantee a

play13:41

reliable energy Supply in the coming

play13:43

years data centers might allow for

play13:45

unprecedentedly large training runs to

play13:47

take place and a 960 megawatt data

play13:50

center would be over 35 times more power

play13:54

than the 27 megaw required for today's

play13:56

training runs we can see that this start

play13:59

is already happening behind the scenes

play14:01

these companies are ramping up for you

play14:03

know 35 times more power needed than

play14:06

current AI models and you can see here

play14:08

that it says that you know some

play14:09

companies are investigating options for

play14:12

gigawatt scale data centers as you know

play14:15

and and basically they're stating that

play14:16

we're going to have gigawatt scale data

play14:18

centers that actually seem feasible by

play14:20

2030 and it says that this assessment is

play14:23

supported by industry leaders and

play14:25

corroborated by recent media reports

play14:27

this is the CEO of next year the largest

play14:30

utility company in the United States

play14:32

recently stated that while finding a

play14:35

site for a 5 GW AI data center would be

play14:38

challenging locations capable of

play14:40

supporting a 1 gaw facilities do exist

play14:43

within the country so they're basically

play14:45

stating that look whilst 5 gwatt AI data

play14:47

centers are pretty insane a 1 gwatt data

play14:50

center the facilities currently do exist

play14:52

within the country and of course if you

play14:54

do remember that you know openai and

play14:56

Microsoft have the 2028 St star game

play14:59

that will require several gaw of Power

play15:02

with an expansion up to 5 gaw by 2030

play15:05

now that's a huge feat and that's going

play15:07

to be really difficult to accomplish but

play15:09

I mean this is you know a race there's

play15:11

going to be lots and lots of money

play15:13

invested in this and you have to

play15:15

understand that they're talking about

play15:17

capturing $60 trillion of economic value

play15:20

so I think a few billion dollars into

play15:22

some data centers is something that

play15:24

they're not going to scoff at so now you

play15:25

can see here that this is where they

play15:27

talk about the future training runs they

play15:29

say that training runs we will presume

play15:31

that they will not likely exceed six

play15:33

months and we will assume that training

play15:35

runs will last around 2 to nine months

play15:37

on the higher end if progress in

play15:39

hardware and software stalls and on the

play15:40

lower end if progress accelerates

play15:42

relative to day so it could be two

play15:44

months or it could be 9 months so this

play15:46

is pretty crazy cuz it seems that you

play15:48

know it's still going to get pretty

play15:49

longer and then of course this is where

play15:51

we get into some incredible statistics

play15:54

it says since the chinchilla scaling

play15:56

laws suggest that one ought to SC scale

play15:59

up data set size and model size

play16:01

proportionately scaling up training data

play16:03

by a factor of 30 times by using the

play16:06

entirety of the index web would enable

play16:08

labs to train models with 30 times more

play16:11

data and 30 times more parameters

play16:15

resulting in 900 times as much comput

play16:19

okay if models are trained to be

play16:21

chinella optimal which is absolutely

play16:23

insane okay and you know people have

play16:25

been saying that we've you know

play16:26

exhausted all our data but we haven't

play16:29

actually done that yet so can you

play16:30

imagine a model being trained with 30

play16:32

times more data 30 times more parameters

play16:34

and 900 times more compute I mean it's

play16:37

going to be truly incredible with as to

play16:39

how these systems are going to be

play16:40

working now like I said before many

play16:43

people have spoken about this data wall

play16:45

which is a thing where you know people

play16:47

are thinking that okay we're going to

play16:50

run out of data but you can see right

play16:52

here that they say that if the recent

play16:54

trend of four times a year scaling you

play16:57

know continues we would run into this

play16:59

data War for TCH data in about 5 years

play17:02

so basically where we completely run out

play17:04

of data but it also does State here that

play17:07

however data from other modalities and

play17:09

synthetic data generation might help

play17:11

mitigate this constraint we will argue

play17:13

that the multimodal data will result in

play17:15

effective data stocks of about 450

play17:18

trillion to 23 quadrillion tokens

play17:21

allowing for impressive training runs

play17:24

and of course synthetic data might

play17:26

enable scaling much Beyond this if AI

play17:28

labs spend a significant fraction of

play17:30

their compute budgets on data generation

play17:33

now the synthetic data conversation is

play17:35

one that's rather interesting because

play17:37

there was this recent report and

play17:39

basically there was this paper that you

play17:42

know actually addresses an issue with

play17:44

synthetic data now basically with

play17:46

synthetic data um there was this issue

play17:48

called Model collapse and I need to show

play17:50

you guys what this is it's not really a

play17:51

real issue but this is something that

play17:53

people always bring up and I'm going to

play17:54

show you guys I know this isn't the best

play17:56

image that you're going to see not from

play17:57

the best article either but essentially

play17:59

what they're stating is that you know

play18:00

you have real images then those real

play18:02

images produce fake images those fake

play18:04

images are used to train another model

play18:05

that produces even more fake images and

play18:07

by the fourth iteration you have a

play18:09

system that collapses essentially um and

play18:12

basically they're saying that you know

play18:13

this lack of human data is going to

play18:15

limit AI progress however um what these

play18:17

studies uh show is that models that are

play18:19

just you know completely just trained on

play18:21

their own data again and again and again

play18:23

they weren't really you know filtering

play18:25

like with humans and stuff like that um

play18:26

this is why I'm talking about this paper

play18:28

cuz this paper came out recently um and

play18:30

this B basically you know um they've had

play18:32

a new method and this method is called

play18:34

reinforcement to improve the quality of

play18:36

AI generated data and this involves

play18:39

having a system which could be a human

play18:41

or an AI which checks the generated data

play18:44

and then only selects the best examples

play18:46

for training future models and basically

play18:48

they provide mathematical proof that

play18:50

under certain conditions using

play18:52

reinforced data can prevent model

play18:54

collapse and even lead to perfect

play18:56

performance in some cases so without

play18:58

reinforcement training on AI generated

play19:00

data would indeed lead to worse

play19:02

performance which is model collapse but

play19:04

with reinforcement and selecting the

play19:06

best AI generated data they could

play19:08

prevent model collapse and sometimes

play19:11

even improve model performance beyond

play19:14

the original model and the quality of

play19:16

both the data and the generator and the

play19:18

reinforcement system are important for

play19:20

good results so whilst many people are

play19:22

thinking that synthetic data is simply

play19:24

this hole that is never going to be

play19:26

filled there is a lot of research that

play19:28

is is out there that suggests that this

play19:30

isn't the truth now what we also do have

play19:33

is this graph that shows us the largest

play19:36

feasible training runs given the

play19:38

different constraints many people you

play19:40

know talk about AI hype and they talk

play19:41

about how AI is just complete overly

play19:43

hyped in terms of the future progress

play19:45

but like I said before these are people

play19:47

who've researched the stuff and they

play19:49

said that this is what the largest

play19:51

feasible training runs are given the

play19:53

actual different constraints so we have

play19:55

different constraints here we've got the

play19:57

power constraints which are you know the

play19:59

energy supplies of course we've got the

play20:01

chip production capacity which is NVIDIA

play20:03

being able to even produce enough chips

play20:05

recently we had news that there were

play20:07

delays on I think the b200s and of

play20:10

course we've got the data scarcity and

play20:12

of course the latency wall now you can

play20:14

see here that they state that the most

play20:15

binding constraints are power and Chip

play20:17

availability and you can see that

play20:19

essentially these are the ones here that

play20:21

are pretty crazy but you can see that it

play20:24

says that data stands out as the most

play20:26

uncertain bottleneck with its

play20:27

uncertainty SP a range of four orders of

play20:30

magnitude you can see on the graph here

play20:31

that data is all the way down here and

play20:33

it's all the way up here so they're not

play20:35

sure but you can see that by 2030 this

play20:38

is where they expect things to be and

play20:39

I'm going to show you guys another image

play20:41

that basically explains everything but

play20:42

basically the worst case scenario okay

play20:45

like the literal worst case scenario is

play20:47

that we have systems that are you know

play20:49

10,000 times greater in terms of the

play20:52

scale so this is you know pretty insane

play20:55

when you actually think about it you can

play20:56

see that there are other areas where we

play20:57

could get to 50 ,000 times greater you

play20:59

know chip capacity 880,000 times greater

play21:02

a million times greater in terms of the

play21:03

latency but um yeah it just shows us

play21:06

that you know by 2030 things are going

play21:08

to get rather incredible and I mean this

play21:10

is taking you know the average you know

play21:12

of all of these and then of course you

play21:14

can see it's brought it down here so

play21:15

it's not like the highest the complete

play21:17

highest but we can see that the 2030

play21:19

compute projection shows we're going to

play21:20

have 10,000 times more compute to train

play21:22

these models by 2030 which means I'm not

play21:25

like that that there's going to be just

play21:27

like an explosion in terms of these

play21:28

models are going to be in terms of their

play21:30

effect now the takeaway from this that

play21:32

you should think about is basically

play21:33

they're stating that by the end of the

play21:35

decade so by 20130 we're going to be

play21:37

able to train a model that is 10,000

play21:39

times larger because gpt2 to GPT 4 if

play21:42

you remember gpt2 to GPT 4 that scale

play21:44

was 10,000 times larger and they're

play21:46

basically saying that we're going to be

play21:47

able to do that by the end of the decade

play21:49

so if you can imagine that with all the

play21:51

progress that we've had just in the past

play21:52

three years which has been quite a lot

play21:54

but now with all the investment now with

play21:56

all the money now with all the eyes on

play21:58

AI with all the major players in

play21:59

robotics with all of the companies en

play22:01

thropic Google Amazon with all of those

play22:03

companies competing the fact that we're

play22:05

also going to have 10,000 times more

play22:07

computer available and the fact that by

play22:09

that time in 2030 we're going to be able

play22:11

to train a model that is going to be

play22:13

10,000 times larger in scale what kinds

play22:16

of systems are we going to have in place

play22:17

I mean it's going to be pretty crazy but

play22:20

I think this you know should let you

play22:22

understand that like even even in the

play22:24

conservative estimates of these people

play22:26

that have done the research it shows us

play22:28

that we're going to have a incredible

play22:30

time in terms of AI so hopefully this

play22:33

video educated you guys a little bit in

play22:35

terms of you know how the future is

play22:36

going to be in terms of compute the full

play22:38

thing is actually really long you can

play22:40

see here that I'm scrolling down for

play22:42

quite some time it's called can AI

play22:43

scaling continue through 2030 and it

play22:46

says we investigate the scalability of

play22:48

AI training runs we identify all of the

play22:50

stuff but you can see right here that

play22:51

guys this is something that is really

play22:54

really long I read through this entire

play22:55

thing it's super super detailed super

play22:58

super they've got so many different re

play23:00

um you know people that have done

play23:01

research on this and you can see that

play23:02

all of the sources are cited here you

play23:04

can you know walk through on the right

play23:06

hand side click through different things

play23:07

sometimes are images but if you do want

play23:09

to do this link will be in the

play23:10

description if you guys have any

play23:11

comments down below let me know what you

play23:13

think about this and if the future is

play23:14

going to be crazy and I'll see you guys

play23:15

in the next one

Rate This

5.0 / 5 (0 votes)

Ähnliche Tags
AI TrendsFuture PredictionsEconomic ImpactMachine LearningAI DevelopmentEpoch AITech InnovationGPT ModelsAI AutomationInvestment Trends
Benötigen Sie eine Zusammenfassung auf Englisch?