The AI Hype is OVER! Have LLMs Peaked?

TheAIGRID
9 May 202428:32

Summary

TLDRThe video script discusses the current perception of AI, particularly generative AI and large language models (LLMs), and counters the notion that AI hype is fading. It references the Gartner Hype Cycle to illustrate the stages of technological maturity and adoption, arguing that while some believe we are at a plateau, significant advancements are still to come. The speaker highlights key developments such as OpenAI's Sora, advancements in voice recognition, and the potential for AI to exceed human performance in complex tasks. They also emphasize the importance of energy and compute capacity as potential bottlenecks for future AI growth. The script suggests that internal developments at companies like OpenAI are far from slowing down, with暗示 (hints) at groundbreaking models like GPT 5 that could revolutionize the field. The summary concludes by emphasizing that despite external appearances, the AI industry is on the cusp of significant breakthroughs that will propel it into a new era of innovation and capability.

Takeaways

  • 🚀 **AI Hype and Progression**: Despite perceptions that AI hype is waning, significant advancements are still being made in the field of generative AI, with large language models (LLMs) continuing to evolve.
  • 📈 **Gartner Hype Cycle**: The Gartner Hype Cycle is a useful tool for understanding the trajectory of technological development, but some argue that AI, particularly LLMs, is not following the traditional pattern of the cycle.
  • 💡 **Innovation in AI**: Companies like OpenAI are making consistent breakthroughs in AI, some of which are not publicly disclosed, indicating that internal progress is much faster than what is visible externally.
  • 🔥 **Energy and Compute Bottlenecks**: Energy consumption and compute capacity are significant bottlenecks for the future of AI, with the need for vast amounts of power and advanced infrastructure to support increasingly complex models.
  • 💼 **Business Dynamics**: The shift from open research to closed, proprietary development means that companies are less likely to share their innovations, leading to a perception that progress is slower than it actually is.
  • 🌐 **Global AI Race**: There is a global race in AI development, with companies investing heavily to stay competitive, which is driving rapid advancements and the potential for new breakthroughs.
  • 🔬 **Advanced Reasoning Engines**: The incorporation of advanced reasoning engines on top of existing models like GPT-4 is pushing the boundaries of what AI can achieve, particularly in complex problem-solving tasks.
  • 📊 **Benchmarking and Competition**: Companies are incentivized to surpass the current benchmarks like GPT-4, leading to a focus on marginal improvements rather than acknowledging the potential for more significant leaps in capability.
  • 🔍 **Internal Developments at OpenAI**: There is an expectation that OpenAI is significantly ahead of its current public releases, suggesting that future releases could bring substantial advancements in AI capabilities.
  • 🔥 **Sora and Other Innovations**: OpenAI's release of Sora, a video generation model, and other undisclosed projects indicate that the company is exploring various AI applications beyond just text-based models.
  • ⏳ **Future Predictions**: Industry leaders like Sam Altman suggest that the current state-of-the-art models will be considered primitive in the near future, hinting at the potential for massive jumps in AI performance.

Q & A

  • What is the current debate about AI on social media platforms like Twitter?

    -The current debate is centered around whether the hype surrounding AI, particularly generative AI, is wearing off and if we have reached an exhaustion point in terms of its capabilities and future developments.

  • What is Sam Altman's stance on the investment in building AGI, despite the high costs?

    -Sam Altman is quoted as saying that he doesn't care if they burn $50 billion a year because they are building AGI, and he believes it will be worth it.

  • What is the Gartner Hype Cycle and how is it relevant to AI technologies?

    -The Gartner Hype Cycle is a graphical representation used to illustrate the maturity, adoption, and social application of specific technologies. It provides a conceptual framework to understand how technologies evolve from introduction to mainstream adoption, which is particularly useful for evaluating emerging technologies like generative AI and large language models.

  • What are some of the current criticisms of large language models (LLMs)?

    -Criticisms include biases in the models, the need for vast data training, high operational and inference costs, environmental impacts, and issues like hallucination where the model generates incorrect or misleading information.

  • What is the 'slope of Enlightenment' in the Gartner Hype Cycle?

    -The 'slope of Enlightenment' is a phase where previous issues with a technology are ironed out. For LLMs, this would involve improvements in model efficiency, reduced biases, and increased reliability.

  • What are some of the advancements in AI that suggest it is not stagnating?

    -null

  • What does Sam Altman imply about the future of AI models beyond GPT-4?

    -Sam Altman implies that future models, such as GPT-5, will represent a significant leap from the current state-of-the-art, suggesting that GPT-4 will be considered 'dumb' in comparison to what is to come.

  • What are the potential bottlenecks that could slow down the progress of generative AI?

    -The potential bottlenecks include energy constraints, due to the high power requirements for training and running large models, and compute capacity, as the demand for processing power may exceed current supply.

  • null

    -null

  • Why might companies be hesitant to publicly release all their AI research and advancements?

    -Companies like OpenAI may keep their research and advancements secret to maintain a competitive edge. Revealing too much could allow competitors to replicate their successes and potentially surpass them.

  • How does the release of GPT-4 as a benchmark affect the AI industry?

    -The release of GPT-4 as a benchmark creates an incentive for other companies to train models that just surpass GPT-4, rather than aiming for a significant leap in capabilities. This can create an illusion of a plateau in AI progress.

  • What does the future hold for AI capabilities according to the advancements and statements made by industry leaders?

    -The future of AI is expected to be transformative, with significant advancements in reasoning capabilities, efficiency, and the potential for AI agents to perform complex tasks at a level surpassing current human capabilities.

Outlines

00:00

🤖 AI Hype and the Gartner Hype Cycle

The first paragraph discusses the perception that the hype around AI might be fading, particularly in the context of generative AI. It references a clip from Sam Altman, emphasizing the commitment to developing AGI despite the costs. The paragraph also introduces the Gartner Hype Cycle as a tool for understanding the trajectory of technological development, from introduction to mainstream adoption. It outlines the cycle's stages, including the technology trigger, peak of inflated expectations, and the trough of disillusionment, which is where many believe generative AI currently stands due to issues like biases, high costs, and environmental impact.

05:02

🚀 The Upcoming Surge in AI Advancements

This paragraph challenges the notion that AI is stagnating and argues that we are on the cusp of significant breakthroughs. It points out that while some benchmarks show AI nearing human performance, there are still complex tasks where AI lags, suggesting room for growth. The speaker disagrees with the idea that we have reached the limits of generative AI and alludes to ongoing advancements in various AI domains, such as voice recognition and video generation, indicating that the field is far from plateauing.

10:03

📉 The Downturn in AI Hype and Future Predictions

The third paragraph focuses on the idea that despite the perceived downturn in AI hype, industry leaders hint at upcoming revolutionary AI models. Sam Altman's statements suggest that current state-of-the-art models like GPT-4 are considered 'dumb' compared to what's coming next, indicating a significant leap in AI capabilities. The paragraph also discusses the potential for energy consumption and GPU production to become bottlenecks in AI development, highlighting the infrastructure challenges that could slow progress.

15:06

💡 The Impact of Energy and Compute Constraints on AI

This paragraph delves into the challenges of energy consumption and the supply constraints of GPUs, which are critical for training large language models. It discusses the potential for energy to be a limiting factor in AI's growth, given the extensive power requirements for training clusters. It also touches on the regulatory hurdles and long lead times associated with building new power plants and infrastructure, suggesting that these factors could slow the pace of AI advancement.

20:07

🌐 The Unseen Progress in AI and the Shift to Closed Research

The fifth paragraph emphasizes that while external appearances may suggest a slowdown in AI progress, internal developments at companies like OpenAI are likely far more advanced. It highlights the shift from open research to closed, proprietary research as companies protect their innovations. The paragraph also mentions the competitive drive among companies to surpass GPT-4, suggesting that the current benchmark does not indicate a plateau but rather a strategic pause before the next significant release.

25:07

🔮 Future AI Capabilities and the Race for Advancements

The final paragraph speculates on the potential for AI to surpass current benchmarks significantly, leading to transformative changes across industries. It discusses the potential integration of advanced reasoning engines with current models like GPT-4 and the impact of iterative agent workflows. The speaker expresses optimism about the unseen advancements within companies like OpenAI and predicts that the next year will bring groundbreaking developments in AI.

Mindmap

Keywords

💡AI Hype

AI Hype refers to the heightened public interest and media coverage surrounding artificial intelligence technologies, particularly generative AI and large language models. In the video, the speaker discusses the perception that the excitement around AI might be fading, but argues against this notion, suggesting that significant advancements are still to come.

💡Generative AI

Generative AI is a branch of artificial intelligence that involves creating new content, such as text, images, or videos, that did not exist before. It is a key focus of the video, with the speaker highlighting how this technology is evolving and disputing claims that its development is slowing down.

💡Gartner Hype Cycle

The Gartner Hype Cycle is a graphical representation that illustrates the life cycle of a technology, from its initial emergence to eventual mainstream adoption. The video uses this concept to discuss the stages of technology adoption, particularly in relation to AI, and to argue that AI is not yet at its peak in terms of public interest and capability.

💡Large Language Models (LLMs)

Large Language Models (LLMs) are AI systems that process and generate human-like language at scale. They are a central topic in the video, where the speaker discusses their current capabilities, the challenges they face, and the potential for future advancements that will significantly surpass current models.

💡Sam Altman

Sam Altman is the CEO of OpenAI and a key figure in the development of AI technologies. In the video, his statements about the future of AI and the willingness to invest heavily in AGI (Artificial General Intelligence) are used to illustrate the commitment and vision of industry leaders in pushing the boundaries of AI.

💡AI Efficiency and Inference Costs

AI Efficiency and Inference Costs refer to how well AI models perform with the resources they use and the expenses associated with running them. The video discusses these costs as a challenge for AI development, highlighting the need for more efficient models that reduce operational and environmental impacts.

💡Bias in AI

Bias in AI refers to the systemic errors or unfairness that can occur in AI systems due to the data they are trained on or the algorithms they use. The video touches on this issue as one of the problems that need to be addressed as AI technology matures, emphasizing the importance of reducing biases for more reliable and fair AI systems.

💡Energy Constraints

Energy Constraints highlight the limitations and challenges posed by the high energy consumption of AI systems, particularly when it comes to training large models. The video suggests that energy costs and infrastructure could become a bottleneck in the rapid advancement of AI, affecting the pace at which new models are developed and deployed.

💡Compute Capacity

Compute Capacity refers to the ability of hardware to perform computational tasks, which is crucial for training and running AI models. The video discusses the need for increased compute capacity to meet the demands of future AI systems, with references to the development of supercomputers to handle the complex tasks of advanced AI.

💡Closed Research Environment

A Closed Research Environment is a situation where research and development are conducted privately within a company, with findings not shared publicly. The video suggests that companies like OpenAI have moved from an open research model to a closed one, keeping their innovations secret to maintain a competitive edge and not give away their advancements to competitors.

💡Multimodal AI Agents

Multimodal AI Agents are systems capable of processing and understanding multiple types of data inputs, such as text, images, and audio. The video speaks to the potential of these agents in performing complex tasks and the significant leap forward this represents for AI capabilities, suggesting that when these agents reach a certain level of performance, it will mark a transformative moment for AI.

Highlights

AI hype is not wearing off; generative AI's capabilities are still growing.

Sam Altman's statement about investing heavily in AGI development signifies ongoing commitment to AI advancement.

Generative AI is not slowing down; it's evolving with new developments in voice recognition and video generation.

Gartner hype cycle is used to illustrate the maturity, adoption, and application of technologies like generative AI.

The peak of inflated expectations for LLMs was reached with GPT 3.5 and GPT 4, indicating significant media attention and public interest.

The trial of disillusionment phase is characterized by the recognition of issues like biases, high operational costs, and environmental impacts of large language models.

The slope of Enlightenment suggests that LLMs are improving in efficiency, reducing biases, and increasing reliability.

AI performance on benchmarks is approaching human levels in areas like image classification and natural language inference, but complex tasks remain a challenge.

Sam Altman hints at significant advancements beyond GPT 4, suggesting that future models will be dramatically more capable.

Energy consumption and GPU supply constraints are potential bottlenecks for the growth of AI systems.

NVIDIA's Blackwell GPU architecture promises significant performance improvements for training large language models.

OpenAI's shift from open research to closed research environment means they are making consistent breakthroughs without immediate public disclosure.

GPT 4 as a benchmark creates an illusion of a plateau, as companies aim to just surpass it before releasing their models.

The incorporation of advanced reasoning engines on top of existing models like GPT 4 is pushing the boundaries of AI capabilities.

Andrew NG's research on iterative agent workflows shows significant improvements in AI performance, indicating ongoing progress in the field.

Internal advancements at companies like OpenAI suggest that there will be shocking releases in the future as they advance beyond current benchmarks.

The next 365 days are predicted to bring about significant changes and advancements in AI, particularly in multimodal agents and open-ended tasks.

Transcripts

play00:00

so one of the current questions that

play00:01

I've been seeing floating around on

play00:03

Twitter and on social media is that AI

play00:07

hype might be wearing off and when I'm

play00:09

talking about AI hype this is referring

play00:11

to the trend that generative AI has I

play00:14

guess you could say reached an

play00:16

exhaustion point now I disagree and I'm

play00:17

going to tell you guys why but I'm going

play00:19

to go over some of the key examples and

play00:21

what people are stating so essentially

play00:23

there was a clip here and a lot of

play00:25

people were stating that this clip from

play00:27

Sam Altman people were stating that the

play00:29

hype is wearing off The Vibes are

play00:31

shifting you can feel it and basically

play00:33

in this clip clip here Sam Alman

play00:36

literally is he's only really stating

play00:38

that you know I don't care if we burn

play00:39

$50 billion a year we're building AGI

play00:42

and it's going to be worth it so it's

play00:44

not a crazy crazy statement I think why

play00:46

people are stating that the hype is

play00:48

wearing off and that Vibes are wearing

play00:50

off and this video is going to be and

play00:52

essentially what this is referring to

play00:53

here is the fact that you know I guess a

play00:56

lot of people are thinking that you know

play00:58

I guess you could say the generative AI

play01:01

is currently slowing down in terms of

play01:03

the capabilities and what we're likely

play01:05

to see in the future now like I said

play01:07

before it's just mainly due to this clip

play01:09

and a bunch of other different factors

play01:11

but let's get into how Cycles work in

play01:13

terms of technology and a graph that a

play01:16

lot of people have been referencing when

play01:17

they talk about llms plateauing so

play01:19

something that I also saw quite a bit

play01:21

being you know passed around as like

play01:23

this infographic is of course the

play01:25

Gartner hype cycle so essentially it's

play01:27

just a graphical rep representation used

play01:30

to illustrate maturity adoption and

play01:32

social application of specific

play01:34

Technologies it basically just provides

play01:36

a conceptual framework that helps

play01:38

stakeholders and individuals understand

play01:40

how Technologies evolve from the time

play01:42

they introduced until they reach

play01:45

mainstream adoption and the hype cycle

play01:47

can be particularly useful for

play01:49

evaluating emerging Technologies like

play01:51

generative Ai and of course large

play01:54

language models so of course we

play01:56

essentially have the technology trigger

play01:58

this phase initially occurs when a new

play02:01

technology is conceptualized or when a

play02:03

significant development makes it

play02:05

publicly known creating interest and

play02:07

significant media attention for example

play02:09

this is when GPT 3.5 chat GPT was

play02:13

released and it demonstrated the ability

play02:15

to create text in the long

play02:19

format and then of course that's when we

play02:22

got to GPT 4 which is stage two so this

play02:25

is where there is the peak of inflated

play02:27

expectations now I've got to be honest

play02:30

it's not just llms that were going crazy

play02:32

at this point there was also 11 labs and

play02:35

other image generation services like mid

play02:37

journey and of course other services

play02:40

like stable diffusion so I would argue

play02:42

that the problem with this is that with

play02:44

generative AI experiencing some kind of

play02:47

increase in the actual media coverage I

play02:50

would say that this is something that is

play02:53

cumulatively increasing in terms of the

play02:55

expectations and that's because like I

play02:57

said already there were many several

play02:59

different categories that came together

play03:01

at the same time and of course this

play03:03

could include llms being at the peak of

play03:06

inflated expectations now I want to say

play03:08

that I do disagree with this hype cycle

play03:11

for AI I do think that this is nowhere

play03:13

near it its peak where it should be but

play03:16

there of course is inflated expectations

play03:18

when a new technology comes to fruition

play03:21

a lot of people may exaggerate what the

play03:24

technology can really do for example

play03:26

some people say it's going to replate

play03:28

replace entire careers or entire tasks

play03:30

and revolutionize entire Industries now

play03:33

whilst yes that might happen in the

play03:34

future I don't think that that is

play03:36

completely happening with GPT 4 and gbt

play03:38

3.5 so some are arguing that this is

play03:41

where we are at currently okay and then

play03:43

of course the trial of disillusionment

play03:45

and this is where Technologies enter

play03:47

this phase when they fail to meet the

play03:49

inflated expectations and stakeholders

play03:51

become increasingly delusion so issues

play03:54

related to the technology start to arise

play03:57

such as the biases and large language

play03:59

models need for vast data training the

play04:01

high operational costs the high

play04:03

inference costs the environmental

play04:06

impacts they become more apparent and

play04:08

then they become criticized and one of

play04:09

the main things that many people are

play04:11

talking about with llms is of course

play04:13

things like the hallucination and the

play04:15

high inference costs and of course the

play04:17

training cost because these models are

play04:19

certainly not cheap and it seems like

play04:21

there's only a few companies that can

play04:23

run and train these large models now of

play04:25

course there we have the slope of

play04:27

Enlightenment and this is essentially

play04:29

where with l M all of these previous

play04:31

issues get ironed out so things like the

play04:33

issues of hallucination and the biases

play04:35

they get ironed out right here and this

play04:37

is where as more experimentations as

play04:39

more implementations occur the market

play04:41

matures and second third generation

play04:43

products appear and with llms this would

play04:46

involve developments that address the

play04:48

earlier criticisms such as improving the

play04:50

model efficiency the inference costs

play04:53

reducing the biases and of course the

play04:55

reliability of the model and this is

play04:57

arguably where people think AI is going

play04:59

to go technology becomes stable and

play05:01

accepted and this means widespread

play05:04

adoption however I think the graph isn't

play05:06

going to look anything like that I think

play05:08

it's probably going to look something

play05:09

like this where we go up we dip a little

play05:11

bit and then we continue a trajectory

play05:13

upwards because like I said before

play05:15

whilst yes that it does seem to many

play05:18

different pieces of Statistics that it

play05:21

looks like we are slowing down in terms

play05:23

of AI versus human performance and even

play05:26

on this Stanford AI index we can see

play05:28

that AI has surpassed human performance

play05:30

on several benchmarks including someon

play05:32

image classification visual reasoning

play05:34

and English understanding yet it Trails

play05:36

behind on more complex tasks like

play05:38

competition level mathematics visual

play05:40

Common Sense reasoning and planning and

play05:43

you can see here that if we take a look

play05:45

at actually how the AI is moving we can

play05:48

see that in image classification visual

play05:50

Common Sense reasoning natural language

play05:52

inference all of these seem to be coming

play05:53

towards the human level Baseline but

play05:56

they don't seem to be going upwards you

play05:58

know like in a crazy level on on the

play05:59

graph and some people would argue that

play06:02

this is because AI generative AI large

play06:04

language models whatever you want to

play06:06

call it if you just want to group

play06:07

everything together that we have reached

play06:09

our limit in terms of where we are and

play06:12

new architectures are going to be needed

play06:14

now I would firmly disagree with this

play06:16

for several reasons that I'm about to

play06:18

explain and I'm going to show you guys

play06:20

some really key evidence on why things

play06:22

are about to actually get very very

play06:24

crazy in the world of AI and why we're

play06:27

about to go into a very very abundant

play06:30

era due to AI so one of the things that

play06:34

many people are not actually taking into

play06:36

account is the fact that currently there

play06:38

are many different things going on in

play06:40

the world of AI that people aren't

play06:41

paying attention to some people are just

play06:44

paying attention to large language

play06:45

models and this doesn't make sense

play06:47

because there are vast and many

play06:49

different categories in which AI is

play06:51

currently exceeding for example in voice

play06:53

recognition and in voice generation

play06:56

opening ey has recently developed their

play06:57

voice engine which was actually

play06:59

something that they developed in around

play07:01

2022 and it was basically a

play07:03

state-of-the-art system that could

play07:04

recreate anyone's voice in addition to

play07:06

that if you do remember open ey also did

play07:10

talk about and quote unquote release

play07:12

their sore up model to Showcase us how

play07:16

far they've come in creating a text

play07:18

video model now I think you have to

play07:20

understand how crazy this is because for

play07:22

people to say that generative AI is

play07:24

stagnating is a crazy statement when

play07:27

literally this year we literally got

play07:30

Sora which blew everyone's Minds in

play07:32

terms of the capabilities so with Sora I

play07:35

still find it absolutely incredible that

play07:37

people can say AI is stagnating because

play07:40

with Sora as you've all seen this was

play07:43

something that was truly just

play07:44

mind-blowing this piece of technology

play07:47

showed us how crazy it is when you get a

play07:50

group of AI researchers dedicated to

play07:52

doing something and I'm going to give

play07:55

you guys a quick memory joke remember

play07:57

that opening ey isn't a company that's

play07:59

Focus fed on video generation this is

play08:01

just a subsection of their company so

play08:04

the video generation aspect was

play08:05

something that I guess they just wanted

play08:07

to see if they could do well at and they

play08:09

literally surpassed state-of-the-art

play08:11

models Google Runway pabs they

play08:14

completely surpassed them and this is

play08:16

absolutely crazy I mean the demo

play08:18

literally absolutely shocked everyone I

play08:21

was truly truly speechless when this

play08:23

technology was here and I'm someone that

play08:25

pays attention to all of the AI news and

play08:27

of course we do have Devon cognition

play08:30

Labs released Devon their first AI

play08:31

software engineer and this was a AI

play08:35

rapper around GPT 4 but it did a few

play08:38

things in a unique way where it was able

play08:40

to surpass GPT 4 on certain software

play08:43

engineering benchmarks so one of the

play08:45

first things the point here I'm making

play08:46

is that you might think that llms are

play08:49

stagnating but that is not the truth at

play08:51

all we're going to get into llms later

play08:53

but overall voice engine Sora and Devon

play08:56

show us that generative AI is really

play08:59

really not going to be stagnating

play09:01

anytime soon but if that didn't convince

play09:03

you let me show you guys some of the

play09:05

recent statements that show you that

play09:06

we're absolutely in for a crazy ride so

play09:09

Sam Alman recently said in an interview

play09:11

at Stanford's entrepreneurship talk he

play09:14

spoke about gp4 now remember GPT 4

play09:17

currently is a state-of-the-art system

play09:20

meaning that it is the best of the best

play09:22

that we can currently get our hands on

play09:23

for public use which means that Sam

play09:26

Alman currently probably has access to

play09:28

Frontier that are being developed by

play09:30

open Ai and remember the big Labs like

play09:33

anthropic and Google are currently

play09:35

behind them in terms of what they're

play09:36

creating so take a listen to this

play09:38

statement open ey is phenomenal chat gbt

play09:41

is phenomenal um everything else all the

play09:42

other models are phenomenal it burned

play09:44

you've burned $520 million of cash last

play09:46

year that doesn't concern you in terms

play09:49

of thinking about the economic model of

play09:51

how do you actually where's going to be

play09:52

the monetization source well first of

play09:54

all that's nice of you to say but Chachi

play09:56

BT is not phenomenal like Chachi BT is

play09:58

mildly embarrassing in a best um gp4 is

play10:01

the dumbest model any of you will ever

play10:03

ever have to use again by a lot um but

play10:06

you know it's like important to ship

play10:07

early and often so if you weren't paying

play10:09

attention there Sam mman literally just

play10:11

said that this is going to be the

play10:13

dumbest model that we will have to use

play10:15

or that we will have had to use by far

play10:18

so he didn't just state that this was

play10:19

going to be an incremental increase he

play10:21

clearly stated that gp4 was dumb he

play10:24

stated this model was not you know great

play10:28

he stated that this model was not that

play10:30

good he clearly stated that GPT 4 a

play10:32

current state-ofthe-art system that

play10:34

people were literally able to get you

play10:36

know increasingly capabilities just by

play10:38

wrapping the system and being able to do

play10:40

software engineering tasks and a lot of

play10:42

people even using GPT 4 to be able to

play10:45

train robotic systems like recently we

play10:47

saw in a research paper and this was

play10:49

literally where we had language model

play10:51

guided Sim to real transfer so we

play10:53

basically had large language models were

play10:55

basically writing the reward functions

play10:57

for an AI system and it was able to do

play10:59

it very effectively and it literally

play11:01

just went from simulation to real life

play11:03

which means that this is going to

play11:04

immediately speed up how quickly we're

play11:07

able to train robots and have them you

play11:09

know doing well in novel environments

play11:11

it's pretty pretty insane so it's an llm

play11:13

guided Sim toal approach so that is

play11:16

pretty crazy and remember I'm guessing

play11:17

that they were using GPT 4 that's

play11:19

actually a study that I haven't covered

play11:21

just yet but the point is is that samman

play11:23

states that the current state-of-the-art

play11:25

model the current model that other

play11:27

industry labs are trying to be millions

play11:29

and millions of dollars he states that

play11:31

that model is bad and it's dumb he

play11:33

didn't just say it was kind of smart he

play11:35

said that it was dumb which leads me to

play11:37

believe that we are truly truly not even

play11:40

scratching the surface for what AI

play11:41

systems are he could have just said

play11:43

future models will be interesting he

play11:44

could have just said they will be you

play11:46

know kind of good but he literally said

play11:48

okay I'm going to play it one more time

play11:49

first of all that's nice of you to say

play11:51

but Chachi BT is not phenomenal like

play11:52

Chachi PT is mildly embarrassing at best

play11:55

um GPT 4 is the dumbest model any of you

play11:57

will ever ever have to use again Again

play11:59

by a lot um but you know it's like

play12:02

important to ship early and often so

play12:04

that just goes to show how clear there's

play12:06

going to be a distinction in the future

play12:08

now he also stated that there will be a

play12:10

massive jump from DPT 3.5 to 4 and there

play12:14

will be a similar jump from GPT 4 to GPT

play12:18

5 so it's important to know the jump

play12:20

from GPT 3.5 to GPT 4 was incredible

play12:23

because of GPT 3.5 limitations it meant

play12:26

that it couldn't be extended to certain

play12:28

tasks but if we have that same jump from

play12:30

GPT 4 to GPT 5 then things are truly

play12:34

about to change there's so many

play12:36

questions uh first of all also amazing

play12:39

it's looking back it'll probably be this

play12:41

kind of historic pivotal moment with 35

play12:44

and four which had your BT maybe five

play12:47

will be the pivotal moment I don't know

play12:49

hard to say that looking forwards we

play12:50

never know that's the annoying thing

play12:52

about the future it's hard to predict

play12:54

but for me looking back GPT 4 Chad GPT

play12:57

is pretty damn impressive like

play12:59

historically impressive so allow me uh

play13:02

to ask what's been the most impressive

play13:04

capabilities of GPT 4 to you and gp4

play13:07

turbo I think it kind of sucks H typical

play13:10

human also gotten used to an awesome

play13:12

thing no I think it is an amazing thing

play13:15

um but relative to where we need to get

play13:18

to and where I believe we will get to uh

play13:20

you know at the time of like gpt3 people

play13:22

were like oh this is amazing this is

play13:24

this like Marvel of technology and it is

play13:26

it was uh you know now we have gp4 and

play13:29

look at GB3 and you're like that's

play13:30

unimaginably horrible um I expect that

play13:33

the Delta between 5 and four will be the

play13:35

same as between four and three and I

play13:37

think it is our job to live a few years

play13:40

in the future and remember that the

play13:41

tools we have now are going to kind of

play13:43

suck looking backwards at them so you

play13:45

can clearly see here that he's basically

play13:47

stating that a year from now when we

play13:49

have gbt 5 or the next level of Frontier

play13:52

Model is released we're going to look

play13:54

back at GPT 4 and think it's pretty

play13:57

pretty bad so as much as people are

play13:59

stating that AI hype is wearing off Sam

play14:01

alman's feeling it there's this Gartner

play14:03

hype cycle it's important to remember

play14:05

the subtle cues the subtle tells the

play14:07

subtle statements that we've clearly

play14:09

seen from industry leaders about these

play14:11

future future models that only they

play14:13

currently have access to now if

play14:15

generative AI was actually slowing down

play14:18

it would be due to a bottleneck and

play14:20

there are only a couple bottlenecks that

play14:23

are really the issues okay one of the

play14:25

issues is of course energy and this one

play14:28

isn't really talked about enough because

play14:29

energy isn't something that people think

play14:31

of but trust me when I say these

play14:33

inference costs are really really

play14:35

expensive and a lot of people you know

play14:37

including Mark Zuckerberg are basically

play14:39

stating that one of the major

play14:41

limitations for future AI systems and

play14:44

where things might actually start to

play14:45

slow down is due to the energy cost the

play14:48

rampant energy cost for these AI systems

play14:50

and just how much electricity they

play14:52

consume and there were even a few rumors

play14:55

talking about a GPT 6 training cluster

play14:57

project that would arguably shut down

play14:59

the power grid something along those

play15:01

lines I know it does sound crazy and I

play15:02

know rumors are rumors but energy is

play15:05

expensive and a lot of it is required to

play15:07

run inference on these large language

play15:09

models which is why they often restrict

play15:11

us I think there was this issue of um

play15:13

GPU production yeah right so even

play15:16

companies that had the money to pay for

play15:17

the gpus um couldn't necessarily get as

play15:20

many as they wanted because there was

play15:22

there were all these Supply constraints

play15:23

now I think that's sort of getting less

play15:26

so now I think you're seeing a bunch of

play15:28

companies

play15:29

think about wow we should just like

play15:31

really invest a lot of money in building

play15:33

out these things and I think that will

play15:34

go for um for some period of time there

play15:37

is a capital question of like okay at

play15:40

what point does it stop being worth it

play15:42

to put the capital in but I actually

play15:44

think before we hit that you're going to

play15:45

run into energy constraints right

play15:47

because I just I mean I don't think

play15:49

anyone's built a gigawatt single

play15:52

training cluster yet I me just to I

play15:54

guess put this in perspective I think a

play15:56

gigawatt it's like around the size of

play15:59

like a meaningful nuclear power plant

play16:01

only going towards training a model and

play16:04

then you run into these things that just

play16:05

end up being slower in the world like

play16:07

getting energy permitted is like a very

play16:11

heavily regulated government function

play16:14

and if you're talking about building

play16:16

large new power plants or large build

play16:18

outs and then building transmission

play16:20

lines that cross other private or public

play16:23

land that is just a heavily regulated

play16:26

thing so you're talking about many years

play16:28

of lead time so if we wanted to stand up

play16:31

just some like massive facility um to

play16:33

power that I I think that that is that's

play16:37

that's a very long-term project so

play16:39

basically what Mark Zuckerberg is

play16:40

stating here is that energy is going to

play16:42

be a huge bottleneck because unlike

play16:44

software where you can make it more

play16:45

efficient and you can do things quickly

play16:47

or you can get a GPU produced quickly

play16:49

trying to build a nuclear power plant it

play16:51

takes time trying to build this these

play16:53

kind of infrastructure it's not

play16:54

something you could just do in a day and

play16:56

it's something that's quite important

play16:58

when it comes to ative AI because as the

play17:00

race continues they're going to be

play17:01

spending a lot more money they're going

play17:02

to be spending a lot more in terms of

play17:05

how much they're spending on cooling

play17:06

because these gpus heat up quite a bit

play17:09

so I think this is something that you

play17:12

know this is probably where the

play17:14

constraints actually will come from in

play17:16

the future and this is where things

play17:18

might actually slow down a little bit

play17:19

because there are going to be a few

play17:21

things that make this truly hard to

play17:23

continue with I think we would probably

play17:26

build out bigger clusters than we

play17:28

currently can if we could get the energy

play17:31

to do it so I think that that's um

play17:34

that's fundamentally money bottl in the

play17:37

limit like if you had a trillion dollar

play17:38

I think it's time right um but it

play17:41

depends on how far the the exponential

play17:42

curves go right like I think a number of

play17:45

companies are working on you know right

play17:46

now I think you know like a lot of data

play17:48

centers are on the order of 50 megawatts

play17:50

or 100 megawatts or like a big one might

play17:52

be50 megawatts okay so you take a whole

play17:54

Data Center and you fill it up with just

play17:56

all the stuff that you need to do for

play17:58

training and you build the biggest

play17:59

cluster you can I think you're that's

play18:01

kind of I think a bunch of companies are

play18:03

running at stuff like that so it will be

play18:04

interesting to see where gen of AI does

play18:05

fall down but like I've said it's not

play18:07

slowing down anytime soon and if you

play18:09

remember open a and Microsoft are

play18:12

building a100 billion Stargate AI

play18:14

supercomputer to power the AGI or ASI

play18:18

now there's also another bottleneck

play18:20

which is what I've titled here which is

play18:22

the compute problem essentially this

play18:24

just means that the compute capacity for

play18:27

AI systems is far too great and it kind

play18:30

of exceeds the demand that we require so

play18:32

that's why they're building out this1

play18:34

billion supercomputer to meet the

play18:36

demands of future generative AI systems

play18:39

or to essentially power the next

play18:41

Industrial Revolution because my oh my

play18:43

if an AGI or air size here it's going to

play18:45

be used pretty much everywhere to power

play18:48

the economy and you're going to need the

play18:50

compute and the infrastructure to do

play18:52

that and currently this is one of our

play18:54

biggest things because openi and

play18:56

Microsoft don't even have enough chips

play18:58

and don't even have enough computes

play19:00

currently to compete with the likes of

play19:02

Google so what we have here is we have

play19:05

updates to the chips and you can see

play19:08

that nvidia's recent Blackwell is pretty

play19:10

pretty incredible and this was one of

play19:13

the most important developments for AI

play19:15

because this accelerates the training of

play19:17

large language models and generative AI

play19:19

systems so the Blackwell GPU

play19:22

architecture with its 208 billion

play19:24

transistors and its enhanced Transformer

play19:26

engine is designed to dramatically

play19:28

increase increase the training of large

play19:30

language models like GPT 4 according to

play19:33

Nvidia Blackwell can provide up to 30

play19:36

times higher performance for generative

play19:37

AI inference compared to the previous

play19:39

h100 gpus four times faster training

play19:42

performance for large language models

play19:44

and this essentially means large

play19:46

language models like GPT 4 which took

play19:48

around 90 days to train on 8, h100 gpus

play19:52

consuming 15 megawatt of power could

play19:54

potentially be just trained in just 30

play19:57

days on two 2,000 black C gpus only

play20:00

using 4 megawatt which is pretty pretty

play20:04

incredible and that curve definitely

play20:06

reminds me of some that we've seen

play20:08

before where things are starting to

play20:10

exponentially increase so the point

play20:13

right here is that these are the actual

play20:15

bottlenecks of generative AI because a

play20:17

lot of people are thinking that you know

play20:19

things are slowing down things are

play20:21

getting worse and trust me guys if you

play20:22

been paying attention things I would

play20:24

argue are actually speeding up

play20:26

internally and one of the things that I

play20:28

didn't even include in this presentation

play20:30

because I forgot about it is the fact

play20:31

that right now it's like open ey lit a

play20:34

match under every other company because

play20:37

now other companies are realizing that

play20:38

whoa there's a huge huge AI race going

play20:41

on and if we partake we could definitely

play20:44

be getting billions and billions of

play20:45

dollars and that means that other

play20:47

companies and other startups are all

play20:49

rushing down the corridor to see if they

play20:51

can get piece of the piie which means

play20:54

that we're about to see a complete

play20:55

Revolution and a complete new industry

play20:57

in terms of all of these products and

play20:59

services now one of the biggest things

play21:01

that I think that most people need to

play21:04

consider is that open AI are no longer

play21:07

open okay and if there's one thing you

play21:09

take away from this video please

play21:11

understand this okay things might be

play21:12

slowing down externally but things are

play21:15

not slowing down internally and what I

play21:17

mean by that statement is that currently

play21:19

we're at a stage where things have moved

play21:21

from an open research environment to a

play21:24

closed research environment the reason

play21:25

this has happened is because opening up

play21:27

they're no longer essentially a company

play21:29

that's just focused on Research they are

play21:31

a business and businesses you know they

play21:33

hide their secrets and they hide their

play21:35

Innovations because they don't want

play21:36

their competitors to have them if open I

play21:38

shared all their secrets then other

play21:40

companies could easily build gbt 4 with

play21:42

remarkable accuracy and op essentially

play21:46

has secret Source the thing is openai

play21:48

also doesn't publicize their research

play21:51

I'm sure breakthroughs are made every

play21:53

single month okay and you have to think

play21:55

about it if openai did their whatever

play21:57

they did with gbt so long ago they must

play22:00

have some secret kind of breakthrough

play22:02

they must have some secret source and

play22:04

they must have something that others

play22:05

don't which essentially means that open

play22:07

ey are making consistent breakthroughs

play22:09

and remember Sora they had Sora we had

play22:12

no news no indication that they were

play22:14

even developing some video AI there was

play22:16

literally no indication whatsoever there

play22:18

was no interview from Sam Alman there

play22:20

was literally nothing we could have

play22:21

picked up on on the fact that they were

play22:22

even training such a model and boom they

play22:25

just you know put it out into the open

play22:26

the point is is that we know we have no

play22:28

idea on what's going on at you know

play22:30

closed AI open AI whatever you want to

play22:32

call it the point is is that internally

play22:34

I can guarantee you guys they are like 2

play22:35

to 3 years ahead from where they are and

play22:38

the point there is that whilst you might

play22:40

think ah they haven't released anything

play22:42

in a while that doesn't mean things are

play22:43

slowing down it just means that they're

play22:45

thinking okay how can we not Shock the

play22:47

public with this next release that we

play22:49

know is literally going to take

play22:51

everything remember other companies are

play22:53

still playing catchup gbt 4 finished

play22:55

training in August of 2022 which means

play22:58

that we are very very lucky because

play23:01

we're going to be in for a real surprise

play23:03

when gbt 5 gets here now another thing

play23:06

to note as well okay is that GPT 4 being

play23:08

The Benchmark does not mean Plateau the

play23:11

problem with this and like I said before

play23:13

this is a business which mean things are

play23:15

going to change GPT 4 is currently the

play23:18

Benchmark which means that companies are

play23:21

incentivized to train their models to

play23:23

surpass GPT 4 and then release that

play23:26

model the reason this creates the

play23:29

illusion that things are plateauing

play23:31

around GPT 4 is because these companies

play23:33

are no longer incentivized to go duly

play23:36

pass GPT 4 they're only incentivized to

play23:39

just beat it and that is because of

play23:41

course with GPT 4 that is something that

play23:44

people State oh it's the best system is

play23:45

the best system so if another company

play23:48

like gemini or anthropic or Google can

play23:51

come out and say look our system

play23:52

surpasses gbt 4 or benchmarks they're

play23:54

going to immediately release that model

play23:57

after it's fine- tuned or after after

play23:58

it's whatever they've done with it and

play24:00

then run with that so that they can

play24:02

Market that and get the customer base

play24:04

because they know that open AI are

play24:06

waiting to release GPT 5 potentially

play24:08

after the elections and that gives them

play24:10

some time to reclaim the market share

play24:13

understand that where GPT 4 is is just

play24:15

an indication of where other models are

play24:17

going to stop and if you think that that

play24:19

is just a pure speculative argument look

play24:22

at how close gp4 is to some of these

play24:24

models you have to understand that if

play24:26

they didn't beat gp4 they wouldn't be

play24:28

releasing these models this is

play24:29

86.4% they literally got it up to 86.8%

play24:33

another one here 92% they got this up to

play24:35

95% okay it's not like it's completely

play24:38

surpassing them and I guess some people

play24:39

like look it all you know slows down

play24:41

around here no they just want it to be

play24:44

as close as possible so that they can

play24:45

get this out as quickly as possible

play24:47

because they know by the time open hour

play24:49

releases next again they're going to be

play24:50

even behind and you can see here that

play24:52

Gemini Ultra a lot of people were even

play24:54

debating this because this was Chain of

play24:55

Thought at 32 um and that what they did

play24:58

in order to beat this because I'm

play25:00

guessing that when they had finished

play25:02

training the model and when they

play25:03

finished fine-tuning it they had to you

play25:05

know increasingly developed certain

play25:07

methods just to get this metric right

play25:09

here and that's why I state that gp4

play25:12

being the Benchmark does not mean we're

play25:14

currently at a plateau at all because

play25:16

it's likely that these companies are

play25:17

just benchmarking their models up to GPT

play25:20

4 so that they can get them out now

play25:22

here's why things are going to go even

play25:24

crazy and remember I said this because

play25:26

the next 365 days are going to be

play25:28

absolutely insane agents are still early

play25:31

okay and someone actually recently

play25:32

created a benchmark where they're

play25:34

talking about multimodal agents for

play25:36

open-ended tasks in real computer

play25:38

environments and essentially with this

play25:40

you can see that humans can accomplish

play25:41

over 72% of the task and the best

play25:44

current AI agent can only do

play25:46

12.24% what happens when AI agents can

play25:49

get to above 80% that is truly going to

play25:52

change everything and with the advanced

play25:54

reasoning and with the advanced

play25:56

capabilities of future models we going

play25:58

to see a future that we've never seen

play26:00

before now there was also something that

play26:02

I covered in a previous video that I'm

play26:04

guessing the majority of people are just

play26:06

completely glossing over and I'm sure

play26:08

it's because during the video I was kind

play26:10

of sick because I had some kind of flu

play26:12

whatever but I still made the video

play26:14

anyways so essentially there was this

play26:16

thing right here okay this is mesa's kpu

play26:19

now this is a little bit speculative

play26:21

because they haven't released too much

play26:22

information but if you check the

play26:23

benchmarks here you can see that this

play26:25

surpasses claw 3 Opus Gemini Ultra and

play26:27

Mr large at all benchmarks okay and this

play26:30

is because they use a advanced reasoning

play26:33

engine on top of gp4 Turbo now this is

play26:36

pretty interesting because this shows us

play26:38

that we are still very early on the

play26:41

reasoning capabilities which is why I

play26:43

argue that samman here says that gp4 is

play26:46

dumb and why he also says here that it

play26:48

was not a very good AI system so there

play26:51

was one demo released by M's kpu in

play26:53

which they showcased an AI system

play26:56

actually doing reasoning with an

play26:57

advanced task and recently on their

play27:00

Twitter I'm not sure why it's not

play27:01

getting any love or any actual you know

play27:03

tweets about it they've shown that this

play27:06

system is able to you know use some

play27:08

reasoning steps and this is their system

play27:10

they're messing around with it and

play27:11

they're showing that it's able to

play27:13

complete a lot of tasks really really

play27:15

correctly so you have to remember that

play27:17

internally things are going at light

play27:18

speed things like qar things like other

play27:20

companies now trying to get a piece of

play27:22

the pie uh incorporating different

play27:24

reasoning engines on top of gbt 4 are

play27:27

going to push things further and

play27:28

remember it was recently that Andrew NG

play27:31

actually spoke about agentic workflows

play27:34

and basically said that GPT 3.5 zero

play27:36

shot was

play27:38

48.3 five and of course it was Andrew NG

play27:41

that did some research and found out

play27:43

that GPT 3.5 zero shot was 48.1% correct

play27:46

GPT 4 zero shot does better at 67% but

play27:50

the Improvement was dwarfed by

play27:52

incorporating an iterative agent

play27:54

workflow and wrapped in an agent Loop

play27:56

GPT 3.5 gets up to 95.1% so the point

play28:00

here is that there's still a lot of

play28:02

different architectures that we haven't

play28:04

fully explored with some of the AI

play28:05

systems that we do have which means that

play28:08

we are far far far away from any sort of

play28:11

plateau and things are going to keep

play28:13

increasing number one we've got the gpus

play28:16

increasing in terms of efficiency we've

play28:18

got the data centers we've got all of

play28:21

these things getting increasingly better

play28:23

and of course we've got the fact that

play28:24

internally open AI they are blisteringly

play28:27

so far ahead that I'm guessing that

play28:29

things are going to be shocking when

play28:30

they are finally released

Rate This

5.0 / 5 (0 votes)

Related Tags
AI HypeGenerative AIGPT ModelsTech InnovationSam AltmanHype CycleAI EfficiencyEnergy BottleneckCompute PowerIndustry TrendsAI AdvancementsTech RevolutionFuture PredictionsAI ResearchBusiness ImpactEconomic ModelLarge Language ModelsAI DevelopmentTech RaceSustainable Tech