New AI Chip, GPT4o, Claude 3.5, SpaceX Double Landing, AI Video Games

Matthew Berman
29 Jun 202412:23

Summary

TLDRThe video script discusses recent AI advancements, focusing on Etched's new AI chip 'Sohu', which promises to outperform GPUs in processing speed and cost. It also covers OpenAI's delayed voice capabilities, Hugging Face's new LLM leaderboard highlighting 'Quen 72b' as a top performer, and Claude 3.5's coding prowess. The script wraps up with AI-generated videos resembling real-time gaming and Apple's decision not to integrate Meta's AI models into Siri due to privacy concerns.

Takeaways

  • ๐Ÿš€ A new AI chip company, Etched, has developed a chip called Sohu that claims to generate over 500,000 tokens per second running Llama 70b, which is specialized for Transformer models.
  • ๐Ÿ”‹ Sohu is said to be more efficient than GPUs, with one server equipped with eight Sohu chips replacing 160 Nvidia H100s, though it's not yet in production.
  • ๐Ÿ’ก The chip's specialization for Transformer models is likened to how ASICs were created for Bitcoin mining, suggesting a shift towards specialized hardware for AI tasks.
  • ๐Ÿ“‰ GPUs are not improving significantly, with a 15% improvement in compute density over four years, indicating a need for specialized chips to enhance performance.
  • ๐ŸŒ Open AI's voice capabilities, which were anticipated to be released, are delayed to further improve the model's content detection and refusal abilities, and infrastructure scalability.
  • ๐ŸŽ™๏ธ Open AI's advanced voice mode is expected to roll out in Alpha to a small group of users in late June, with a full rollout planned for the fall.
  • ๐Ÿ† Hugging Face has launched a new open LLM leaderboard, with Quen 72b emerging as the top performer, indicating the dominance of Chinese open models in AI.
  • ๐Ÿ“Š There's a concern that AI model makers are focusing too much on public benchmarks, potentially at the expense of overall model performance.
  • ๐Ÿฅ‡ CLA 3.5 (Sonaut) has achieved the top spot in coding and hard prompts, showcasing its capabilities against other leading models like GPT-40 and Gemini 1.5 Pro.
  • ๐ŸŽฎ AI-generated video content, resembling a Call of Duty game, demonstrates the potential future of video games, though real-time generation requires significant computational power.
  • ๐Ÿ”„ Reports suggest that Apple was in talks with Meta AI to integrate Llama 3 into Siri but has since decided against it due to privacy concerns, despite Apple's capacity to host the model themselves.

Q & A

  • What is the new AI chip company mentioned in the script called, and what is its claim to fame?

    -The new AI chip company is called Etched, and it claims to be able to generate over 500,000 tokens per second running Llama 70b, with a chip named Sohu that is specialized for Transformer models.

  • How does the Sohu chip compare to Nvidia's H100 in terms of performance and efficiency?

    -One server with eight Sohu chips is said to replace 160 Nvidia H100s. Sohu is more than 10 times faster and cheaper than Nvidia's next-generation Blackwell GPUs, running over 500,000 Llama 70b tokens per second compared to H100's 23,000 tokens per second.

  • What does the script suggest about the future of AI models and hardware specialization?

    -The script suggests that within a few years, every large AI model will run on custom chips, which are more than 10 times faster and cheaper than current GPUs, indicating a shift towards specialized hardware for AI models.

  • Why is OpenAI delaying the release of its advanced voice mode?

    -OpenAI is delaying the release of its advanced voice mode to improve the model's ability to detect and refuse certain content, and to further enhance the user experience and infrastructure to scale to millions of users while maintaining real-time responses.

  • What is the significance of the new open LLM leader board announced by Hugging Face's CEO?

    -The new open LLM leader board is significant as it provides a comprehensive evaluation of major open LLMs, with Quen 72b emerging as the top performer, indicating a shift in the dominance of AI models and the importance of specialized benchmarks.

  • What is the current status of the integration talks between Apple and Meta's AI models for Siri?

    -According to recent reports, Apple is no longer considering integrating Meta's AI models into Siri due to privacy concerns, despite previous talks suggesting otherwise.

  • What does the script imply about the potential impact of AI-generated content on the future of video games?

    -The script implies that AI-generated content, as demonstrated by the realistic AI-rendered video, could revolutionize the video game industry by enabling highly realistic and immersive gaming experiences.

  • What is the script's perspective on the importance of specialized AI chips like Sohu for the future of AI development?

    -The script highlights the importance of specialized AI chips like Sohu for the future of AI development, suggesting that they will become the standard for running large AI models due to their superior performance and cost-effectiveness.

  • What is the script's view on the current state of GPUs and their limitations in AI model performance?

    -The script suggests that GPUs are not improving at a rate that matches the needs of AI model performance, with compute density only improving by 15% in the past four years, indicating a need for more specialized hardware.

  • How does the script describe the potential of AI in creating realistic video content, as shown in the AI-generated video?

    -The script describes the potential of AI in creating realistic video content as impressive and mind-blowing, with the AI-generated video showcasing high-quality visuals and sound that are almost indistinguishable from real footage.

  • What is the script's opinion on the role of benchmarks in evaluating AI models?

    -The script suggests that benchmarks are crucial for evaluating AI models, but there is a concern that model makers might be focusing too much on major public benchmarks at the expense of overall model performance.

Outlines

00:00

๐Ÿš€ Revolutionary AI Chip 'Sohu' and Its Impact on the Future of AI

The script discusses the emergence of a new AI chip company called Etched, which has developed a chip named 'Sohu' capable of generating over 500,000 tokens per second using the Llama 70b model. This chip is specialized for Transformer models, similar to how ASICs were created for Bitcoin mining. 'Sohu' promises performance that surpasses Nvidia's H100 GPUs, suggesting a future where AI models will predominantly run on custom chips. The company emphasizes the importance of specialization in improving performance and reducing costs, and they predict a shift towards hardware tailored for AI models, potentially rendering general-purpose GPUs obsolete for AI tasks.

05:02

๐ŸŽ™๏ธ Open AI's Delayed Voice Capabilities and the New Open LLM Leaderboard

The script addresses the delay in Open AI's voice capabilities, which were expected to launch but have been postponed to ensure high safety and reliability standards. Open AI plans to roll out advanced voice mode in Alpha to a select group of users and aims for broader access by fall. Additionally, the CEO of Hugging Face introduces a new leaderboard for evaluating major open LLMs, revealing that Quen 72b outperforms others, and there's a concern that model makers might be focusing too much on benchmark scores rather than overall performance. The script also highlights the impressive performance of Anthropic's CLA 3.5 model, which has secured top spots in various categories.

10:02

๐ŸŽฎ AI's Role in the Future of Video Games and Apple's Decision on AI Integration

The script explores the potential of AI in creating realistic and immersive video game experiences, as demonstrated by an AI-generated video that mimics the quality of a Call of Duty game. It also discusses the recent reports about Apple's consideration to integrate Meta AI's Llama 3 model into Siri, which was later dismissed due to privacy concerns. The summary touches on the implications of such integrations and Apple's stance on privacy, suggesting that Apple might have chosen to maintain control over user data by not proceeding with the integration.

Mindmap

Keywords

๐Ÿ’กAI chips

AI chips, or artificial intelligence chips, are specialized hardware designed to accelerate machine learning tasks and AI applications. In the video, the company 'Etched' is mentioned for developing a chip called 'Sohu', which is capable of generating a high volume of tokens per second for AI models, specifically for the Transformer model architecture. This is a significant development as it suggests a shift from using general-purpose GPUs to specialized chips for AI, enhancing performance and efficiency.

๐Ÿ’กTransformer models

Transformer models are a type of deep learning architecture that has become foundational in the field of natural language processing. The script refers to the Sohu chip as the first specialized chip for these models, indicating a move towards hardware tailored to the specific needs of AI models that use the Transformer architecture, such as llama 70b, which is mentioned in the context of the chip's capabilities.

๐Ÿ’กGPUs

GPUs, or Graphics Processing Units, are specialized electronic hardware designed to handle complex mathematical and graphical calculations. They have been widely used for AI and machine learning tasks due to their parallel processing capabilities. The video discusses the limitations of GPUs in comparison to specialized AI chips, suggesting that the latter may offer superior performance for certain AI tasks.

๐Ÿ’กNvidia H100s

Nvidia H100s are high-performance GPUs designed for AI and machine learning workloads. In the script, they are used as a benchmark to illustrate the superior performance of the Sohu chip, with one server equipped with eight Sohu chips replacing the computational power of 160 Nvidia H100s.

๐Ÿ’กASICs

ASICs, or Application-Specific Integrated Circuits, are custom-made chips designed for a specific function or set of functions. The script mentions ASICs in the context of Bitcoin mining, where they became more cost-effective than using general GPUs. It suggests a similar trend may occur with AI chips, where specialized hardware will become more prevalent for AI tasks.

๐Ÿ’กOpen AI

Open AI is a research organization focused on the development and application of AI technologies. The video discusses Open AI's voice capabilities, which are in development and expected to be rolled out in an advanced mode for Chat GPT, indicating the organization's ongoing work in improving AI interaction through natural language and voice processing.

๐Ÿ’กChat GPT Plus

Chat GPT Plus is a subscription service that offers enhanced features for users interacting with the Chat GPT AI model. The script mentions that the advanced voice mode will initially be rolled out to a small group of Chat GPT Plus users, highlighting the service's role in testing and deploying new AI functionalities.

๐Ÿ’กHugging Face

Hugging Face is a company that focuses on natural language processing and provides tools and libraries for AI development. The CEO of Hugging Face is mentioned in the script for announcing a new open LLM leaderboard, which evaluates and ranks the performance of various AI models, indicating the company's role in benchmarking and advancing AI capabilities.

๐Ÿ’กCLA 3.5

CLA 3.5, or Claude 3.5, is an AI model developed by Anthropic, a company focused on AI safety and capabilities. The video discusses its performance on various benchmarks, particularly noting its top ranking in coding and hard prompts, showcasing the model's advanced capabilities in these areas.

๐Ÿ’กAI-generated video

The script refers to a video rendered completely using AI, which demonstrates the potential of AI in creating realistic and high-quality visual content. This AI-generated video is likened to a Call of Duty game, indicating the future possibilities of AI in the gaming industry and its ability to create immersive and realistic experiences.

๐Ÿ’กPrivacy concerns

Privacy concerns are highlighted in the context of Apple's decision not to integrate Meta AI's Llama model into Siri due to potential privacy issues. This underscores the importance of data protection and user privacy in the development and integration of AI technologies, especially when considering partnerships between tech companies.

Highlights

AI chip company 'Etched' claims its new chip 'Sohu' can generate over 500,000 tokens per second running Llama 70b, outperforming Nvidia GPUs.

Sohu is the first specialized chip for Transformer models, similar to how ASICs were created for Bitcoin mining.

Etched's Sohu chip is not yet in production, but promises more than 10 times the performance of Nvidia's next-generation GPUs.

Open AI's voice capabilities are delayed for further improvements, including the model's ability to detect and refuse certain content.

Open AI plans to roll out advanced voice mode in Alpha to a small group of users in late June, with full access expected in the fall.

Hugging Face's CEO introduces a new open LLM leaderboard, revealing that 'Quen 72b' is the top-performing model.

The new Sohu chip is expected to make every large AI model run on custom chips within a few years, marking a shift in AI hardware specialization.

AI Builders may be focusing too much on main evaluations at the expense of overall model performance.

CLA 3.5 SAA secures the top spot in coding and hard prompts, showcasing its competitive edge against other models like GPT-40.

Anton OA demonstrates Claude 3.5's performance against GPT-40 in coding, highlighting its success in task and full project completion.

Elon Musk shares a video of two rockets landing for reuse, showcasing a significant achievement in space engineering.

AI-generated video content is becoming increasingly realistic, hinting at the future of video games as predicted by Nvidia's CEO.

Apple was reportedly in talks with Meta AI to integrate Llama 3 into Siri but has since dropped the idea due to privacy concerns.

The decision to not integrate Meta AI's models into iPhones may also be influenced by Apple's criticism of Meta's privacy practices.

AI labs are spending millions optimizing kernels for Transformers, indicating a significant investment in AI model efficiency.

Startups are utilizing special Transformer software libraries to enhance features like speculative decoding and tree search.

Once Sohu hits the market, it is expected to reach a point of no return for AI model efficiency and specialization.

Transcripts

play00:00

again it's just been a few days since my

play00:02

last news video and so much has happened

play00:04

we're going to be talking about AI chips

play00:06

open AI Apple Cloud 3.5s dominance and

play00:10

what looks to be the future of video

play00:12

games let's get into it the first story

play00:14

today is about a new AI chip company

play00:16

called etched and it claims to be able

play00:19

to generate over

play00:20

500,000 tokens per second running llama

play00:24

70b so this chip is called sohu I

play00:27

believe and it lets you build products

play00:28

that are impossible on gpus one server

play00:31

with eight of these

play00:33

replaces 160 Nvidia h100s and this is

play00:36

all assuming this is true because I

play00:38

don't think this is actually in

play00:39

production yet now here's how it's able

play00:42

to do that soou is the first specialized

play00:44

chip as6 for Transformer models now if

play00:47

you remember in the earlier days of

play00:49

crypto mining specific chips were

play00:51

created for Bitcoin mining and these

play00:54

chips were especially made just to crack

play00:58

that algorithm and this chip is specific

play01:00

to the Transformer model and they even

play01:02

say that we get way more performance

play01:05

sohu can't run CNN's convolutional

play01:08

neural networks lstms ssms or any other

play01:11

AI model every major AI product chbt CLA

play01:15

Gemini Sora is powered by Transformers

play01:17

within a few years every large AI model

play01:19

will run on custom chips it is more than

play01:22

10 times faster and cheaper than even

play01:24

nvidia's Next Generation Blackwell gpus

play01:28

one so hu server runs over 500,000 llama

play01:31

70b tokens per second 20 times more than

play01:34

an h100 23,000 tokens per second and 10

play01:37

times more than a B2 45,000 tokens per

play01:40

second and here are the benchmarks now

play01:42

interestingly enough they do not compare

play01:44

themselves to Gro which I wonder why and

play01:47

it says here gpus aren't getting better

play01:49

they're just getting bigger in the past

play01:51

four years compute density has only

play01:52

improved by 15% next gen gpus are now

play01:56

counting two chips as one card to double

play01:58

their performance with with Mo's loss

play02:00

slowing the only way to improve

play02:02

performance is specialization and they

play02:04

specifically call out Bitcoin mining

play02:05

here so when Bitcoin miners hit the

play02:07

market in 2014 it became cheaper to

play02:09

throw out gpus than to use them to mine

play02:11

Bitcoin and that's exactly what happened

play02:14

everybody started transitioning to as6

play02:16

they go on to say Transformers have a

play02:18

huge moat we believe in the hardware

play02:20

Lottery the architecture that wins is

play02:21

the one that runs fastest and cheapest

play02:23

on Hardware I'm not really sure what

play02:26

they meant by Hardware lottery in this

play02:28

context typically the hardware Lottery

play02:30

means when you get a GPU some of them

play02:32

are just made better than others AI labs

play02:34

to spend hundreds of millions of dollars

play02:36

optimizing kernels for Transformers

play02:38

startups use special Transformer

play02:39

software libraries like trt llm and VM

play02:42

which offers features built on

play02:44

Transformers like speculative decoding

play02:46

and tree search Once soou hit the market

play02:49

we will reach the point of no return

play02:50

Transformer Killers will need to run

play02:52

faster on gpus than Transformers run on

play02:54

so hu if that happens we'll build an

play02:56

Asic for that tool and on their website

play02:59

scale is all you need for super

play03:00

intelligence so if you want to learn

play03:02

more about soou and the technical

play03:04

details check out the website et.com and

play03:07

I'll drop a link to that in the

play03:08

description below next it seemed like

play03:11

the open AI voice capabilities were

play03:13

taking forever to come out and now we

play03:15

know why they're actually delaying them

play03:18

and as a reminder the GPT 40 voice

play03:20

capabilities were essentially that movie

play03:22

Her hey chat GPT how are you doing I'm

play03:25

doing fantastic thanks for asking how

play03:27

about you all right so we're sharing

play03:29

update on the advanced voice mode we

play03:31

demoed during the spring update which we

play03:33

remain very excited about we plan to

play03:35

start rolling this out in Alpha to a

play03:36

small group of chat GPT plus users in

play03:38

late June but need one more month to

play03:40

reach our bar for launch for example

play03:42

we're improving the model's ability to

play03:44

detect and refuse certain content this

play03:46

is disappointing this is very similar to

play03:48

what it felt like when they announced

play03:50

and demoed Sora and then we basically

play03:52

have no idea when that's actually coming

play03:54

and I'm guessing they were able to

play03:55

jailbreak it we're also working on

play03:57

improving the user experience and

play03:58

preparing our infrastructure to scale to

play04:00

Millions while maintaining real-time

play04:02

responses as part of our iterative

play04:04

deployment strategy we'll start the

play04:05

alpha with a small group of users to

play04:07

gather feedback and expand based on what

play04:09

we learned we are planning for all plus

play04:11

users to have access in the fall so

play04:13

that's great right around the corner

play04:15

exact timelines depend on meeting our

play04:16

high safety and reliability bar we are

play04:19

also working on rolling out a new video

play04:20

and screen sharing capabilities we

play04:22

demoed separately and will keep you

play04:24

posted on that timeline chat gpt's

play04:26

advanced voice mode can understand and

play04:27

respond with emotions and non-verbal

play04:29

cues moving us closer to realtime

play04:30

natural conversations with AI great and

play04:33

Billow wall responds with not sure what

play04:35

the point of flexing a live demo was if

play04:37

the intention was to delay the launch

play04:39

like this clearly the product wasn't

play04:41

ready I could probably say the same

play04:43

thing about Microsoft's recall which was

play04:47

recalled both of these things I was

play04:49

incredibly excited about but now we have

play04:52

to wait next Clen the CEO and co-founder

play04:55

of hugging face has now announced a new

play04:58

open llm leader board we burned 300

play05:01

h100s to rerun new evaluations like MML

play05:04

U pro for all major open llms some

play05:07

learnings interestingly quen 72b is the

play05:10

king and Chinese open models are

play05:12

dominating overall you all know that I

play05:14

tested quen 72b and yes it did perform

play05:17

very well I would like to see a coding

play05:19

specific flavor of quen 72b previous

play05:22

evaluations have become two easy for

play05:24

recent models that is something that I

play05:26

found at least with my own llm rubric

play05:28

and I asked you for suggestions for new

play05:30

tests in my previous video keep them

play05:32

coming drop it in the comments below if

play05:33

you have suggestions for new tests that

play05:35

I should use going forward on these open

play05:37

source models there are indications that

play05:39

AI Builders have started to focus on the

play05:41

main evaluations too much at the expense

play05:43

of model performances on other ones now

play05:46

interestingly a lot of you have

play05:48

mentioned and suggested that some of

play05:50

these model makers might actually use my

play05:52

tests to train their models I didn't

play05:54

really think that was true because I'm

play05:56

not actually doing a formal Benchmark

play05:58

and it's just just a little YouTube

play06:00

channel but it seems like open- Source

play06:03

model makers are focusing on these major

play06:06

public benchmarks and bigger is not

play06:08

always smarter so that is something that

play06:11

I've been thinking a lot about lately

play06:12

think about this the Llama 3 88b Model

play06:16

is substantially better than the Llama 2

play06:19

7B model so just slightly bigger but

play06:23

much better performance and they've

play06:24

decided to cover the following General

play06:27

tasks knowledge testing reasoning on

play06:29

short and long context complex math

play06:31

abilities and tasks well correlated with

play06:34

human preference like instruction

play06:35

following and to do that they have these

play06:37

benchmarks mlu Pro GP QA M Sr which I've

play06:42

never heard of multip soft reasoning

play06:45

very cool math if eval and BBH and here

play06:48

are the results as you can see here quen

play06:50

72b instruct is number one metal llama

play06:54

370b instruct number two 53 medium 4K

play06:58

instruct number three ye 1.5 34b number

play07:02

four and so on so if you want to learn

play07:05

more if you want to read about the

play07:06

details if you want to see the

play07:07

benchmarks themselves I'll drop a link

play07:09

to hugging face and this specific page

play07:11

in the description below and speaking

play07:14

about benchmarks and leaderboards by

play07:16

lm.org CLA 3.5 SAA has just made a huge

play07:19

leap securing the number one spot in

play07:21

coding arita hard prompts Arena and

play07:23

number two in the overall leaderboard

play07:25

the new Sonet has surpassed Opus at five

play07:27

times lower cost in competitive with

play07:29

with Frontier models GPT 40 and Gemini

play07:31

1.5 Pro across the board now in a

play07:33

previous video I tested claw 3.5 Sonic

play07:36

and yes it is the best model I've ever

play07:38

tested in fact that was the video in

play07:40

which I asked for new tests because it

play07:42

completely demolished my tests so huge

play07:45

congrats to anthropic and now it is

play07:48

number one in coding number one in hard

play07:50

prompts and number two overall and

play07:53

remember this is the sonnet model this

play07:55

is 3.5 Sonet Opus their largest model is

play07:59

still coming 3.5 Opus that is so very

play08:02

cool to see I'm now paying for Claude

play08:04

and that is my go-to model and Anton OA

play08:07

has also showed Claude 3.5s performance

play08:09

versus GPT 40 on coding and this is

play08:13

build success GPT 40 wins Claude 3.5

play08:17

wins on task success and full project

play08:19

success let's see what else he says

play08:21

about it code that compiles fails a bit

play08:23

more for Claude small difference and

play08:25

pass human QA fails significantly less

play08:27

for Claude And even more interesting is

play08:30

the realistic benchmarks qualitative

play08:32

results so he says Claud more verbose

play08:35

nice for long pieces of code makes

play08:36

generation slower which is interesting

play08:39

because I guess the overall generation

play08:41

because the code is longer it might make

play08:43

sense but from what I've noticed the

play08:44

actual tokens per seconds seem to be

play08:46

faster with Claude and generally not

play08:48

desirable in an agent setting which is

play08:50

interesting and I haven't tested if you

play08:52

want to see me test that let me know in

play08:53

the comments does not follow

play08:55

instructions in large prompts as

play08:56

reliably as GPT 40 it tends to miss

play08:58

crucial instr instructions for example

play09:00

how to format output we experimentally

play09:02

switched to Claud for a right long code

play09:05

sub agent at lovable unfortunately we

play09:07

decided to revert one of the things that

play09:09

I'm going to be testing going forward

play09:10

for all the models is its ability to

play09:12

format its output in specific formats

play09:14

and really Json next Elon Musk posted an

play09:18

absolutely incredible video of two

play09:21

rockets Landing for reuse at the same

play09:25

time look at this video This Is The

play09:28

Stuff of s sence fiction imagine how

play09:31

much science and engineering had to go

play09:33

into this to get these massive Rockets

play09:35

to actually be able to come back down to

play09:37

earth land for later reuse so

play09:40

congratulations to SpaceX really cool

play09:43

accomplishment keep up the amazing work

play09:45

and I see Matt wolf commented this will

play09:47

never not be impressive to me completely

play09:49

agreed next Twitter user chubby who

play09:52

always posts really interesting stuff

play09:54

posted this video that is rendered

play09:57

completely using AI this is not real

play09:59

engine this is actually Ai and the

play10:02

quality is mind-blowing so what we're

play10:04

looking at is an AI generated video of

play10:06

what looks to be kind of a Call of Duty

play10:08

is game the sound is AI generated the

play10:10

visuals are AI generated I can see a

play10:13

little bit of morphing but overall it

play10:15

looks incredible and very very realistic

play10:19

now the amount of compute it will take

play10:22

to do this in real time is going to be

play10:25

tremendous and so we're not quite here

play10:27

yet but as Jensen the CEO of Nvidia has

play10:30

said this is truly the future of video

play10:33

games and will really take video games

play10:35

to the next level and just because

play10:37

people were doubting that this was an

play10:39

actual video game created by AI here is

play10:42

another clip of it let's look at this

play10:43

one so here's another version again you

play10:46

could see a little bit of morphing a

play10:48

little bit of clipping here and there

play10:49

but overall it looks really good it

play10:52

again is kind of like a Call of Duty is

play10:55

game all right and last in the quickest

play10:57

turnaround in reporting ever just a

play10:59

couple days ago it was reported that

play11:01

Apple was in talks with meta aai to

play11:04

integrate llama 3 into Siri the same way

play11:07

that open AI is integrated into Siri so

play11:10

not a deep integration just simply an

play11:12

API call but just a day or two later it

play11:15

is now reported that they are not

play11:18

considering that anymore and

play11:19

specifically over privacy concerns which

play11:21

is really surprising because if Apple

play11:24

actually took the Llama model hosted it

play11:26

themselves they would be able to control

play11:28

the Privacy so maybe that means they

play11:30

were depending on meta to actually

play11:32

operate the inference endpoint so days

play11:34

after the Wall Street Journal reported

play11:35

that apple and meta were in talks to

play11:36

integrate the latter's AI models

play11:38

Bloomberg's Mark German said that the

play11:40

iPhone maker was not planning any such

play11:42

move and they were in talks with

play11:44

multiple companies to explore

play11:45

integration and I know they were also in

play11:48

talks with Google or at least it was

play11:49

reported as such they sheld the idea of

play11:52

putting meta AI models on iPhones over

play11:54

privacy concerns and the report also

play11:56

noted that partnering with the social

play11:58

networking company won won't do a lot of

play11:59

good for Apple's image given that the

play12:01

copertino based company has continuously

play12:03

criticized meta's privacy practices very

play12:06

true but again I think they could have

play12:08

easily hosted the model themselves them

play12:10

being apple and all the privacy concerns

play12:13

they would have control over so I'm not

play12:15

sure why they didn't do that so that's

play12:17

it for today thanks for watching if you

play12:19

enjoyed this video please consider

play12:20

giving a like And subscribe and I'll see

play12:22

you in the next one

Rate This
โ˜…
โ˜…
โ˜…
โ˜…
โ˜…

5.0 / 5 (0 votes)

Related Tags
AI ChipsOpenAITransformer ModelsAI BenchmarksTech NewsVideo GamesAI SpecializationCloud ComputingSpaceX RocketsAI Generated Content