New FREE AI Video Generator & Feature Length AI Films!

Theoretically Media
7 Mar 202413:38

Summary

TLDRこの動画は、無料で使える新しいAIビデオジェネレーター「ハイパー」について紹介しています。Google DeepMindの元社員が開発し、1,900万ドル以上の資金調達を成功させたプラットフォームです。テキストからビデオ、画像からビデオ、そしてビデオのリペインティングなどの機能を提供しています。ユーザーインターフェイスは簡単で、HD版とSD版のビデオジェネレーションのオプションがあります。HD版は約2秒間生成され、SD版は4秒間まで生成できます。また、研究によると、AIが完全に生成する映画が近い将来実現される可能性があるとされています。

Takeaways

  • 🌟 新しいAIビデオジェネレーター「ハイパー」が登場し、現在完全無料で利用できます。
  • 🤖 ハイパーは、Google DeepMindの元社員であるYia MeowとZuu Wongによって開発されました。
  • 💰 ハイパーは1,900万ドル以上の資金調達を行い、AIビデオ競争で大きな存在感を見せています。
  • 📝 ハイパーはテキストからビデオ、画像からビデオ、そしてビデオのリペインティング機能を提供しています。
  • 🎨 ハイパーのインターフェースはシンプルで、フルHDのビデオジェネレーションや標準定義のオプションなど、多様な機能があります。
  • 📹 ハイパーで生成されるHDビデオは約2秒間で、延長方法としてエディターでのスローモーションを提案しています。
  • 📚 研究では、GPT-4と画像からテキストのモデルを利用して、詳細な脚本と対応するビジュアルを生成する「ムービーLLM」が提案されています。
  • 🎬 ムービーLLMは、映画のプロット、スタイル、キャラクターを生成し、一貫性のあるスタイルでシーンを制作することを目指しています。
  • 🔍 ムービーLLMは、1,000以上の映画と60,000以上のトレイラーのデータセットを利用して、より良い映画の概要を生成しています。
  • 💬 ミッドジャーニーのCEO、David Holtzは、不審なアカウントからのアクセスが原因で24時間サービス停止が発生したと説明し、Stabilityの従業員からのアクセスを禁止しました。
  • 🔥 StabilityのImad Mustakは、チームがスクラッピングを行っていないと主張し、自社のデータセットと拡張機能に満足しています。

Q & A

  • ハイパーとはどのようなプラットフォームですか?

    -ハイパーは、テキストからビデオを生成する新しいビデオ生成プラットフォームで、元Google DeepMindのメンバーによって作られました。

  • ハイパーの開発に携わった人物は誰ですか?

    -ハイパーは、Yia MeowとZuu Wongという元Google DeepMindのメンバーによって開発されました。

  • ハイパーはどのように資金調達を行ったのですか?

    -ハイパーは、総額1900万ドル以上の資金調達を行いました。

  • ハイパーで提供されている機能は何ですか?

    -ハイパーではテキストからビデオ、画像からビデオ、そしてビデオのリペイント(再描画)を提供しています。

  • ハイパーのインターフェイスはどのようになっていますか?

    -ハイパーのインターフェイスは比較的直截了しており、ライトモードとダークモードの切り替え、プロンプトボックス、ビデオ生成オプションが用意されています。

  • ハイパーで生成されるHDビデオの長さはどのくらいですか?

    -ハイパーで生成されるHDビデオは、現在は約2秒程度の短いものとなっています。

  • ハイパーの標準解像度(SD)モデルとHDモデルの違いは何ですか?

    -SDモデルは、より長い(最大4秒)ビデオを生成できますが、HDモデルのように動的な結果を提供するわけではありません。

  • ハイパーのコミュニティフィードで見られる成果物はどのようなものですか?

    -ハイパーのコミュニティフィードでは、高品質なビデオクリップやアニメーション、スタイリッシュなシーンなどが見られます。

  • ハイパーのビデオを長くするためにどのような方法がありますか?

    -ハイパーのビデオを長くするために、非線形エディター(例:Premiere ProやDaVinci Resolve)を使用して、ビデオのスピードを低下させ、オプティカルフローを有効にすることで、ビデオを延長できます。

  • ムービーLLMとは何ですか?

    -ムービーLLMは、GPT-4と画像からテキストのモデルを利用して、詳細な脚本と対応するビジュアルを生成する研究プロジェクトです。

  • ムービーLLMがどのように機能するのですか?

    -ムービーLLMは、GPT-4で映画のテーマや概要、スタイル、キャラクターを決定し、その後、シーンごとに詳細な脚本を生成します。その後、Stable Diffusionを通して、シーン、キャラクター、場所のスタイルを固定し、一貫性を保ちながらキーフレームを生成します。

  • ムービーLLMが使用したデータセットは何ですか?

    -ムービーLLMは、1000以上の映画と60,000以上のトライラーからデータを作成したMovieNetデータセットを使用しています。

Outlines

00:00

🌟 AIビデオジェネレーターの紹介

この段落では、無料で使える新しいAIビデオジェネレーター「Hyper」について紹介されています。Google DeepMindの元社員であるYia MeowとZuu Wongが立ち上げたプラットフォームで、1,900万ドル以上の資金調達を成功させています。テキストからビデオへの変換、画像からビデオへの変換、ビデオのリペイントなど、様々な機能を提供しています。ユーザーインターフェースはシンプルで、標準的なオプションとHDバージョンのビデオジェネレーションが可能です。しかし、HDバージョンのジェネレーションは2秒程度の短い制限があります。

05:00

🎬 AIビデオの品質と特徴

この段落では、HyperのAIビデオジェネレーションの品質と特徴について詳しく説明されています。特に、車の描写やゴーストトレインのような創造的な要素が好評を得ています。また、AIビデオの短さに関連して、非線形エディターを使用してクリップを延長する方法も提案されています。最後に、完全なAI映画の生成に向けた研究と、その過程でのスタイル固定化プロセスについても触れられています。

10:02

📢 Mid JourneyとStabilityの対立

この段落では、Mid JourneyのCEOであるDavid Holtzが、24時間のサービス停止がStabilityの従業員によるボットのような行為によるものであると説明し、Stabilityの全従業員がMid Journeyサービスの使用を禁止したことを発表しています。Stabilityの代表であるNick S. PierreとImad MustakがTwitterで対立し、データセットの品質やモデルの性能について主張し合っています。また、Mid Journeyがビデオモデルをトレーニング中であり、StabilityがStable Diffusion 3をリリースする予定についても言及されています。

Mindmap

Keywords

💡AI视频生成器

AI视频生成器是一种利用人工智能技术自动创建视频内容的工具。在视频中,提到了一个新的、完全免费的AI视频生成平台,它允许用户通过文本提示生成视频。这个工具的出现预示着AI在创意产业中的应用越来越广泛。

💡文本到视频

文本到视频(Text to Video)是一种AI技术,它可以根据用户提供的文本描述自动生成相应的视频内容。这种技术在视频制作和内容创作领域具有革命性的意义,因为它极大地降低了视频制作的门槛。

💡视频重绘

视频重绘(Video Repainting)是指对现有视频内容进行风格上的改变或创新,使其呈现出全新的视觉效果。这种技术在艺术创作和视频编辑中有着广泛的应用。

💡Mid Journey

Mid Journey是一家提供AI创作工具的公司,它的服务在视频中被提及,因为与稳定性(Stability)公司之间发生了一些争议。这个争议涉及到了数据抓取和使用的问题。

💡稳定性(Stability)

稳定性(Stability)在这里指的是一家与Mid Journey发生争议的公司。这个名词在视频中被用来描述两家公司之间的紧张关系。

💡GPT-4

GPT-4是一种先进的自然语言处理模型,它能够理解和生成人类语言。在视频中,GPT-4被用于电影LLM(Movie LLM)项目,以帮助生成电影剧本和视觉内容。

💡风格固定化(Style Immobilization)

风格固定化是一种艺术和技术手段,用于确保在创作过程中保持一致的视觉风格。在视频中,这一概念被用于AI生成的电影中,以确保电影的视觉风格和角色形象保持一致。

💡特征长度电影(Feature-Length Movies)

特征长度电影指的是时长符合电影标准的作品,通常在90分钟以上。在视频中,讨论了AI技术在生成这种长度电影方面的潜力。

💡数据集(Dataset)

数据集是用于训练和测试机器学习模型的数据集合。在视频中,提到了MovieNet数据集,这是一个包含超过1000部电影和60000个预告片的庞大数据集。

💡合成数据(Synthetic Data)

合成数据是通过算法生成的数据,它可以模拟真实世界的数据特征。在视频中,提到了合成数据在AI模型训练中的应用。

💡非线性编辑器(Nonlinear Editor)

非线性编辑器是一种视频编辑软件,它允许用户以非线性的方式编辑视频,即可以在任何时间点进行编辑,不受时间顺序的限制。

Highlights

New free AI video generator introduced.

Hyper developed by Yia Meow and Zuu Wong, ex-Google Deep Mind.

Hyper raised over $19 million.

Offers text to video generation and image to video.

Hyper interface includes light/dark mode and various video generation options.

Standard definition model allows up to 4 seconds of video.

Community feed showcases diverse user-generated content.

Potential for AI-generated feature-length movies discussed.

Movie LLM research paper uses GPT-4 and text-to-image models.

Style immobilization process for consistent movie generation.

MovieNet dataset used for training in the paper.

Mid Journey CEO addresses 24-hour outage and bot-like behavior.

Ban on Stability AI employees using Mid Journey.

Mid Journey video model in training, reported to be good.

Stable Diffusion 3 release by Stability AI imminent.

Transcripts

play00:00

hey everyone today I have a brand new

play00:02

totally free AI video generator it's

play00:05

pretty impressive and I've got some

play00:07

interesting information on who's behind

play00:09

it and yes I should say totally free for

play00:11

now you guys have been around the block

play00:13

long enough you know how this works plus

play00:15

we're going to take a look at the latest

play00:16

research on generating fulllength AI

play00:19

films yep it's coming all that plus

play00:21

drama at Mid Journey yeah they've got

play00:23

bad blood with stability. a this one is

play00:26

pretty wild it's actually rumored to be

play00:28

the plotline for the second season of

play00:30

Netflix's beef okay lots to cover let's

play00:32

dive in first up is hyper. a a new video

play00:36

generation platform that you can start

play00:38

using right now totally for free this

play00:40

one is brought To Us by Yia meow and zuu

play00:43

Wong two former Google Deep Mind alums

play00:46

who have teamed up to form this platform

play00:49

raising over $19 million so it does look

play00:52

like hyper is set to be quite a

play00:54

contender in the AI video race hyper

play00:58

offers text to video uh very ni nice

play01:00

clean smooth movement here we're going

play01:01

to take a deeper look at that but

play01:03

clearly I mean obviously it does these

play01:04

animated Styles very well it also does

play01:07

image to video in this example uh they

play01:10

actually showcase that you can generate

play01:12

images on the platform and then animate

play01:14

that I have not seen that on the

play01:15

platform yet but we are going to take a

play01:17

look at some image to video examples and

play01:19

also somewhat surprisingly they offer

play01:21

video repainting or you know video in

play01:24

painting essentially they actually had

play01:25

this kind of neat demo of repainting on

play01:28

the site with you know somebody pouring

play01:29

the smoothie mixture into a bowl and

play01:31

then repainting it to be like this

play01:33

watercolored uh you know koi fish thing

play01:36

uh yeah very creative like it a lot

play01:37

getting started the interface is pretty

play01:39

straightforward there is a light dark

play01:40

mode down here by the way uh change it

play01:42

to dark because once again only

play01:44

Psychopaths use light mode your prompt

play01:46

box is down below and you have a number

play01:48

of different options in terms of how you

play01:50

want to generate your video uh either in

play01:53

a full HD version animate your image

play01:55

which is basically image referencing

play01:57

repainting your video and then two more

play01:59

options for creating with text prompts

play02:01

or image referencing but this time in

play02:03

standard definition there is a reason

play02:05

why we'll get to that in one second

play02:07

finally at the end of the options is

play02:08

extend your video uh this obviously has

play02:10

not been enabled yet but will be coming

play02:12

soon diving in in my last video I did

play02:15

kind of a quick sketch music video of

play02:17

some creepy dolls working in a factory

play02:20

for that I ended up utilizing Pika so I

play02:22

decided to take the same prompt and run

play02:24

it through hyper just to see what the

play02:27

results would look like so pretty much

play02:29

the exact same prompt which is stop

play02:31

motion facel dolls working in a factory

play02:33

Gothic dark uh yielded these results

play02:35

which looks really good so to note

play02:38

Generations in this HD format tend to be

play02:41

around 2 seconds I know it's super short

play02:43

I do have a bit of a workaround for that

play02:46

we'll get to that in just one minute

play02:47

once you're in the generation and

play02:49

creation area if you want to swap out

play02:51

modes you can simply come up to this

play02:52

creation mode up here and then you know

play02:55

you have all of your various templates

play02:56

once again so taking that same prompt in

play02:58

the standard definition definition model

play03:00

you'll note that we now have additional

play03:03

templates uh up here that we can add in

play03:05

basically what this is doing is it's

play03:07

just going to add keywords into your

play03:09

prompt so uh say we're going with this

play03:12

in steampunk uh by hitting that we now

play03:14

have weighted steampunk style uh

play03:16

mechanisms fantasy gear decoration Etc

play03:19

we also have additional options here uh

play03:21

for example you can lock the seed number

play03:23

uh there is controls for motion levels

play03:25

let's just crank that all the way up and

play03:27

then the duration on the standard

play03:30

definition version you can take your

play03:31

Generations up to 4 seconds running that

play03:34

but removing the steampunk prompts uh

play03:36

yields this which yeah that looks pretty

play03:38

good I would put it pretty much on par

play03:40

with the Pika outputs that I got in the

play03:43

last video so these are definitely two

play03:45

different models image to video also

play03:46

looks really good I took this Dune

play03:48

inspired like heavily Dune inspired

play03:51

image that I generated up in mid Journey

play03:53

mostly because I have Dune fever I have

play03:54

not seen it yet hopefully next week

play03:56

anyhow running that into hyper and then

play03:58

adding the prompt man looks to camera

play04:01

intense gaze and running that in the HD

play04:03

mode yielded this which yes while short

play04:06

does look very good now taking that same

play04:09

image and running it through the SD

play04:11

model with the exact same prompt yielded

play04:14

this the results from the SD version are

play04:16

just a little bit more on the I know

play04:18

kind of static and Bland side uh

play04:20

additionally Timothy shalom's stunt

play04:22

double here has kind of an animated look

play04:24

to it his eyes are kind of doing the

play04:26

Pennywise thing as well uh and just you

play04:28

know overall the HD version is

play04:30

definitely far superior to this that

play04:33

said I do think that you can still get

play04:34

some really cool results out of the SD

play04:36

model uh for example this is a prompt of

play04:39

just uh a wizard standing in an

play04:40

enchanted forest and I use the gibli

play04:43

preset and yeah this looks really really

play04:45

nice hopping over to the Community Feed

play04:47

to see the breath of what hyper is

play04:49

capable of uh I ran across this output

play04:51

which looks really really good um yeah I

play04:54

mean I would hesitantly call this

play04:56

somewhere in the neighborhood of Sora

play04:58

quality although you you know granted

play05:00

much shorter something that I noticed is

play05:02

that the cars are all maintaining

play05:03

consistency like they're they're not

play05:05

morphing all over the place and there

play05:06

aren't a few of them like driving

play05:07

backwards although I do have a bone to

play05:09

pick that there are what like five

play05:11

Outward Bound lanes and one inward bound

play05:13

Lane there's one thing Sim City taught

play05:15

me is that this is going to end in

play05:16

disaster sticking with the imaginative

play05:18

side this ghost train in the sky looks

play05:20

really really good this interior zoom

play05:23

out also looks really good although

play05:24

obviously you know Limited in time it's

play05:26

it's also not like morphing out too much

play05:29

everything is staying consistent and

play05:31

stable this extreme closeup also caught

play05:33

my eye yeah I mean pun not intended uh

play05:35

but yeah this is actually text video

play05:37

it's not image to video so that's what

play05:39

I'm saying about it approaching sore

play05:42

level quality although just you know not

play05:44

as long few more quick ones yeah walk

play05:46

cycle here looks really really good uh

play05:48

all six legs look like they are actually

play05:50

present and not just like sort of

play05:51

floating along a surface another walk

play05:53

cycle or I should say step cycle uh in

play05:56

sort of an animated style it's really

play05:58

solid and finally cribbing one of the

play05:59

famous Sora text prompts fly over of a

play06:02

California Gold Rush town uh we end up

play06:05

with this which yeah I mean this looks

play06:08

really really good now I should say it's

play06:09

still AI video and there is still

play06:12

weirdness abound uh taking our old

play06:14

friend the man in the blue business suit

play06:16

walking down a busy City street gets us

play06:18

this uh where our guy is like

play06:21

moonwalking down the center lane of a

play06:24

city street definitely not safe uh

play06:26

additionally we have this car over here

play06:29

that's actually driving backwards but

play06:30

you know maybe they just made a wrong

play06:32

turn down a one-way street but again at

play06:34

the cost of free I mean roll to your

play06:36

heart's content until you find something

play06:37

that you like overall I think the

play06:39

results out of hyper are super

play06:41

impressive yeah there is that 2cond

play06:43

limitation on the HD versions uh but

play06:46

that is a bridge that will be crossed at

play06:48

one point or another in the meantime uh

play06:50

if you want a quick hack in terms of

play06:52

extending your shots you can always

play06:54

bring your clip into a nonlinear editor

play06:57

like Premiere that we have here or Da

play07:00

Vinci resolve and extend it out in

play07:02

Premiere what you'll want to do is uh

play07:04

just right click on it come to speed

play07:06

duration uh take the speed down to 50%

play07:09

and then just make sure that Optical

play07:11

flow is turned on here Da Vinci also has

play07:14

something pretty similar uh but yeah as

play07:15

you can see the results are pretty good

play07:17

A little slowed down but I mean not bad

play07:19

ultimately to get the best results you

play07:21

should probably use something like topaz

play07:23

video Ai and listen I know it is super

play07:25

expensive but it is also kind of the

play07:28

best in class for for this particular

play07:30

job and that pretty much is the

play07:32

smoothest and best result that you can

play07:34

possibly get and again I know the 2C

play07:36

thing is a bit of a bummer but again

play07:38

that is just temporary and even now I

play07:40

think you can craft together a pretty

play07:42

solid narrative by using a combination

play07:45

of 2cond HD shots extending them out a

play07:48

bit and 4 second SD shots uh to vary out

play07:53

your shot lengths anyhow moving on from

play07:55

like two second shots to feature length

play07:58

movies uh we have movie llm enhancing

play08:01

long video understanding with AI

play08:03

generated movies now this one is in the

play08:05

paper research Arena but it does set the

play08:07

stage for what will be eventually coming

play08:10

which is you know fully generated AI

play08:12

movies from a prompt we actually have

play08:14

seen something kind of similar to this

play08:16

in a previous video namely LTX studio uh

play08:19

if you missed that video it is linked

play08:21

Down Below movie llm leverages The Power

play08:24

of GPT 4 and text to image models to

play08:26

generate detailed scripts and

play08:28

corresponding visuals the work path of

play08:31

movie llm is the movie plot generation a

play08:34

style immobilization process and video

play08:37

instruction by data generation movie llm

play08:39

leans on GPT 4 for sort of the initial

play08:42

breakdown of the film um with GPT coming

play08:45

up with the theme the overview movie

play08:47

style frame level description and the

play08:50

characters of the film from there GPT 4

play08:53

takes that output and generates up Epoch

play08:56

chapters essentially scenes for the most

play08:58

part uh at which point it takes it

play08:59

through another process where it will

play09:01

generate out you know chapters for each

play09:04

scene detailing both the characters the

play09:07

actions that are happening and the

play09:08

location uh from there it actually takes

play09:11

that output and then begins generating

play09:13

dialogue for each of those scenes the

play09:16

real key to all of this is in the style

play09:18

immobilization process in which uh

play09:21

essentially keywords are then extracted

play09:23

from each of those chapter epochs the

play09:25

characters and the plot summary uh run

play09:27

through stable diffusion to generate up

play09:30

the various you know scenes characters

play09:32

and locations uh but then taken through

play09:35

the uh textural inversion process and

play09:38

the immobilized style embedding in order

play09:41

to essentially lock the look of the

play09:43

entire film into one consistent or what

play09:47

multiple consistent characters and

play09:48

multiple consistent locations from there

play09:51

it's taken back out through generation

play09:53

guiding and through another stable

play09:54

diffusion process to generate out key

play09:57

frames for each one of the scenes now

play10:00

the paper did not provide any video

play10:02

examples but we can see key frame

play10:04

comparisons uh namely uh the movie llm

play10:08

output is down here where you know it

play10:10

does it looks like it's very consistent

play10:12

in terms of style and in character

play10:14

whereas in the other two examples uh

play10:16

this custom diffusion one in particular

play10:17

is a little bit all over the place in

play10:19

terms of its overall style and look just

play10:22

zooming in here for a minute because the

play10:23

custom diffusion model actually did make

play10:25

me laugh a little bit like in the second

play10:27

example where he's supposed to be a

play10:29

blacksmith he just kind of looks like a

play10:31

drummer for like an 80s rock band

play10:33

interestingly in the paper they do note

play10:35

that there isn't a lot of resources out

play10:37

there in terms of training models for

play10:39

extremely long like featurelength movies

play10:42

uh but they ended up using the data from

play10:45

the movie net data set which apparently

play10:48

is a massive data set of over 1,000

play10:51

movies and 60,000 trailers now to note

play10:54

the movies and trailers aren't actually

play10:56

within this data set it's just the data

play10:59

from those films and trailers by

play11:01

utilizing the movet data set movie llm

play11:04

is able to basically come up with better

play11:06

synopses for its own films weirdly

play11:10

flexed with an example by using the film

play11:13

splice I've seen splice and that's a

play11:15

pretty odd choice to go with for an

play11:17

example if you haven't seen it I pretty

play11:19

much say you know you can take a pass on

play11:21

that one uh what you should check out is

play11:22

natali's other movie Cube that movie is

play11:25

really good ring out with some pretty

play11:27

odd mid Journey news uh yesterday's

play11:29

office hours started off on a pretty

play11:31

weird note mid Journey CEO David Holtz

play11:34

explained that the 24-hour outage they

play11:36

had on Saturday was caused by botlike

play11:40

behavior from Paid accounts he then said

play11:43

that this originated from stability. a

play11:45

employees basically scraping for images

play11:49

and text prompts at which point he

play11:51

announced that he was effectively

play11:53

banning all stability. employees from

play11:56

using the mid-journey service Nick s

play11:58

Pierre took to Twitter as this was

play12:00

happening and offered stability. imod

play12:03

mustak the opportunity to comment imad's

play12:06

response was uh what mod shortly

play12:08

followed up with and I'm just quoting

play12:10

directly here it's Twitter there are

play12:11

grammatical and spelling errors glore

play12:14

but you know it again it's Twitter uh

play12:16

very confusing how two accounts would do

play12:18

this team also hasn't been scraping as

play12:21

we have been using synthetic and other

play12:23

data given sd3 outperforms all other

play12:26

models anyway I am a big mid journey and

play12:29

David fan which is why I back them at

play12:31

the start with the grant to pay for the

play12:34

beta on E we go IM mod followed up from

play12:37

there with if anyone did do this on team

play12:41

have asked we'll dig also happy if mid

play12:43

Journey reaches out direct it's not

play12:46

great but obviously not a DDOS attack

play12:49

but unintentional certainly not

play12:52

instructed by us/ stability though

play12:55

really happy with our data set and the

play12:57

augmentation we have on that a little

play13:00

while later David did respond to Ahad

play13:02

saying sent you some information to help

play13:04

with your internal investigation

play13:06

definitely interesting I do hope the

play13:08

boys can resolve this issue I mean we

play13:10

have enough going on with Elon and Sam

play13:12

we do not need any more drama meanwhile

play13:15

mid journey is still training their

play13:16

video model it is reported to be quite

play13:19

good and stability will be releasing

play13:21

stable diffusion 3 into the wild uh

play13:23

pretty much at any moment so that wraps

play13:26

up another crazy week in Creative AI as

play13:28

as always I thank you for watching my

play13:30

name is

play13:36

Tim

Rate This

5.0 / 5 (0 votes)

Related Tags
AI動画ジェネレーターハイパーGoogle DeepMindテキストから動画画像アニメーションビデオリペインティングAI技術動画制作クリエイティブAIハイドパフォーマンスビデオエディットAI映画制作
Do you need a summary in English?