New AI Video That Does Everything!

Theoretically Media
28 Feb 202410:36

Summary

TLDRThe video script unveils the latest advancements in AI-powered creative tools and platforms. It introduces LTX Studio, a holistic platform that solves challenges faced in creating AI-generated films and videos, offering features like text-to-video generation, shot editing, dialogue and sound integration, and auto-editing. Updates to platforms like Pika and Runway, enabling lip-syncing and motion brush enhancements, are also discussed. Notably, Google's Genie, a text-to-2D game generator trained on game footage, showcases the potential for unsupervised learning in AI-powered content creation.

Takeaways

  • 👍 LTX Studio is a new AI video platform announced that offers features like text-to-video, shot editing, camera controls, music/dialogue/sound effect integration, video inpainting, storyboarding, casting, and auto-editing.
  • 🎥 Pika AI has added a lip-sync feature that can generate 3-4 second audio clips or sync to provided MP3 files for characters in the video.
  • ✏️ Runway AI has introduced a new quality-of-life feature in Motion Brush that allows for auto-selecting and controlling specific areas of an image.
  • 🎮 Google has released Genie, a text-to-2D game generator that can create simple platformer games from text or images, trained on 200,000 hours of video game footage.
  • 🔮 Runway's CEO hinted that their AI outputs may soon be better than the highly impressive Sora AI.
  • 🌟 The video emphasizes the importance of continuously creating and using available tools instead of waiting for the perfect one.
  • 🤖 The video showcases various examples and use cases of the new features from LTX Studio, Pika AI, and Runway AI.
  • 🔄 LTX Studio allows for iterative polishing and rearranging of shots while creating AI films.
  • 🎬 LTX Studio can generate entire storylines and scripts, though users can still write their own.
  • ⏱️ The video mentions that LTX Studio is set to release by the end of March, with an opportunity for early access through a provided link.

Q & A

  • What is LTX Studio and who developed it?

    -LTX Studio is a new AI video platform developed by Light Tricks. It's designed as a holistic platform to solve various challenges in creating longer form AI films or videos, integrating external requirements under one roof.

  • What makes LTX Studio unique compared to other platforms?

    -LTX Studio distinguishes itself by offering comprehensive tools for AI film and video creation, including camera controls, editable shots, and on-platform music, dialogue, and sound effects, making it a one-stop solution.

  • What are some of the key features of LTX Studio?

    -Key features include text-to-video conversion, camera control, storyboard view, consistent character casting, lighting adjustments, title card generation, and auto editing for dialogue, sound effects, and music.

  • How can users access LTX Studio early?

    -Users can gain early access to LTX Studio by using a special link provided by the video's presenter, who received early access as a partner through Light Tricks.

  • What does the updated Pika platform offer?

    -The updated Pika platform introduced a new feature that allows adding lip sync to video outputs, with audio generations limited to 3 to 4 seconds but with a workaround for extending this.

  • How does the new motion brush feature from Runway improve the user experience?

    -Runway's new motion brush feature automates the selection of certain image areas for control, significantly speeding up the editing process and making it easier to apply effects and adjustments.

  • What is Genie by Google, and what can it do?

    -Genie is a text-to-2D game generator released by Google, trained on 200,000 hours of video game footage. It can create simple games from text prompts or images, focusing on old school NES platformer style games.

  • How does the LTX Studio manage camera movements in videos?

    -LTX Studio allows users to control camera movements including horizontal, vertical, pan, roll, and zoom, providing greater flexibility and creativity in video production.

  • Can LTX Studio generate an entire script for a film?

    -Yes, LTX Studio can write and generate an entire storyline and script for you, but it also allows users the flexibility to write their own storylines if they prefer.

  • What was the response of Runway's CEO to the comparison with Sora's output quality?

    -The CEO of Runway, Cristobal Valenzuela, responded with optimism, stating that Runway's outputs would look 'better' than Sora's, suggesting confidence in future improvements to Runway's technology.

Outlines

00:00

🎬 LTX Studio: A Groundbreaking AI Video Platform

The input introduces LTX Studio, a new AI video platform that provides a comprehensive solution for creating long-form AI films or videos. It is a holistic platform that brings various external requirements under one roof. The video showcases LTX Studio's capabilities, including text-to-video generation, shot-to-shot panels, editable prompts, timelines for music, dialogue, and sound effects, video inpainting, storyboarding, casting, title card generation, camera movement controls, auto-editing, and script generation. It highlights the platform's ability to create full stories efficiently. The input also mentions that while impressive, LTX Studio serves a different purpose than Sora, which generates individual shots, and emphasizes the importance of continually creating and utilizing available tools.

05:02

👄 Pika's Lip Sync and Runway's Motion Brush Update

Pika has updated their platform with a new feature that allows adding lip sync to output videos. The feature enables users to select from various AI voices or upload their own audio files for lip-syncing. The summary showcases examples of Pika's lip sync feature with different characters and discusses potential workarounds for extending audio durations. Additionally, Runway has introduced a quality-of-life update to their Motion Brush module, allowing for auto-selection of image areas for precise control. The input illustrates the update's capability by recreating the iconic dolly zoom shot from the film Jaws and provides a practical use case example.

10:05

🎮 Google's Genie: Text-to-2D Game Generator

Google has released Genie, a text-to-2D game generator. Genie can generate simple, old-school NES-style platformer games based on text prompts or images. While the games are limited in scope and run at a low frame rate, the underlying technology is remarkable. Genie was trained unsupervised on 200,000 hours of publicly available video game footage, enabling it to create entire game worlds. Although the input suggests that Genie itself may not be accessible to the public, the underlying research and technology are expected to be incorporated into future products and advancements.

Mindmap

Keywords

💡LTX Studio

LTX Studio is presented as a holistic AI video platform by Lightricks, designed to address the challenges in creating longer-form AI-generated films or videos by consolidating external requirements under one roof. It signifies a breakthrough in video production, offering tools for text-to-video conversion, camera controls, and editable scenes, demonstrating how it streamlines the creative process by integrating various aspects of filmmaking into a single platform.

💡Text to video

Text to video refers to the capability of transforming written prompts into video content, a feature highlighted early in the script with examples of high-quality output. This technology showcases the AI's ability to understand and visually represent narrative elements, allowing creators to generate video scenes directly from textual descriptions.

💡Camera controls

Camera controls within LTX Studio allow users to direct the virtual camera's movement, such as tracking towards a subject or changing angles, as if handling a real camera. This feature enriches the storytelling by enabling dynamic visual storytelling elements, like the mentioned 'track in toward villain,' thereby adding depth and cinematic quality to AI-generated videos.

💡Editable shots

Editable shots refer to the capacity to modify generated video scenes in terms of content and narrative focus, as demonstrated when changing a 'futuristic space drama' to a 'New York City courtroom drama.' This flexibility is crucial for filmmakers and content creators, allowing them to refine and pivot their storytelling without starting from scratch.

💡Storyboarding

Storyboarding is described as a feature allowing creators to view and organize their entire film from a bird's-eye perspective. It enables the swapping and insertion of shots, highlighting the platform's utility in planning and iterating film projects, thus facilitating a more efficient and creative workflow for users.

💡Casting and in-painting

Casting and in-painting refer to the process of generating and modifying characters within scenes, ensuring consistency across different shots. This feature simplifies the creation of coherent narrative visual content, demonstrating the AI's capability to adapt characters and settings according to the user's needs, like changing a green car to a red one or modifying character appearances.

💡Auto editing

Auto editing is a feature that automatically syncs music, dialogue, and sound effects with the video content, enhancing the production value by ensuring that audio elements complement the visual narrative effectively. This innovation highlights the move towards a more integrated and user-friendly approach to video editing within the AI platform.

💡Script generation

Script generation is a capability of LTX Studio to create narrative scripts for videos, offering users the option to either use AI-generated storylines or input their own. This flexibility emphasizes the AI's role in supporting creative processes, catering to both users who seek full automation and those who wish to maintain control over their story's narrative.

💡Pika platform

Pika platform is mentioned as having introduced a new feature for adding lip-sync to videos, showcasing advancements in AI-generated audio and its synchronization with visual content. This development demonstrates the ongoing enhancements in making AI tools more sophisticated and useful for creators, especially in creating engaging and lifelike characters.

💡Genie

Genie, introduced by Google, is a text-to-2D game generator, exemplifying the expansion of AI capabilities into the realm of video game development. Although simplistic and limited in scope, Genie's ability to generate games from text or images showcases the innovative use of AI in creating interactive and engaging content, based on unsupervised training from extensive gaming footage.

Highlights

LTX Studio by Lightricks is a new holistic AI video platform that solves challenges in making longer-form AI films by bringing all external requirements under one roof.

LTX Studio offers text-to-video generation, shot-to-shot panel editing, music, dialogue, and sound effects integration, video inpainting, storyboarding, casting, title card generation, camera movement controls, and auto-editing.

LTX Studio can generate an entire storyline and script, but users can also write their own.

LTX Studio is different from Sora, which generates individual shots, while LTX Studio is a platform for creating full stories.

Pika has updated their platform with a new feature that allows lip-syncing to generated audio or user-provided MP3 files.

Runway has introduced a new quality-of-life feature in Motion Brush that allows for auto-selection of certain areas of an image for better control.

The new Runway feature enables creating cinematic effects like the dolly zoom shot from Jaws.

Runway's CEO hinted that their outputs may soon look better than Sora's.

Google has released Genie, a text-to-2D game generator that can create simple platformer games from text or images.

Genie was trained unsupervised on 200,000 hours of video game footage and can generate games, showcasing the potential for world-creation through AI.

The speaker encourages always creating and using available tools, rather than waiting for specific products.

LTX Studio is set to release by the end of March 2024, and users can sign up for early access using the provided link.

The speaker demonstrated the new Pika lip-syncing feature using different voices and an MP3 file.

The speaker showcased the new Runway Motion Brush feature by recreating the iconic dolly zoom shot from Jaws.

The speaker highlighted the unsupervised training approach used for Genie and its potential implications for AI world-creation.

Transcripts

play00:00

so it looks like we have a new AI video

play00:02

platform this one was just announced

play00:04

today and it does a lot honestly like

play00:07

all the things uh trust me you're going

play00:09

to want to check this one out because we

play00:10

haven't seen anything like it yet we've

play00:12

also got some big updates to paa and Gen

play00:15

2 plus Google has a text to 2D game

play00:18

generator which you know on its own is I

play00:20

guess kind of cool but it's really when

play00:22

you dig in that you see the

play00:23

ramifications of what this really does

play00:26

okay lots to cover let's dive in kicking

play00:28

off we have LT X Studio by light tricks

play00:31

this is something that I've been

play00:32

predicting for a while a holistic

play00:34

platform that solves a lot of the

play00:36

challenges that we face when we're

play00:37

making longer form AI film or video uh

play00:40

kind of bringing all of the external

play00:41

requirements in Under One Roof as a

play00:44

quick FYI I did get early access to LTX

play00:47

Studio as a partner through lightricks

play00:50

uh stay tuned because I've got a way for

play00:52

you guys to get Early Access as well see

play00:54

I'm always looking out for you uh but I

play00:56

think as we go through this video you'll

play00:58

see exactly why I'm excited about it and

play01:00

listen I know Sora we'll talk about that

play01:03

in a minute but for now let's step

play01:04

through this video the video kicks off

play01:06

with some text to video examples

play01:08

everything looks really good here uh I

play01:10

did notice that in this shot the prompt

play01:13

calls out track in toward villain uh yes

play01:16

there is camera controls in LTX Studio

play01:19

we'll talk about that more in just a

play01:21

minute from there it moves into this

play01:23

shot to shot panel where you know we get

play01:25

three shots when we prompt futuristic

play01:27

space drama but here is where things get

play01:29

pretty interesting continues generating

play01:31

and we end up with 12 shots now what's

play01:33

interesting is that these are all

play01:34

editable as well uh as seen in the next

play01:37

example where uh they change the prompt

play01:39

out to New York City courtroom drama and

play01:42

yeah everything repopulates now what's

play01:44

interesting is that if you actually

play01:46

scrub through this you'll see that our

play01:48

vaguely Cyber punan City here actually

play01:50

becomes an east coast city and you know

play01:52

our first shot uh kind of that back

play01:55

angle of the character turns into uh is

play01:58

that uh I guess like John Doe from Seven

play02:01

I don't know I kind of want to see this

play02:02

movie now now here is the really cool

play02:04

part because as you can see here we now

play02:06

have uh timelines for music dialogue and

play02:09

sound effects and these all happen on

play02:11

platform so there's no need to generate

play02:13

externally and bring them in this is all

play02:15

One-Stop shopping it's a quick FYI uh

play02:18

this is actually not the UI you know

play02:20

this is a Sizzle reel that tends to

play02:22

happen I do have a look at the actual UI

play02:24

coming up in just a minute we next get a

play02:26

shot of video imp painting with this

play02:28

shot of like bootleg Josh Haron and the

play02:30

green car in the background where

play02:32

obviously via text prompt we have now

play02:33

changed the green car to a red car

play02:36

storyboarding is another feature that I

play02:37

think is super cool you know you can

play02:39

basically get a bird's eyye view of your

play02:41

entire film and you can actually even

play02:43

swap shots around and if you need to

play02:45

insert shots uh you can do so as well

play02:48

which is obviously hugely valuable when

play02:50

you're iterating and polishing your film

play02:52

if you're wondering about timing of your

play02:53

shots yeah you can do that as well we'll

play02:55

take a look at that in one second here

play02:57

we've got a bit on casting it's you know

play02:58

basically video in painting but the

play03:01

important thing to stress here is that

play03:03

you're getting consistent characters we

play03:05

can swap around lighting we can generate

play03:07

some solid title cards for our film

play03:09

everything here looks spelled correctly

play03:11

so really this just comes down to user

play03:13

input meaning all of my titles are going

play03:15

to be misspelled we have camera movement

play03:17

controls as well as you can see we have

play03:19

controls for horizontal vertical our pan

play03:21

our roll and our zoom and LTX studio

play03:23

also has Auto editing this is likely to

play03:26

time all of the dialogue uh sound

play03:28

effects and music

play03:30

just as a quick FYI the UI that we saw

play03:32

in the sizzle reel is not what LTX

play03:35

Studio actually looks like uh it's

play03:37

fairly common for a launch video to

play03:39

prioritize obviously these are the

play03:41

features that are available not

play03:43

necessarily like this is what the

play03:44

platform looks like but they did send

play03:47

this over to me so that I could show you

play03:48

what the platform actually looks like

play03:50

right now to be honest I actually prefer

play03:52

this over sort of a flashy iPad type

play03:55

look as the video closes out it does

play03:57

mention script which wasn't highlighted

play03:59

in the the sizzle reel but yes LTX

play04:01

Studio can you know write and generate

play04:03

an entire storyline in script for you

play04:05

but importantly it doesn't have to you

play04:08

can still write your own storyline

play04:10

that's something that's kind of

play04:11

important to me okay now for the

play04:12

elephant in the room Sora uh yes Sora is

play04:16

amazing but a we don't have it yet and B

play04:20

this is doing something different Sora

play04:22

generates individual shots whereas this

play04:24

is a platform where you can create full

play04:26

stories in my last video I talked to

play04:28

Nico from C who had some really great

play04:30

advice for anyone that wanted to make an

play04:32

AI film or really do anything creative

play04:35

always be creating it's not about one

play04:36

idea it's not about one product AI or

play04:38

whatever it doesn't matter you need to

play04:40

always be making stuff it art is not

play04:42

about I made one movie I'm good it's

play04:44

about I like to make movies so I make

play04:45

movies every day so don't wait for the

play04:47

tools use the ones that are available to

play04:49

you and make awesome stuff to which you

play04:52

are probably wondering when does LTX

play04:54

Studio actually released well I've got

play04:55

good news not too long away end of March

play04:58

but I've also got some good news you can

play04:59

sign up for Early Access if you use the

play05:02

link down below moving on Pika have

play05:05

updated their platform with a new

play05:06

feature that allows you to add lip sync

play05:09

to your output videos audio Generations

play05:12

are limited to about 3 to 4 seconds

play05:14

although there is kind of a work around

play05:16

that I'll show you in just one second so

play05:18

uh let's give lip syncing a shot we're

play05:19

going to take our old friend Daniela van

play05:21

Denon dressed as a pirate we haven't

play05:22

seen her in a while and drop her into

play05:25

the prompt box uh from here you can just

play05:27

hit this lip sync button and that will

play05:30

provide you with a number of different

play05:32

drop-down voices that you can try out so

play05:35

uh we're going to try out Demi we type

play05:39

in some text and before you know it we

play05:40

got Daniela speaking please stop using

play05:42

me in your videos it's getting silly in

play05:44

general I kind of find the AI voices to

play05:46

be well very AI voice sounding uh but

play05:48

luckily you can actually add in your own

play05:51

MP3 files as well and it will lips sync

play05:54

to that so let's give that a shot so

play05:56

taking some audio of the real Daniela

play05:58

van Denon and popping it in uh take a

play06:00

listen to that real

play06:03

quick she's Dutch I have no idea what

play06:06

she was saying there I will say that it

play06:07

can also get a bit finicky depending on

play06:09

the face detection for example uh with

play06:11

this astronaut here um well I'll I'll

play06:13

just play it I wasn't originally going

play06:15

to get a brain transplant but then I

play06:16

changed my mind I cracked me up so while

play06:19

I did get a whopping six seconds out of

play06:20

it uh obviously the face was kind of

play06:22

paralyzed by the way two points to the

play06:24

Spaceman for the dad joke but I did get

play06:25

a pretty good Indiana Jones X never ever

play06:29

the spot so I will say it's all

play06:31

definitely a work in progress but it is

play06:33

progress in the right direction uh

play06:34

here's a couple of quick examples Dave

play06:35

alova posted this one up show me the

play06:38

money and P themselves posted this up uh

play06:40

which kind of looks like a young John

play06:41

Truro in a liveaction Ratatouille

play06:45

remake that is a deep cut Dexter's

play06:47

Laboratory reference one idea in terms

play06:50

of extending out your lip sync audio is

play06:53

that you can potentially run one

play06:54

generation and then just slide this

play06:56

whole thing over uh and then run a

play06:58

second generation and then kind of

play06:59

splice them together in some editing

play07:02

software afterwards not going to lie

play07:04

might be a bit of a pain but it is a

play07:05

potential workaround at least for now

play07:07

all in all keep up the great work Pika

play07:09

moving on not to be outdone Runway have

play07:12

introduced a new quality of life feature

play07:15

to their motion brush uh mostly what

play07:17

this allows for is to Auto Select

play07:19

certain areas of your image uh so that

play07:22

you can have control over them uh it is

play07:24

definitely a quality of life thing it

play07:26

just makes things a lot faster you can

play07:28

always erase sections as well uh this is

play07:31

all demonstrated by Rich Klein AI now as

play07:34

soon as I saw that I figured that this

play07:35

would make a pretty good excuse to try

play07:37

out the dolly Zoom shot from Jaws the Z

play07:40

shot if you will it was also used in

play07:42

Vertigo the way this effect was achieved

play07:44

in camera is that the camera itself

play07:46

would dolly in on Roy shider while at

play07:49

the same time they would be zooming out

play07:51

on the lens so it creates that sort of

play07:53

weird parallaxing vertigo is effect so

play07:56

taking a screenshot and bringing it into

play07:58

the motion brush mod module as you can

play07:59

see we can just sort of you know grab

play08:01

selections and just click and uh those

play08:04

sections are now part of our brush one

play08:07

from there I just took a brush one and

play08:09

cranked the proximity up meanwhile I

play08:11

just selected the entire background and

play08:13

cranked the proximity in the opposite

play08:15

direction running that got us this which

play08:18

actually is not that bad I mean it yes

play08:20

it it is weird and warpy and all of that

play08:23

but I mean it's also one of the most

play08:24

iconic film shots of all time so you are

play08:27

judging at a very high bar still that is

play08:29

a cinematic technique that you know does

play08:31

work in Runway so you can feel free to

play08:33

bring that into your own AI films in

play08:36

terms of a more practical use case uh

play08:38

for this new tool Nicholas Newbert

play08:40

posted up a speedrun of him working with

play08:42

it you can see him here quickly

play08:44

selecting various tools using uh

play08:46

proximity Ambience uh the horizontal and

play08:48

vertical as well as camera movement on

play08:51

it uh and then when he generates yeah

play08:52

this is looks very very good rounding

play08:55

out and Runway USS I don't know if you

play08:57

guys caught this but uh someone asked

play08:58

crystal ball valenzula the CEO of Runway

play09:01

if uh runway's outputs would would look

play09:04

as good as Sor is anytime soon and his

play09:07

response was better so whether that was

play09:09

bravado or if we actually will be seeing

play09:12

like Sora level outputs coming out of

play09:14

runway in the next couple of months I

play09:15

don't know it's all just very exciting a

play09:17

lot of stuff is moving very quickly and

play09:19

yeah it this couldn't be a better time

play09:21

to be sharpening your skills with AI

play09:24

video rounding out Google have released

play09:26

well Google released Genie which is a

play09:29

text to 2D game generator so yeah

play09:32

game-wise it is very much in the old

play09:34

school NES platformer style um you are

play09:37

not going to be text prompting for

play09:39

halflife 3 here but here is what's kind

play09:41

of cool about it is that Genie was

play09:43

actually trained unsupervised on 200,000

play09:46

hours of publicly available video game

play09:49

footage and it can generate games either

play09:51

off of text or via an image the games

play09:54

are obviously very simple and apparently

play09:56

only running at one frame a second so

play09:57

it's not like it would be a very

play09:59

enjoyable experience to play anyways but

play10:01

the idea behind it is pretty remarkable

play10:04

in that what Jeanie's doing is actually

play10:07

world creating again Google being Google

play10:10

I don't think that we'll ever get a

play10:11

chance to play with Genie but I think

play10:12

the important part here is the

play10:14

underlying research which I think will

play10:16

appear in some way in the future oh and

play10:19

if you did happen to miss the last video

play10:20

I did where I gave a talk about AI to

play10:23

Hollywood I do invite you to check that

play10:25

out that video was a lot of fun it's

play10:27

coming up next I thank you for watching

play10:29

my name is

play10:34

Tim

Rate This

5.0 / 5 (0 votes)

Related Tags
AI VideoAI CreativityTech UpdatesVideo EditingFilmmakingGame DevelopmentFuturistic MediaInformativeInnovativeExciting