Massive AI News from Stability, Adobe, Ideogram, and More!

Theoretically Media
1 Mar 202410:18

Summary

TLDRThis week brought major AI developments, including Morph Studio's new AI filmmaking platform for easily generating and editing video; EmoTalker's impressive lip sync tool for animating still images; Adobe's Project Music AI for music generation and editing; an update to the free AI image generator Idiogram; insights into how YouTube's Sora video generator works via reverse engineering; and the first fully AI-generated remake of Terminator 2, created by 50 artists, which will premiere theatrically in LA before being live streamed.

Takeaways

  • 😲 Morph Studio partners with Stability AI to create a new AI filmmaking platform
  • 🎥 EmoTalker adds impressive lip sync to still images along with heavy eyeliner!
  • 🎵 Adobe releases Project Music AI for music generation and extension
  • 🔍 Researchers start reverse engineering Sora for similar video outputs
  • 🎞️ The first fully AI-generated remake of Terminator 2 gets a theatrical release
  • 🖼️ Idiogram 1.0 upgrades aesthetics and adds a handy prompt helper
  • 👄 Pik also has a new lip sync feature for images and video
  • 🌌 Sora looks amazing but likely won't be accessible for a while
  • 💸 EmoTalker leverages Stable Diffusion, trained on lots of data
  • ✏️ Morph's platform allows sharing templates and workflows

Q & A

  • What new video generation platform was announced?

    -Morph Studio partnered with Stability AI to create a new AI filmmaking platform called Morph Cloud.

  • What does the new platform Morph Cloud aim to do?

    -Morph Cloud aims to streamline the filmmaking process by allowing users to regenerate shots, share workflow templates, and have more control over the editing process using AI.

  • What is the name of the new lip sync tool featured?

    -The new lip sync tool is called Emo Talker, created by Alibaba.

  • What makes the Emo Talker lip sync tool special?

    -Emo Talker can add lifelike lip sync to any still image and also expresses emotion through facial expressions and eye movement.

  • What new AI music project did Adobe release?

    -Adobe released an AI music project called Project Music, developed with UC Berkeley and Carnegie Mellon.

  • What does Project Music allow users to do?

    -Project Music allows users to extend and vary musical compositions using AI.

  • What is the name of the free AI image generator covered?

    -The free AI image generator is called Idiogram.

  • How does Idiogram compare to other generators?

    -Idiogram specializes in coherent text generation within images and released a new 1.0 model with upgraded aesthetics.

  • What is the historical event related to AI mentioned?

    -The first fully AI-generated remake of a full length film (Terminator 2) will be screened in theaters.

  • Who created the AI remake of Terminator 2?

    -The remake was a collaborative project among 50 different artists using AI to generate the film.

Outlines

00:00

😲 New AI Film Production Platform

Paragraph 1 introduces a new AI film production platform called Morph Studio that partners with Stability AI. It has a node-based interface that connects different AI-generated video clips, allows style transfer, and gives control over the blend. The platform aims to create a community for sharing templates.

05:00

💋 Emo Talker Adds Lip Sync and Emotion

Paragraph 2 covers Emo Talker, a new lip sync tool from Alibaba that adds lifelike facial expressions and lip sync to still images. It was trained on extensive video and image data and leverages Stable Diffusion. The results are impressive but it currently only works with still images.

10:01

🎵 Adobe Project Music AI for Music Generation

Paragraph 3 discusses Adobe's new Project Music, an AI system for music generation and manipulation developed with UC Berkeley and Carnegie Mellon. It has capabilities like extending and changing music but there is no timeline yet for if/when it would be incorporated into Adobe products.

Mindmap

Keywords

💡Creative AI Tools

This refers to artificial intelligence systems and platforms that are designed to enhance and augment human creativity, especially in areas like art, music, writing etc. The video discusses updates and new releases from companies like Anthropic, Stability AI, Pika, Runway ML etc that are building these creative AIs.

💡Video generation

The ability to algorithmically generate or synthesize video content is a growing capability of AI systems. The video talks about new platforms like Morph Studio that allow users to control and tweak AI-generated video.

💡Lip sync

This refers to synchronizing facial animations, especially lip movements, with audio tracks like speech or music. The video demonstrates EmoTalker, an AI tool that can add realistic lip sync to still images.

💡Adobe

A leading technology company known for creative software tools like Photoshop. The video discusses Adobe's new AI music project called Project Music that allows AI-assisted music generation.

💡Idiogram

A free AI image generation platform focused on text-to-image capabilities. The video covers Idiogram's new 1.0 release which enhances both visual quality and text interpretation.

💡Sora

An AI video/animation model developed by Anthropic that produces highly realistic motion and speech animation from text prompts. The video speculates on how soon Sora-like capabilities may become available.

💡Terminator 2

A popular sci-fi action film. The video discusses an ambitious fan-made remake of Terminator 2 that is entirely AI-generated, marking a historic achievement.

💡Theatrical screening

The fan-made Terminator 2 remake will be the first full AI-generated film to be screened in an actual movie theater, representing a milestone for creative AI.

💡Parity

Making fun of something by imitation. The Terminator 2 remake is described as a parody, implying it humorously imitates the original film in various ways.

💡Rogues Gallery

A humorous term for a group of colorful, maverick individuals. It's used to describe the team of talented artists who worked together on the AI-generated Terminator remake.

Highlights

Morph Studio and Stability AI partnered to create an AI filmmaking platform with a node-based workflow

EmoTalker adds lifelike lip sync to still images along with emotive facial expressions

Adobe released Project Music, an AI music generator developed with UC Berkeley and Carnegie Mellon

Idiogram, the free AI image generator, released version 1.0 with improved aesthetics and text rendering

YouTuber MKBHD got early access to test out Google's video generator Sora

Researchers reverse engineered Sora, finding it uses spacetime latent patches for continuity

50 artists created the first AI-generated remake of Terminator 2 to be screened in theaters

Morph Studio's filmmaking platform has a simple node-based workflow for chaining AI models

EmoTalker was trained on 250 hours of video and 150 million images to achieve accurate lip sync

Idiogram received $80 million in funding to keep improving the free AI image generator

Researchers believe Sora-like video generation will be possible within months despite no public release

The AI-generated Terminator remake required extensive post work to meet theatrical standards

Morph aims to foster a community for sharing workflow templates on their filmmaking platform

Adobe's Project Music can extend and vary AI-generated music samples

Idiogram's new magic prompt feature suggests text to accompany images

Transcripts

play00:00

so it's been a pretty crazy week for

play00:02

Creative AI Tools in my last video I

play00:04

went over LTX Studio as well as the

play00:06

updates to Pika and Runway So today

play00:09

we're hitting all the other stuff well I

play00:11

mean at least as much as I can pack in

play00:13

today we've got news from stability. a

play00:15

on a new video generation platform a

play00:18

really impressive new lip sync tool a

play00:21

big release from Adobe in the music

play00:22

space details on how Sora actually works

play00:26

and the first fully AI generated film to

play00:29

be screened in movie theater all right

play00:31

grab a cup of coffee and buckle up

play00:34

kicking off morph Studio have partnered

play00:36

with stability AI to create a AI film

play00:38

making platform that has a really kind

play00:41

of cool workflow taking a look at this

play00:43

shot from a video that they've released

play00:45

you can see that it sort of has a

play00:46

vaguely comfy UI node-based structure to

play00:50

it but I do assure you this is much more

play00:52

simple than comfy UI if you've never use

play00:54

node-based workflows it's it takes a

play00:56

minute to get your head wrapped around

play00:58

but once you do it actually makes a lot

play01:00

of sense since you can see everything

play01:01

visually laid out here we have three

play01:03

different videos um with the style

play01:05

transfer on the third one and then as

play01:07

you connect them together you can export

play01:09

them out having control over the amount

play01:12

of influence each one gives not calling

play01:14

anybody out here but I did catch a typo

play01:16

in that first prompt a grill is looking

play01:18

at the camera uh again I'm not one to

play01:21

judge you guys catch me misspelling

play01:22

stuff all the time it does look like

play01:24

this video generator will be able to

play01:26

spell uh for example in that first video

play01:28

morph Cloud we laid see the prompt come

play01:31

up that says a cloud that spells morph

play01:33

Billows out morph's co-founder XII I I

play01:37

hope I pronounced that correctly uh said

play01:39

filming editing and post- production

play01:40

used to be separate steps in traditional

play01:42

film making but AI blurs the boundaries

play01:45

of these stages and turns them into one

play01:47

continuous process if you aren't happy

play01:49

with the shot you can regenerate it on

play01:51

our canvas AI has introduced a new

play01:54

workflow to Film Production the platform

play01:56

aims to create a Vibrant Community by

play01:58

allowing users to to share their

play02:00

workflow templates with one another via

play02:03

the gallery this one does fall under

play02:04

weight list alert I just signed up

play02:06

myself so once I get access I will

play02:08

definitely be bringing you a full look

play02:11

and I know any new video model that

play02:12

comes out in the back of everyone's head

play02:14

is like Sora Sora sore I I know we'll

play02:16

talk about that in just a second next up

play02:18

we have Emo talker which will not only

play02:20

add lip syncing to any still image but

play02:23

it also add heavy eyeliner to all of

play02:25

your

play02:28

characters

play02:31

seriously no matter what kind of music

play02:32

you're into go listen to the Black

play02:34

Parade it is a 1010 album that

play02:36

transcends any genre good is good anyhow

play02:38

emo talker which is actually emote

play02:41

portrait alive which is not an anacronym

play02:43

for emo is brought To Us by Alibaba

play02:46

let's take a quick look at it in action

play02:47

here crying is the most beautiful thing

play02:49

you can do I encourage people to cry I

play02:52

cry all the time and I think it's the

play02:54

most healthy expression of how you're

play02:57

feeling there are a number of other

play02:58

examples that you can check check out at

play03:00

the link down below most of them have

play03:02

music on them though and like in this

play03:04

case this is Eminem's rap God uh which

play03:08

is a very quick way for you know uh

play03:10

Marshall's lawyers to show up on the

play03:11

channel doorstep with a copyright strike

play03:13

so I can't play it here overall the

play03:16

thing that I'm actually super impressed

play03:17

with is kind of the emotive aspects of

play03:20

emo talker's performance there is still

play03:22

some issues I feel with like the lip

play03:24

flap but this is also an Eminem track in

play03:26

which he is rapping extremely fast uh

play03:30

but there is some issues with kind of

play03:32

like the lip movement tracking but where

play03:34

I think it really flies is with animated

play03:36

or you know kind of CG characters as we

play03:38

see in this example from the sleeper

play03:41

game Detroit Being Human game is really

play03:43

great when I was a kid I feel like you

play03:46

heard the thing you heard the term don't

play03:48

cry you don't need to cry digging into

play03:51

the paper emo talker was apparently

play03:52

trained on 250 hours of video and more

play03:56

than 150 million images it can also

play04:00

speak essentially in multiple languages

play04:02

it does apparently also use stable

play04:04

diffusion as its foundational framework

play04:06

although that said the results are very

play04:08

impressive but there are limitations for

play04:10

one you can only lips sync to still

play04:12

images so you can't like resync to video

play04:16

the paper also notes that they did not

play04:18

use any explicit control signals to

play04:20

control character movement which as they

play04:23

note uh may result in the inadvertent

play04:25

generation of other body parts such as

play04:26

hands leading to artifacts in the video

play04:29

so you know basically weird AI video my

play04:32

favorite emo talker has not been

play04:33

released yet but Pik did release their

play04:35

lip sync feature I did cover that in the

play04:37

last video link is down below moving on

play04:39

Adobe have released a new AI music

play04:42

project called Project Music gen AI

play04:45

Control I mean Adobe come on you got to

play04:47

step it up with the names here Photoshop

play04:49

Symphony InDesign Rhymes Lightroom

play04:51

lullabies I mean I'm not giving you any

play04:53

more for free call me Project Music was

play04:55

developed in collaboration with

play04:56

researchers at the University of

play04:58

California and Carnegie melon they

play05:00

released a promo video explaining some

play05:02

of the things that you can do with

play05:04

Project Music uh we'll take a look at

play05:06

the section on extending

play05:13

music all right here's the the

play05:15

lengthened

play05:25

one project music is just a research

play05:28

project we may see it in the future but

play05:31

not yet sliding back over to imagery

play05:33

idiogram the free AI image generator

play05:36

that spells better than I do has

play05:39

released a 1.0 update yeah this one is

play05:41

really cool idiogram always kind of sits

play05:43

in my back pocket when I'm trying to

play05:45

generate up something with text uh you

play05:47

know mid Journey claims that it's doing

play05:49

text but usually it's still kind of a

play05:51

garbled mess idiogram has always really

play05:53

had an edge on that front the new 1.0

play05:56

model has really upped idiograms

play05:59

aesthetic game and it actually now has a

play06:01

magic prompt button that you can turn on

play06:03

that kind of fills out your text if you

play06:04

want to use it the idiogram Community

play06:06

Feed is one of my favorites amongst all

play06:09

of the image generation Community feeds

play06:11

uh this one is actually really cool this

play06:12

is from Devil's tuna or this awesome

play06:14

penguin with the text cool people play

play06:16

base and use After Effects I don't know

play06:18

who made this but whoever did you are a

play06:19

cool person the best part about idiogram

play06:21

is that it is free it allows for 25

play06:24

Generations per day obviously if you

play06:26

want more you can then move into one of

play06:28

their subscription tiers idiogram also

play06:30

just secured $80 million in additional

play06:32

funding so that's great news hopefully

play06:34

it keeps that free tier free moving over

play06:36

to some sore news YouTube's own mares

play06:39

brownley got to play with it uh these

play06:40

are some of the generations that he got

play06:42

out of it I guess when you have 18.5

play06:45

million subscribers on YouTube you get

play06:46

to play with Sora and yes that is my

play06:49

subtle way of asking for like 18.4

play06:51

million of YouTube kindly hit the

play06:52

Subscribe button anyhow as great as Sora

play06:55

looks I still do not think that we will

play06:57

be getting it anytime soon although in

play06:59

my my last video I did mention that

play07:01

someone asked Chrystal ball valenzula

play07:02

Runway ml's CEO if there would be sore

play07:07

like outputs coming out of Runway

play07:08

anytime soon and he did say better on

play07:10

the heels of that a paper was recently

play07:12

released with a group of researchers who

play07:15

were basically reverse engineering what

play07:18

they saw in Sora I'm still going through

play07:20

the paper it is obviously very dense and

play07:22

it makes my head hurt quite a bit the

play07:25

paper does indicate that Sora utilizes

play07:27

SpaceTime latent patches which basically

play07:30

Break Down the video into smaller

play07:32

controllable pieces that it can

play07:34

understand in both space and time which

play07:36

allows for that sort of smooth

play07:38

continuity again the paper is super

play07:40

dense and while I have read a number of

play07:42

white papers at this point you know I'm

play07:44

still pretty much I'm like a caveman

play07:46

that's sitting in at a meeting at JPL

play07:48

you know I I can I can nod but you know

play07:50

at the same time do I really understand

play07:52

what's happening here but my overall

play07:54

point is that very smart people have

play07:56

already started to pull it apart and

play07:58

they are in the process of developing

play08:00

their own Sora like models even if open

play08:02

AI does not release Sora in say the next

play08:04

6 months I do think that we're going to

play08:07

see something that looks like Sora

play08:10

appearing within that time frame or

play08:13

maybe a little bit later basically

play08:14

anytime between tomorrow 6 months from

play08:16

now or some point in the future I've

play08:18

stopped predicting things because I'm

play08:20

always wrong I'm always wrong rounding

play08:21

out a historical event to my knowledge

play08:24

at least the first fulllength AI

play08:27

generated film to be screened out of

play08:29

theater so a little while back I was on

play08:31

the nerdy novelist podcast where I said

play08:33

this I think a full feature movie is

play08:35

kind of out of a fion currently with

play08:38

with the way the technology is right now

play08:39

although there are some lunatics right

play08:40

now that are creating um their own

play08:43

version of Terminator 2 wh generated by

play08:46

Ai and just to be clear when I say

play08:47

lunatics I do mean that as a term of

play08:49

endearment but yeah they did it it is 50

play08:52

artists doing essentially a cover

play08:53

version of Terminator 2 wholly generated

play08:56

by AI uh you know obviously as we see

play08:59

here you know a lot of extra work went

play09:01

into that AI but it is all still AI

play09:04

generated the entire project is

play09:06

basically this Rogues gallery of really

play09:08

awesome and talented artists uh many of

play09:11

whom have been featured on this channel

play09:13

as well uh so yeah kudos to every single

play09:16

one of you for pulling this Insanity off

play09:19

the film will have a theatrical Premiere

play09:22

in Los Angeles at the newart theater on

play09:24

March 6th uh but don't worry if you

play09:26

don't live in Los Angeles or near Los

play09:28

Angeles you can still see the movie

play09:30

online there'll be a live stream for it

play09:32

on March 9th in which the cast and crew

play09:35

will be in attendance so you can watch

play09:37

it alongside them I did also want to

play09:39

point out that because this is being

play09:40

theatrically screened there was like a

play09:42

ton of work put into it even after

play09:45

everything was done because you know you

play09:46

still have to conform the picture to

play09:48

theatrical standard and like the sound

play09:50

mix has to be ready for you know a

play09:53

theater system overall from the stuff

play09:55

that I've seen from this remake I mean

play09:56

it's it's a parody it's hilarious

play09:59

definitely please do check it out link

play10:01

is down below well that's it for this

play10:02

week I mean that's it like there was

play10:04

like 80 things that happened this week

play10:05

and I did not even get to everything but

play10:07

I don't know we'll see what's in store

play10:08

for next week I thank you for watching

play10:10

my name is

play10:16

Tim

Rate This

5.0 / 5 (0 votes)

Benötigen Sie eine Zusammenfassung auf Englisch?