Easy Guide To Ultra-Realistic AI Images (With Flux)
Summary
TLDRThe video explores the impressive advancements in AI-generated images, particularly with Stable Diffusion 3, which creates hyper-realistic images that can easily blend in with social media feeds. It discusses the use of 'Aurora' models to enhance image quality and realism, and how combining these with platforms like f.aai and Runway ML can produce convincing AI-generated videos. The host shares their trials with different tools, providing insights into the process of generating and animating ultra-realistic AI characters.
Takeaways
- 😲 AI-generated images have become incredibly realistic, making it difficult to distinguish them from real photos.
- 🎨 The script discusses Stable Diffusion 3, a model known for creating hyper-realistic images.
- 🤖 Flux, a part of the AI model, is highlighted for its ability to produce images that look like casual snapshots rather than professionally composed photos.
- 🖼️ The script mentions instances where AI-generated images have a 'plastic' or 'shiny' appearance, which can be a giveaway that they're not real.
- 🔍 The use of 'Aurora', a low-rank adapter that enhances image quality, style, or character consistency, is explained as a way to improve AI-generated images.
- 🌐 The script explores different platforms like Comfy UI, f.a, and Runway ML for generating and animating AI images.
- 💸 There's a mention of the cost associated with using AI image generation services, with some platforms offering initial credits for new users.
- 🎥 The process of animating AI-generated images to create realistic videos is discussed, including the use of Lum's Dream Machine and Runway ML's Gen 3.
- 📈 The script suggests that with the right settings and a bit of fine-tuning, it's possible to generate highly realistic AI images and videos.
- 🔧 The video script serves as a tutorial for AI enthusiasts, providing insights into the latest tools and techniques for creating ultra-realistic AI content.
Q & A
What is the main topic discussed in the script?
-The main topic discussed in the script is the advancement in AI-generated images, specifically focusing on the capabilities of Stable Diffusion 3 and the use of Aurora (AI model fine-tuning) to create highly realistic images and videos.
What is Stable Diffusion 3 known for?
-Stable Diffusion 3 is known for generating highly realistic images that can be difficult to distinguish from real photographs, especially when scrolling through social media platforms like Instagram.
How does the speaker describe the quality of AI-generated images from Stable Diffusion 3?
-The speaker describes the AI-generated images from Stable Diffusion 3 as 'really good' and 'phenomenal,' noting that they are so realistic that they could easily be mistaken for real photographs taken by a person.
What is Aurora and how does it enhance AI-generated images?
-Aurora, also known as a low-rank adapter, is a tool that can be thought of as a filter or plugin used to fine-tune AI-generated images. It allows for targeted improvements in image quality, style specificity, or character consistency without the need for extensive computational power or complete retraining of the foundational AI model.
What is the purpose of using Aurora in combination with Stable Diffusion 3?
-The purpose of using Aurora in combination with Stable Diffusion 3 is to enhance the realism of the generated images by adding extra information that improves skin, hair, and wrinkle details, making the images look more lifelike.
How does the speaker attempt to recreate the ultra-realistic AI-generated images?
-The speaker attempts to recreate ultra-realistic AI-generated images by using the flux realism Aurora model on the f.aai site, adjusting the guidance scale, and then using Runway ML to animate the generated images.
What challenges does the speaker face when generating realistic AI images with flux inside of Glyph?
-The speaker faces challenges such as the images having a 'plastic shininess' to the skin and not looking as realistic as those generated by others using Aurora. The speaker also notes the lack of options to add Aurora within the Glyph workflow builder.
What is the significance of the 'guidance scale' in the AI image generation process?
-The 'guidance scale' is significant in the AI image generation process as it determines the level of detail and realism in the output. The speaker found that setting the guidance scale to two produced more realistic results compared to the default setting.
How does the speaker evaluate the quality of the AI-generated videos?
-The speaker evaluates the quality of the AI-generated videos by looking at the realism of the movements and the consistency of the generated images, noting issues like the floating microphone and the unnatural stillness of the microphone in the animations.
What are the two main tools the speaker uses to animate AI-generated images?
-The two main tools the speaker uses to animate AI-generated images are Runway ML and Lum's Dream Machine.
What conclusion does the speaker draw about the current state of AI-generated videos?
-The speaker concludes that while the AI-generated videos are impressive, they may require multiple attempts or 'rerolls' to achieve the highest level of realism, and that some of the ultra-realistic videos circulating might be cherry-picked for their quality.
Outlines
🤖 AI's Leap in Realistic Image Generation
The speaker expresses amazement at the recent advancements in AI-generated images, particularly those from Stable Diffusion 3. They note that these images are so realistic that they could easily be mistaken for photographs on social media platforms like Instagram. The speaker highlights the imperfections in the images, such as off-center compositions, which paradoxically contribute to their authenticity. They mention that while some images still have issues with body proportions, these can be resolved with a few attempts. The speaker also discusses the use of additional tools like Aurora to enhance image quality, turning AI-generated images into highly realistic videos.
🎨 Enhancing AI Image Realism with Aurora
The speaker delves into the use of Aurora, a low-rank adapter that acts as a filter or plugin to improve the realism of AI-generated images. They explain that Aurora allows for targeted improvements in image quality, style, or character consistency without extensive computational power or retraining of the AI model. The speaker contrasts the results from using Aurora with those from the foundational model alone, noting the significant difference in realism. They also discuss the limitations of using Aurora within certain platforms like Glyph and suggest alternative methods to utilize Aurora, such as through Comfy UI or cloud-based services like f.aai.
📹 Animating AI-Generated Humans for Realistic Videos
The speaker explores the process of animating AI-generated images to create ultra-realistic videos. They demonstrate the use of Runway ML's Gen 3 Alpha to animate an AI-generated image, highlighting the challenges in achieving a realistic result, such as the unnatural movement of objects like a microphone. The speaker also compares the animation results from Runway with those from Lum's Dream Machine, finding the former to produce more convincing animations. They conclude by summarizing the steps to create realistic AI-generated videos, emphasizing the potential for further refinement and the excitement of these new tools for AI enthusiasts.
Mindmap
Keywords
💡AI Generated Images
💡Stable Diffusion 3
💡Realism
💡Flux
💡Aurora
💡Comfy UI
💡Runway ML
💡Lum's Dream Machine
💡F.aas.ai
💡Inference
Highlights
AI-generated images have become incredibly realistic, making it difficult to distinguish them from real photos.
Images from Stable Diffusion 3 are setting new standards for realism in AI-generated art.
Flux AI model is praised for creating highly realistic images that mimic snapshots from a phone.
The imperfections in AI-generated images, such as off-center compositions, contribute to their realistic appearance.
Some AI-generated images can have wonky proportions, especially when depicting full body shots.
Reddit users have been pushing the boundaries of AI image realism with tools like Flux and Aurora.
Aurora, a low-rank adapter, is used to fine-tune AI models for specific styles or character details.
Aurora models enhance image quality by improving details like skin texture, hair, and wrinkles.
The use of Aurora in combination with Flux allows for the creation of ultra-realistic AI-generated images.
Glyph app workflow builder allows users to utilize the Flux Pro version for free, but lacks Aurora integration.
Comfy UI offers complex workflows for fine-tuning AI-generated images with Aurora, but can be overwhelming for beginners.
Falling AI (f.aai) is a cloud-based service that allows users to run AI models like Flux and Aurora.
Falling AI provides a free credit for new users to experiment with AI model generation.
The Flux Realism Aurora model on Falling AI enhances the realism of generated images significantly.
Adjusting the guidance scale in Falling AI is crucial for achieving the desired level of realism.
Runway ML's Gen 3 is used to animate AI-generated images, creating ultra-realistic videos.
Lum's Dream Machine is another tool for animating AI-generated images, though results may vary.
The video concludes with a summary of the easiest methods to achieve ultra-realistic AI-generated images and videos.
Transcripts
man these AI generated images have been
really good lately I mean check out what
we're getting we got this one I mean
this one is just phenomenal check that
out like if you were just scrolling on
Instagram you would never know that that
was AI generated we've got this work of
art I
mean fantastic I cannot find anything
wrong with this image at all no no I'm
just screwing with you these are all
from stable diffusion 3 which from this
point on will probably always be known
for images like this and things like
this but truly for reals AI images have
gotten really really good I talked about
this in a recent news video about how
flux came out and people were figuring
out how to make flux even more and more
realistic we got images like this and
like this one and this one here and
here's another one I mean as you can see
it's getting harder and harder to tell
when an image was generated with AI
these are all flux and flux is just
absolutely insane at creating super
realistic images I think the fact that
they're not like perfectly composed like
they don't look like a professional
photographer took them is sort of what
gives them that feeling of all right
this looks like just a random snapshot
that somebody took it looks real like if
you were just scrolling Instagram and
you saw this without looking super super
closely you probably wouldn't know that
was AI I mean look how it's sort of
off-centered like someone taking a quick
iPhone picture would probably do now
these are all really really good there
are a few exceptions of when it starts
to get really kind of wonky and that's
when you try to get more of the body in
the shot then the proportions start to
look a little bit off but even that
often just takes a few rerolls and you
get something that looks decent now I
didn't generate any of these by the way
these are all ones that I found on
Reddit but then people on X started
taking this to another level and taking
these realistic looking images and
animating them and making them into
realistic looking videos you see a video
like this and you know there's no sound
to it but you just scroll past this this
does not look AI generated to me here's
another one of an AI generated woman
talking on stage and then the same like
ponytail dude over on the right and once
again this was generated with flux and
then it looks like they used lum's dream
machine to take that image and turn it
into a video here's another video that I
came across that looks like somebody at
a TED Talk and their paper actually has
the date on it now I kept coming across
a lot of this stuff and I was having a
really hard time getting the images that
I generated with flux to look ultra
realistic I was using this glyph doapp
workflow Builder here because it
actually lets me use the glyph pro
version for free but when I run a prompt
through the glyph version of flux I'd
get images like this which to be honest
is actually really really good really
realistic but it's got this like almost
plastic shininess to the skin that we
weren't getting in some of those other
images here's another output that tried
to generate that same guy and once again
it is pretty dang realistic but it
doesn't look like this quality here
here's the one that I just made here's
the quality of the one that was shown
off on Reddit this one hard to tell that
it's fake this one I mean just the
colors of it look off that Skin's got
like a little plasticky feel to it this
to me looks like an AI image I don't
know if I've just seen so many now that
I'm better at spotting them but there's
definitely a quality difference between
this and this the one that I generated
here came straight out of glyph with no
extra filters or anything special
running from it it's just the prompt
into glyph giving me this image the ones
in this example on the other hand they
used what's called Aur now aora is a low
rank adapter you can think of it as
almost like a filter or Plugin on top of
the normal image generation so flux is
the foundational image model which
generates the image the Laura is like
some extra sort of fine-tuning
information on top of that training
here's how perplexity explains allora
allora is used to train the model on
specific Concepts Styles or characters
allowing for targeted improvements in
image quality style specificity or
character consistency Laura models are
typically small in size ranging from 2
to 500 megabytes and can be easily
integrated into existing models to
enhance their performance it allows them
to customize their AI mod models to
produce unique art styles or improved
image quality without requiring
extensive computational power or a
complete retraining of the model so some
examples they gave Here style
specialization training the model to
generate image in a specific style such
as anime or oil painting character
specialization training the model to
generate images of specific characters
such as Mario or SpongeBob or quality
improvements enhancing the overall
quality of the generated image such as
improving the detail or Texture so
somebody basically trained one of these
lauras which works in combination with
flux without needing to retrain flux
entirely it can just add the additional
information that's needed to get to the
desired result that they're going for so
with this example here they used Aur
from Excel lab which apparently affects
the skin the hair and the wrinkles to
make the images look more realistic same
thing with this image here it was using
the same Laura to get this sort of extra
realism out of the image however when
I'm using flux inside of glyph I can
show you when I look at my actual glyph
workflow here there is no special
add-ons here there is no luras happening
even under Advanced controls we don't
even have the option to add luras in and
even if I click add block there's no
options to work with luras within here
as well something that I imagined glyph
will probably add in in the future if I
had to guess but right now we don't get
that option we get what comes straight
out of the flux foundational model
without the benefit of using that EXT ex
ra realism Laura that the people on
Reddit were using now one option to be
able to use the luras would be to use
something like comfy UI you've probably
seen a few of these comfy UI workflows
they look like kind of spaghetti bowls
with these lines going everywhere they
get complicated really really quickly
and are over the head of most people I
even struggle to wrap my head around
them once they start to get more complex
than you know three or four blocks the
other way to use this Laura would be to
use a site like
f. now this is a service similar to
replicate or like what you'd get on
hugging face spaces where you can
actually run AI models but you're using
their Cloud to run them you're using
like theile AI Cloud to run the
inference to run the processing on these
AI images now they have the standard
flux one pro model here so if you want
to just use flux Pro you can use it but
we're going to run into those issues
where if I want a super realistic image
it's not going to look as realistic as
what we're seeing because it doesn't
have that additional Laura information
on it however somebody did add the flux
realism Laura inside of f. a or foul.
here now one thing to note when you do
first start using this site this fall. a
it's not free to run the inference and
to use their Cloud computers it costs a
few cents when you do it so every time I
run flux over here on this website it
cost 32 cents or for about $1 you can
run it 29 times here's the thing though
when you first sign up as of right now
as of the recording of this video they
actually give you $2 worth of credit so
that you can get in here and play around
with this so if you do want to play with
it yourself you've got a couple bucks
you know 60 70ish Generations before you
need to start spending out of your own
pocket but once you're logged into this
fall. site you can go to fall. a/m
models and see all the various models
that are available to use here and at
the time of this recording flux one Dev
and flux real ISM are both towards the
top for you to test out and use right
now you can also use the flux one pro
which I believe is slightly more
expensive yeah so it's 5 cents per
generation typically but let's go ahead
and use the flux realism Laura here I've
already got the prompt plugged in here
that gets us a similar image to the dude
with the ponytail and under additional
settings one thing I did notice is that
you can leave the number of inference
steps at 28 here but if you leave the
guidance scale at the default 3.5 it
actually doesn't look that great it
starts to look shiny and plasticky and
and more unrealistic I found the sweet
spot to be about two I would get in here
and play with it if you don't get an
image that looks like what you want you
can play with this CFG scale here but
two was The Sweet Spot for the realism
for me I go higher than that and it
starts to look a little bit more fake so
let's go ahead and change that and click
run and we get images like this which
are pretty dang realistic looking the
forehead is still a little bit shinier
than I'd like but it's pretty dang good
the next thing I wanted to do was I
wanted to animate them like we saw in
the other videos that were circulating
all over X so let's go ahead and
download this image and we'll jump over
to Runway ml.com and I'll animate it
with Gen 3 so let's go ahead and click
started on gen 3 Alpha I can grab my
image here and drag it in the one that
we just created it wants me to crop it
so I'll go ahead and crop it like that
so his whole head is in the picture and
then I'm just going to grab the exact
same prompt that I had originally paste
it in here it's a little bit too long it
goes past their 500 here so I'm going to
get rid of the last sentence let's go
ahead and generate this see what it
gives us and here was my first attempt
with the video oh the mag the microphone
did a little magic trick there where it
just floats in midair if that didn't
just happen it's actually pretty decent
looking I mean it's a good video oh oh
right until that moment right there I
think it's a pretty solid video and then
after the floating microphone incident
it actually still looks pretty dang
solid and realistic as well I generated
one more time because this first video
that I made I accidentally had it set as
the last frame being this image that's
why the video sort of starts zoomed out
and then moves all the way to finishing
on the frame that we had here that was a
mistake I meant to set it as the first
frame so I generated it again where this
is now the first frame and here's the
version we get out of that this one is a
lot more realistic the microphone is
kind of a little too still for my taste
right he's moving a little too much and
that microphone is just like solidly
held there no matter how much he moves
his head or mouth around which just like
looks off to me it's pretty good I mean
the fingers get a little wonky here but
this is really how they're making those
videos that you're seeing all over X
right now of ultra realistic but not
real people videos the other way to do
it outside of using Runway would be to
use lum's dream machine so I went ahead
and plugged in my image into lum's dream
machine here you H the exact same prompt
and I mean the results are not quite as
good as what we got out a Runway you can
see the face sort of goes wonky towards
the end like what oh my God what just
happened there so not great Runway to me
did a lot better job unfortunately the
way AI is right now I'm fairly certain
that most of the videos that you're
seeing on X where somebody took a ultra
realistic AI generated human and made a
video of that person speaking it was
probably a little cherry-picked they
probably had to do a few rerolls like if
I generated two or three more times I
bet one of the videos that came out of
it would be just totally perfect and
hard to tell that it was AI generated
but I wanted to make a quick fun
breakdown and see if I can recreate some
of what I saw there are ways to get it
way more dialed in right if you want to
use one of those comfy UI workflows you
can get way more dialed in using
something like that but I wanted to to
try to figure out the sort of quickest
path from A to Z here the easiest way
that I've found to do it so far is to go
to this file. aai site use the flux
realism Laura model make sure you have
your guidance scale set to two but if
you don't get what you're looking for
play with that one a little bit use that
slider until you get something that
looks the way you want it then you can
take the image that it generates pull it
into Runway and now you're getting that
ultra realistic video mine don't look as
good as the examples at the beginning
but again I really really think some of
that stuff was cherry-picked just some
fun toys for us AI nerds to play with
hopefully you learned about some new
tools that maybe you hadn't heard of yet
this fall. a site is actually a new one
to me uh my buddy angry penguin who was
on that flux video with me last week is
the one who pointed this site out to me
and now I've been playing with it and
using this flux realism Laura to really
dial in these realistic AI generated
images that are coming out pretty dang
solid so anyway just thought I'd shoot a
fun video nerd out with you for a few
minutes today and show you what I've
been playing around with hope you
enjoyed it if you did like this video
subscribe to this channel more videos
like this will show up in your feed and
uh that's it really appreciate you see
you in the next one bye-bye
浏览更多相关视频
FLUX Ai | How To Create Ultra Realistic Images & Videos | Flux Ai Tutorial
How to make AI ART for FREE in 2 Minutes - NSB Pictures
24 Intelligenze Artificiali PAZZESCHE da provare nel 2024
Best AI Video Generator in 2024 (Top 5 Tools We Recommend!)
How I Created Realistic AI Influencer from Scratch | AI Model
Free ways To Access Best AI Image model yet! | Flux AI Online
5.0 / 5 (0 votes)