A1111 - Best Advice for amazing Stable Diffusion Images

Olivio Sarikas
11 Sept 202323:43

Summary

TLDRThis video tutorial offers valuable tips for achieving stunning AI-generated images. It guides viewers on selecting the best models, utilizing negative embeddings, and leveraging various models and lora adjustments for optimal results. The host also explains how to use the interface effectively, including adjusting sliders for clip skip and choosing the right upscale models for high-resolution fixes. Additionally, the video covers the use of extensions like 'a detailer' for face tracking and 'control net' for body tracking, providing a comprehensive guide to enhance image generation.

Takeaways

  • 😀 Choose the right AI model for better results based on community ratings, hearts, and downloads.
  • 🔍 Read model descriptions and prompts to understand how to effectively use them for desired outcomes.
  • 📚 Utilize negative prompts and embeddings to exclude unwanted elements from AI-generated images.
  • 🎨 Explore different models like Realistic Vision, ref-animated, and deliberate for various artistic styles.
  • 🔧 Understand the importance of settings such as sampler, CFG scale, and high-res fix for image quality.
  • 📸 Use the 'after detailer' extension for automatic face tracking and higher-resolution in-painting.
  • 🌟 Discover and apply settings from successful images by checking prompts and parameters used.
  • 🔗 Recognize that models, lora styles, and embeddings must be compatible with the base model version you are using.
  • 🛠️ Adjust the weight of lora styles to prevent overemphasis and achieve a balanced result.
  • 🖼️ Use the image-to-image tab for making subtle changes to existing images, like altering ethnicity while keeping the same composition.
  • 🔧 Employ extensions like 'a detailer' for face tracking, 'control net' for pose tracking, and 'Roop' for face replacement to enhance image detail and customization.

Q & A

  • What is the main purpose of the video?

    -The main purpose of the video is to share tips and tricks for achieving amazing results with AI, covering both basics for beginners and advanced techniques for experienced users.

  • Why is choosing the right AI model important for the outcome of the AI-generated images?

    -Choosing the right AI model is important because a better model generally leads to better results in terms of image quality and accuracy. The model's ratings, hearts, and downloads can indicate its popularity and effectiveness within the community.

  • What are 'negative embeddings' and how are they used in AI image generation?

    -Negative embeddings are small models trained on content that the user does not want to appear in the image. They are used in the negative prompt to ensure that the AI-generated image avoids certain elements.

  • How can one find suggestions for using a specific AI model?

    -Suggestions for using a specific AI model can be found by scrolling down on the model's page, where there is information on the best positive and negative prompts, as well as other settings like the sampler and CFG scale.

  • What is the role of the 'after detailer extension' mentioned in the script?

    -The 'after detailer extension' is used for automatic face tracking and in-painting the face in higher resolution, enhancing the detail in the facial features of the AI-generated image.

  • Can you explain the significance of the 'CFG scale' in AI image generation?

    -The CFG scale determines how closely the AI sticks to the provided prompt. A lower value allows the AI to take more creative freedom, while a higher value makes the AI adhere more strictly to the model's training data.

  • What is the 'high-res fix' and how does it improve the quality of AI-generated images?

    -The 'high-res fix' is a feature that allows users to upscale the resolution of an AI-generated image for better quality. It involves using specific upscale models and adjusting settings like denoise strength to enhance image details.

  • What are 'loras' and how do they influence the style of an AI-generated image?

    -Loras are smaller versions of models that can influence the style of an AI-generated image. They add specific details or characteristics to the image, such as additional sharpness or artistic flair.

  • How can one adjust the settings for the 'image to image' tab in the AI software?

    -To adjust the settings for the 'image to image' tab, one can change the prompt, adjust the denoise strength, and use various extensions like 'a detailer' or 'control net' to modify specific parts of the image.

  • What are some tips for using the 'in painting' tab to modify parts of an AI-generated image?

    -In the 'in painting' tab, one can use a brush tool to paint over areas they wish to modify. It's important to adjust the render size to match the area being modified and to set an appropriate denoise strength to allow for changes while maintaining image quality.

  • How can one ensure that the AI-generated image matches the desired ethnicity or facial features?

    -To match the desired ethnicity or facial features, one can use the 'image to image' tab to adjust the prompt accordingly and set an appropriate denoise strength that allows the AI to make the necessary changes while staying close to the original image.

  • What is the purpose of the 'Roop' extension in AI image generation?

    -The 'Roop' extension allows users to replace the face in an AI-generated image with a photo of a real person. This can be useful for creating images that have a specific person's likeness but may require the original image to be close to the body and head shape of the person for a natural result.

  • Why is it recommended to use a lower batch size or batch count for users with slower computers or older GPUs?

    -Using a lower batch size or batch count helps prevent overloading the computer or GPU, as it allows images to be rendered one at a time or in smaller groups, improving performance on less powerful systems.

Outlines

00:00

🤖 Choosing the Right AI Model for Amazing Results

The video script begins by emphasizing the importance of selecting the right AI model to achieve the best results. It suggests looking at community ratings, hearts, and downloads to identify popular and effective models. The script introduces the concept of models and luras (smaller model versions influencing style), and provides tips on using specific models like 'Realistic Vision' and 'ref animated'. It also discusses the use of negative embeddings to exclude unwanted elements from AI-generated images and mentions the significance of choosing the right sampler and CFG scale for high-quality outputs.

05:00

🔍 Navigating the Interface and Enhancing Image Quality

This paragraph delves into the user interface of the AI software, focusing on adjusting settings like clip skip and selecting appropriate vae models. It explains how to add quick settings for clip stop and sdvae, which are crucial for customization. The script also covers the high-res fix feature, which improves image resolution and quality using upscale models like '4X Ultra Sharp'. The importance of setting the right denoise strength for image rendering is highlighted, and the process of placing models, luras, embeddings, and vae files into the correct folders within the software is outlined.

10:01

🎨 Working with Models, Luras, and Embeddings

The script discusses the compatibility of models, luras, and embeddings with different versions of AI software, such as SD 1.5 and sdxl. It explains the process of adding negative embeddings to the prompt field and adjusting the weight of luras to achieve better results. The paragraph also covers the use of the CFG scale to control how closely the AI adheres to the given prompt, and introduces the image to image tab for making subtle changes to existing images, such as altering ethnicity while maintaining the same composition.

15:01

🖌️ Advanced Image Editing with Inpainting and Extensions

This section introduces advanced image editing techniques available in the AI software, such as the inpainting tab for making detailed changes to specific parts of an image. It explains how to use the 'only mask' feature and adjust render size for局部 image re-rendering. The script also highlights various extensions that enhance the AI's capabilities, including 'a detailer' for face tracking and replacement, 'control net' for body tracking, and 'Roop' for face replacement using photos. Additionally, it provides tips for users with slower GPUs or older computers on how to use tiling for upscaling images without overwhelming their hardware.

20:01

🛠️ Optimizing AI Image Generation for Different Hardware

The final paragraph offers advice for optimizing the AI image generation process based on the user's hardware capabilities. It suggests using the 'batch count' and 'batch size' settings to manage the rendering of multiple images, which can be beneficial for those with slower or older GPUs. The script also mentions using the 'SD upscale selected' script for tiling-based upscaling to减轻GPU负担. The video concludes with a call to action for viewers to leave a like and comment if they have further suggestions or enjoyed the content.

Mindmap

Keywords

💡AI Models

AI Models refer to the various artificial intelligence algorithms designed for specific tasks, such as image generation. In the context of the video, AI models are crucial as they determine the quality and style of the generated images. The script mentions checking ratings, hearts, and downloads to identify the most beloved models by the community, indicating their effectiveness and popularity.

💡Prompts

Prompts are the textual descriptions or commands given to AI models to guide the generation of content. The video emphasizes the importance of positive and negative prompts, which help the AI understand what to include or avoid in the image. For instance, the script describes how to use 'best positive prompt' and 'negative prompt' to guide the AI in creating desired images without undesired elements.

💡Negative Embedding

Negative Embedding is a technique used in AI where a model is trained to avoid certain features or elements in the generated content. The script explains that this is done by textual inversion, which is a form of training on what not to include in the image. The video mentions downloading negative embeddings and using them in prompts to refine the AI's output.

💡Sampler

A Sampler in AI refers to the method used to select elements from a probability distribution, which is crucial in determining the final output of the AI model. The script provides examples of different samplers like Euler a or DPM++ SDE, which are used to influence the randomness and detail in the AI-generated images.

💡CFG Scale

CFG Scale stands for Control Flow Guidance Scale and it is a parameter that influences how closely the AI adheres to the provided prompts. The video script mentions that a CFG scale of 3.527 is good, indicating that it is a recommended setting for achieving a balance between adherence to the prompt and creative freedom in the AI's output.

💡High-Res Fix

High-Res Fix is a term used in the video to describe a process of improving the resolution and quality of AI-generated images. The script explains using upscale models like '4X Ultra Sharp' for enhancing image details, especially faces, which are otherwise not rendered well at lower resolutions.

💡Loras

Loras, short for 'low-resolution artifacts,' are smaller versions of AI models that can influence the style and details of the generated images. The video distinguishes between different types of Loras, such as 'XL More Artful Laura XL,' which is designed to work with specific base models like 'sdxl 1.0' and adds an artistic touch to the images.

💡Extensions

Extensions in the context of AI image generation are additional tools or functionalities that can be used to enhance or modify the AI's output. The script mentions several extensions like 'after detailer extension' for automatic face tracking and in-painting, and 'control net extension' for body tracking, which allow for more detailed and controlled image manipulation.

💡In-Painting

In-Painting is a process where missing or selected parts of an image are filled in or modified by the AI. The video script describes using the in-painting tab to make changes to specific parts of an image, such as adding sunglasses or changing facial expressions, providing a way to alter existing images without affecting the rest of the scene.

💡Batch Processing

Batch Processing refers to the ability to render multiple images or variations at once. The script discusses using 'batch count' and 'batch size' to manage the rendering of images, which can be particularly useful for users with slower or older GPUs, as it allows for rendering images sequentially rather than in parallel.

Highlights

The importance of choosing the right AI model for better results based on community ratings, hearts, and downloads.

Understanding the use of models with detailed instructions and suggestions for positive and negative prompts.

The role of negative embeddings in refining AI-generated images to exclude undesired elements.

The significance of samplers and CFG scale in achieving high-quality AI image generation.

How to utilize the high-risk fix with specific models for enhanced image details.

Insights on using extensions like After Detailer for automatic face tracking and inpainting.

Recommendations for models like 'ref animated' and their compatibility with different VAEs and negative embeddings.

The application of different models such as 'Deliberate' for photorealism and 'Epic Realism' for natural lighting.

The concept of 'loras' as smaller model versions that influence the style of AI creations.

Guidance on matching the correct version of models, loras, and embeddings for optimal AI performance.

How to navigate the Automatic1111 interface for effective use of additional parameters like clip skip and VAE model selection.

The process of implementing high-res fix for image upscaling using recommended models like '4X Ultra sharp'.

Instructions on where to place different AI components such as models, loras, embeddings, and upscalers in the Automatic1111 folder structure.

The use of 'textual inversion' for negative prompts and how to integrate it with AI models.

Adjusting the weight of 'loras' to prevent overemphasis and achieve a balanced AI-generated image.

The function of the 'image to image' tab for making subtle changes to existing AI images.

How to use the 'in painting' tab for modifying specific parts of an AI-generated image.

The integration of extensions like 'a detailer' for face tracking and 'control net' for body tracking in AI image generation.

The use of the 'Roop' extension for replacing faces in AI images with real photos.

Strategies for handling AI image rendering on older computers or with slower GPUs using tiling and batch processing.

Transcripts

play00:00

today I want to share with you some of

play00:02

the best tricks to get amazing results

play00:04

with AI this video is great for

play00:06

beginners but there's also a lot of

play00:08

stuff in here for the pros so let's get

play00:10

started to get amazing results you want

play00:13

to make the chop for you as easy as

play00:16

possible and this starts with the choice

play00:18

of the model because the better the

play00:20

model is the better your results will

play00:23

look of course there's a lot of

play00:25

different models here so when you go to

play00:27

a Civic AI you can also see here how

play00:30

many ratings does it have how many

play00:32

hearts how many downloads does it have

play00:34

so that you can see which are the most

play00:37

beloved models by the community those

play00:40

are often working the best of course

play00:42

there's different categories I want to

play00:44

show you some of my favorite models for

play00:46

example realistic vision is one that I

play00:49

really like and here already I want to

play00:52

show you something that is really

play00:54

important to understand how to work with

play00:57

such a model so for that you want to

play00:59

scroll scroll down and here you have a

play01:02

lot of information of how to use the

play01:05

model for example here is a suggestion

play01:08

for the best positive prompt all of

play01:10

these different words that you want to

play01:12

use next to the description of what you

play01:15

want to have as the image then there is

play01:17

also here a suggestion for the negative

play01:21

prompt in here on what you want to use

play01:23

for all these words so that you get the

play01:25

best out of that model there's two

play01:27

different suggestions in here so that is

play01:29

pretty awesome and you can also see here

play01:32

in Orange there is a negative embedding

play01:35

this is a very small Model A textual

play01:37

inversion as it is called this is

play01:40

trained basically on the stuff you don't

play01:42

want to have in the image and that is

play01:43

then put into the negative prompt so

play01:46

here you can click on that link and this

play01:48

will then go to the page where you can

play01:50

download the negative embedding we will

play01:53

talk about it in a minute where to put

play01:54

that then when you scroll deeper here

play01:56

you can also see you have here for

play01:59

example the sampler you want to use so

play02:01

either Euler a or DPM plus plus sde

play02:05

carers below that it says CFG scale of

play02:09

3.527 is good and then also for the high

play02:12

risk fix use the 4X Ultra sharp model

play02:15

for that again I'm gonna show you in a

play02:18

second where to put that there's a lot

play02:21

of information here that you can use but

play02:24

there is even more information that you

play02:26

can use so when you scroll up you can

play02:28

see these images and you can check out

play02:30

the images that you like for example

play02:32

this is really nice with the neon light

play02:34

in the background so click on that and

play02:37

when you click on that image you can see

play02:39

here on the right side The Prompt that

play02:41

has been used the negative prompt that

play02:43

has been used the sampler the model the

play02:46

CFG scale the steps that have been used

play02:49

and the seed you can see here it says in

play02:51

the first comment the after detailer

play02:53

extension was used to create this image

play02:55

that extension is something that does

play02:58

for example automatic face tracking and

play03:01

then automatically in painting the face

play03:03

in a higher resolution so that's also

play03:06

very nice I have a video about that

play03:08

right here to check out that extension

play03:10

other models I like are for example ref

play03:12

animated now again you want to scroll

play03:15

down and here on that page you'll find

play03:17

something really useful so when we

play03:19

scroll down again you find here a lot of

play03:21

interesting information but on top of

play03:23

that you see that this model works well

play03:25

with the vaes orange mix and the klf8

play03:29

anime 2 model and then also the Blast 2

play03:33

vae so that you can use that below that

play03:36

you find here the negative embeddings

play03:38

that I am so fond of because they are

play03:40

really helpful here's a very nice list

play03:43

of them you can click on these links to

play03:45

download them again I'm going to show

play03:47

you where to put all of that in a second

play03:49

so you really want to check out all of

play03:52

that information and for each of these

play03:54

models you want to scroll up click on

play03:56

images you like here to look what kind

play03:59

of settings prompt negative prompt has

play04:01

been used to create that and you want to

play04:03

start with this information and then

play04:06

change the prompt to work your way into

play04:08

that model to get amazing results right

play04:12

from the start now there's other models

play04:14

out there like for example deliberate

play04:17

and this can be trained on different

play04:19

styles for example this can do

play04:21

photorealism and digital painting at the

play04:23

same time there is also epic realism

play04:26

which is really good has a little bit

play04:28

more natural looking light so that's

play04:30

also very nice and then another thing

play04:33

that is really important is different

play04:35

lauras you want to use now what loras

play04:38

are is that they are smaller versions of

play04:41

models and they can basically influence

play04:44

the style that a model can create one

play04:48

that I like really much I have a video

play04:50

about that here is the XL more Artful

play04:54

Laura XL also means that this Laura is

play04:57

supposed to be used with the sdxl model

play05:00

you can see here on the right side in

play05:02

the information the type is Laura the

play05:05

base model is sdxl 1.0 so that means

play05:10

this does not work without SD 1.5 model

play05:13

and it does not work with a St 2.x model

play05:17

it has to be a sdxl model on the other

play05:20

hand here we have another Laura it's

play05:23

called a detailed tweaker now what this

play05:25

does is that it adds some more details

play05:28

to the image so that can be very useful

play05:31

but this one as you can see on the right

play05:34

side is for the base model SD 1.5 so not

play05:38

for 2.x and not for sdxl and you will

play05:42

find that most models and most lures at

play05:45

the moment on cvdi are trained for SD

play05:49

1.5 they have a lower resolution but

play05:52

often still have a very good image

play05:55

quality so you can absolutely start with

play05:58

SD 1.5 and get an amazing results out of

play06:01

that often even better than sdxl at the

play06:04

moment because like I said there are so

play06:07

many different loras for that so many

play06:09

different models that are already

play06:11

trained for that now let's have a look

play06:13

at the interface of automatic 1111 what

play06:16

are important things you might want to

play06:18

know here so first of all we have up

play06:21

here a slider for clip Skip and we have

play06:24

a Chooser for the vae model you can see

play06:27

I have for example orange mix and I have

play06:30

the vae 840 000 and also 560 000 so I

play06:35

can choose them here so they fit to the

play06:37

model and I can play around with that

play06:40

where do you get these sliders well you

play06:42

want to go to settings and then on the

play06:45

left side you want to scroll down until

play06:46

you find user interface you click on

play06:48

that you scroll down a little bit more

play06:50

until you find quick settings and in

play06:53

here you can add these additional

play06:55

parameters so when you click in an empty

play06:57

space you get a long list from where you

play06:59

can can select these different editions

play07:02

and what you want to add here is clip

play07:05

stop at last layer and also

play07:08

sdvae so these two choices when you add

play07:11

them they will add these two additional

play07:14

choices up here and then you want to

play07:16

click of course on apply settings so

play07:18

that the settings are saved and then

play07:19

when you reload the UI you have these

play07:22

two choices up here another thing this

play07:25

is really important for good quality is

play07:27

the high rest fix because usually with

play07:31

the SD 1.5 model you are for example

play07:34

rendering with 512 by 768 so with the

play07:39

high res fix if you click on that this

play07:41

is opening you have some additional

play07:42

settings here you want to use some good

play07:46

upscale models so I suggest to you the

play07:49

4X Ultra sharp model the 4X nmkd model

play07:54

and also the 8X nmkd model both of them

play07:58

are super scary male versions I will

play08:01

link all of that below the video and

play08:03

then of course you want to upscale two

play08:05

times and you want to set the denoise

play08:07

strength between 0.2 or 0.35 not too

play08:13

high because then it changes the image

play08:14

too much although you can also

play08:16

experiment with that now I want to show

play08:18

you the difference this is an image I've

play08:20

rendered with just the 512 by 768

play08:24

resolution and as you can see the face

play08:26

doesn't look too good the other details

play08:29

are kind of okay but overall the image

play08:31

could be better now here I have the high

play08:35

res fix and this as I think we can all

play08:38

agree looks amazing the face is fixed

play08:41

everything looks beautiful this is not

play08:44

using restore face this is just using

play08:46

the high-res fix in that case I'm using

play08:49

the 4X nmkd model so at this point we

play08:52

have talked a lot about different models

play08:55

and loras and embeddings where do you

play08:58

put them in the automatic take 11 11

play09:00

folder now the easiest one for that is

play09:03

the negative embeddings these go into

play09:06

the embeddings folder in here so you

play09:09

just want to double click and then you

play09:11

save them in here you can see I have

play09:13

already a bunch in here then for the

play09:16

models you want to go into the models

play09:18

folder in here

play09:19

and in there there's two important

play09:22

folders there is this stable diffusion

play09:24

folder this is where all of your models

play09:26

go for normal rendering and then up here

play09:29

you have the Laura folder this is where

play09:32

all of your loras go so that's very

play09:34

important other folders that are

play09:36

important in here for you is the es

play09:38

argon folder this is where all of the

play09:41

different app scalar models go and then

play09:43

down at the lower end you find that the

play09:45

vae folder this is of course where all

play09:48

of your vae models go there's another

play09:51

important folder for you this is for the

play09:54

control net extension because that has

play09:57

its own models for control net so for

play10:00

that you want to go into the extensions

play10:02

folder and then in there you want to

play10:04

look for the SD web UI control net

play10:07

folder and in there you want to look for

play10:09

the models folder and this is where all

play10:11

of your control net models go I'm going

play10:14

to provide a link under the video where

play10:16

you find a lot of different appscaler

play10:18

models here for example you can see the

play10:19

the 4X Ultra sharp upscaler when you

play10:23

scroll down a little bit more you see

play10:24

here the 4X nmkd super scale model is

play10:28

different versions of that and then also

play10:30

down here you'll find the 8X nmkd super

play10:33

scale model so that you can experiment

play10:36

with these upscales to see which of them

play10:38

give you the best results now when we

play10:41

talk about all these different models

play10:43

and lures and also control that models

play10:45

the important thing to remember here is

play10:48

that they have to fit the version that

play10:51

you want to use so if you want to have a

play10:54

SD 1.5 model you also need to have an SD

play10:58

1.5 Lora an SD 1.5 negative embedding

play11:02

and also for control net you want to

play11:04

have a SD 1.5 control net model and of

play11:07

course when you're working with sdxl the

play11:10

Lora has to be sdxl and the control net

play11:12

model also has to be sdxl now this is

play11:16

not true for the vae and this is all

play11:19

also not true for the upscaler that you

play11:22

use in high-res upscaling these models

play11:26

can be used on any model now let's look

play11:28

at some other interesting facts here you

play11:31

can see down here I have a tab for

play11:33

Generation but next to it there is

play11:35

textual inversion and there's also Laura

play11:37

textual inversion is the embeddings and

play11:40

this is mostly used for the negative

play11:43

embedding so when you click on that you

play11:45

can see here I have a list of the

play11:47

negative embeddings that I can use you

play11:49

want to click here into your negative

play11:51

prompts field and then click on one of

play11:54

these negative embeddings and as you can

play11:55

see here this is added now this does not

play11:58

use any kind of bracket or any kind of

play12:00

other information it is just simply the

play12:03

name of the negative embedding this is

play12:06

however different for your Laura so now

play12:08

we're on the Laura tab you want to click

play12:10

into your positive prompt and now when I

play12:13

click here for example on ADD detail you

play12:16

can see that I have the pointy brackets

play12:18

but also at the end I I have a weight of

play12:21

1 here now often what you want to do

play12:24

here is to maybe try this out to lower

play12:26

it for example to 0.8 or 0.7 as a weight

play12:31

the reason for that is that often these

play12:33

lures are over trained so they are

play12:35

actually too strong to be used and you

play12:38

want to reduce the weight to get a good

play12:40

looking result with that also you will

play12:42

sometimes find that the lower you try to

play12:44

use does not want to play well with the

play12:47

model you're using so to combat that to

play12:51

still get good results there are two

play12:53

ways you can try to fix that one is to

play12:57

have higher amount of sampling steps

play12:59

here than you would usually use so for

play13:02

example use 40 or 50 steps and another

play13:05

thing you can try is down here you have

play13:07

the CFG scale so for that you want to

play13:10

lower that a little bit for example to

play13:12

five four or even three to see if this

play13:15

will improve the image if nothing helps

play13:18

the Laura just doesn't want to work with

play13:20

your model also here's a little

play13:22

explanation for the CFG scale and what

play13:24

it actually does the CFG scale means how

play13:27

close the AI is sticking to the prompt

play13:30

that you are writing so when you have a

play13:33

lower value of that the AI is taking

play13:37

more freedom with the image but it can

play13:40

also often create better results because

play13:42

the AI can stick closer to what the

play13:45

model was trained on when you set a too

play13:48

high value for example 14 15 16 you

play13:51

might get a lot of problems with the

play13:53

results in the image and it might not

play13:55

look good now here's another interesting

play13:57

thing I want to show you on the right

play13:58

side you can see we have here these

play14:00

colorful buttons and one of them when

play14:02

you Mouse over assess image to image tab

play14:05

so when you click on that the image and

play14:07

all of the image information is sent to

play14:10

this image to image Tab and this image

play14:13

to image method is actually surprisingly

play14:16

powerful so here I have the image we

play14:19

have seen before now I'm changing the

play14:22

prompt up here instead of writing

play14:25

beautiful French woman I'm writing

play14:27

beautiful black African woman and I'm

play14:30

sitting down here my denoise strength to

play14:33

0.5 now again for the denoise strength

play14:36

the lower the value is the closer it

play14:38

sticks to the original image the higher

play14:41

the value is the more it looks different

play14:43

from the original image so with 0.5 it

play14:47

stays relatively close but also allows

play14:50

the AI to make changes and as you can

play14:52

see here we have a beautiful French

play14:54

woman on the left side and then when I

play14:57

click here on the image I have as a

play14:58

result we have a beautiful black African

play15:01

woman as a result she's wearing the same

play15:04

clothing it is the same scene with the

play15:07

same background so that you can do

play15:08

subtle changes in the image for example

play15:11

changing the ethnicity but the

play15:14

composition and the rest of that stays

play15:16

the same and that can be a very powerful

play15:18

tool to to adjust your image and play

play15:21

around with an image you already have

play15:23

and you like but you want to improve on

play15:25

that so image to image is really

play15:27

powerful to get better results another

play15:29

thing you can do here is to use this

play15:32

button that looks like a little

play15:33

painter's palette to send it to the in

play15:36

painting tab now the in painting tab is

play15:38

also in the image to image tab but down

play15:41

here you can see you have different tabs

play15:43

and one of them is called in paint you

play15:45

can see again I have the image here so

play15:48

as you can see on the right side here we

play15:50

have an icon for a paintbrush I have a

play15:52

slider here for the size of the brush

play15:54

and with that for example I can paint

play15:56

out the face to replace it or I can put

play15:59

sunglasses on it or I can have an older

play16:02

person anything I want to do also for

play16:05

any other part of the image now to make

play16:08

this work and this is really important

play16:09

you want to scroll down here and you

play16:12

want to click here on only mask and then

play16:14

also you want to adjust here the render

play16:17

size so often for example in this case I

play16:20

have something that is almost round so

play16:22

for the render size I would set for

play16:24

example 512 by 512 or 768 by 768 because

play16:30

this is only rendering that part and

play16:32

then in painting it into the existing

play16:35

image so don't use too high resolutions

play16:37

here because it is going to lead to bad

play16:40

results also another thing here is

play16:42

because we want to change the image you

play16:45

need to have the denoise strength high

play16:47

enough to enable that you want to play

play16:49

around with that but for example a value

play16:52

between 0.4 and 0.75 is good with that

play16:56

to see how soon can I actually change

play17:00

that so of course the closer the changes

play17:02

to the original image for example I want

play17:04

to have the same face but I want to have

play17:06

a angry expression for the face I can do

play17:09

with a lower value of the denoise but if

play17:13

I want to have a more intense change of

play17:16

the face for example making it the face

play17:18

of a 70 year old woman you want to have

play17:21

a higher denoise value to be able for

play17:24

the AI to actually create that kind of

play17:26

image now at that point let's go back to

play17:28

our text to image Tab and we want to

play17:30

scroll down because I want to show you

play17:32

some of the most useful extensions you

play17:35

want to have now one of them we already

play17:37

talked about is called a detailer and

play17:40

here for example you can track the face

play17:43

the hence the person in total and also

play17:46

more choices for the face now often this

play17:49

works good here for the face what I can

play17:52

do here for example is I put a Laura in

play17:55

here for example here I have a Laura for

play17:57

Nicki Minaj and then I can just replace

play18:01

the face with Allura of Nicki Minaj onto

play18:05

that body now the reasoning for that the

play18:08

why you want to use this extension is

play18:11

you want to use the Nicki Minaj Laura in

play18:15

your main prompt with a low value for

play18:17

example 0.5 so so that the face is

play18:21

attempted by the AI but also the body is

play18:24

fitting the face of the character but

play18:27

then on top of that you still give

play18:30

enough freedom for the model that you

play18:32

are using so it's not overwhelmed by the

play18:35

Laura so you get the full quality of the

play18:37

model but then also you get the body

play18:40

shape and the face shape the head shape

play18:43

from the Laura now afterwards in the

play18:46

second step with the a detailer which of

play18:49

course you need to enable here this is

play18:51

then rendering only on the face the face

play18:55

of the character you want to have so

play18:57

that you have the full quality of the

play19:00

model on the rest of the image and then

play19:03

mainly the lower quality on the face of

play19:05

the person so it looks as much as the

play19:08

person as possible another important

play19:10

extension of course is control net with

play19:12

that you can do all kinds of tracking I

play19:15

have multiple videos about that that you

play19:17

should check out you can use as second

play19:20

extension with that so here you can see

play19:22

I'm using open pose and this is tracking

play19:25

the body of the person on the left side

play19:28

in my preview here but I have another

play19:31

extension called the open pose editor

play19:34

installed this gives me this little edit

play19:37

button here so when I click on that I

play19:40

have here now a preview of everything

play19:42

and I can move these points around to

play19:45

adjust them if I feel like this tracking

play19:48

hasn't done a good job or I want to

play19:51

change the pose a little bit or there

play19:53

are some body parts missing so you can

play19:54

see down here I have person one I can

play19:57

scroll down and there are some

play19:59

informations here like the left hand the

play20:01

right hand that have not been tracked so

play20:03

I can add them to the image I can post

play20:06

them in here and then when I'm happy

play20:08

with everything I click up here on the

play20:11

send post to control net button so let's

play20:13

do some changes here and put this arm up

play20:15

here like so

play20:18

and then when I send this you can see

play20:21

that now this has updated down here and

play20:24

this can now be used to generate the

play20:27

image so this is very very useful as an

play20:29

extra extension another extension you

play20:31

want to use here is the Roop extension

play20:34

again I have a video about that here now

play20:36

what this does is that you can use any

play20:39

kind of photo and then the AI will use

play20:42

the information of that face of the

play20:45

person to render it onto the AI image

play20:48

now the good thing here is that you only

play20:51

need a single image to put it onto the

play20:53

body onto the face of any kind of AI

play20:56

image that you have rendered the

play20:58

downside of that of course is that the

play21:00

body type the head type and the rest of

play21:03

the person does not match the face or

play21:06

the body that the person actually has

play21:08

because it's only replacing the face so

play21:11

you might want to try to render an AI

play21:14

image it is as close as possible to the

play21:16

body of the person and also the head

play21:18

shape and then use Roop to put the face

play21:20

on that because otherwise it's gonna

play21:22

look a little bit strange as a result so

play21:25

here you can see an example of rupe in

play21:28

action this is the original image with

play21:30

the original face of the character and

play21:33

here we have the same image but now we

play21:35

have rendered the image of the other

play21:37

character on top of that the rest of the

play21:40

image as you can see is completely the

play21:43

same now here's another useful

play21:45

information especially for people with

play21:47

slower gpus older computers when you

play21:51

send your image to image to image to

play21:54

upscale it or upscale it even further

play21:57

what you can do instead of using the

play22:00

size down here is that you leave this as

play22:03

the original size you scroll a little

play22:05

bit lower you go here to script where it

play22:08

says SD upscale selected from that

play22:10

pop-down list here and what this is

play22:13

doing is it is creating tiles and this

play22:17

is also used using the app scalers that

play22:20

we have downloaded before words for

play22:22

example you can see here the 8X nmkd

play22:26

super scale upscaler this is using a

play22:29

scale factor of two and what this is

play22:31

doing is it is splitting it up into

play22:33

smaller tiles so that you render the

play22:36

images one after another and afterwards

play22:39

they are combined into one big image and

play22:43

last but not least I have another advice

play22:45

here for you for the text to image tab

play22:47

but also for the image to image tab when

play22:49

you scroll down here you have batch

play22:51

count and batch size this allows you to

play22:54

render multiple images or image

play22:56

variations at the same time now what

play22:59

batch size is doing is that these images

play23:01

are rendered parallel at the same time

play23:04

if you have a slow computer and older

play23:06

GPU that might be a problem for you on

play23:09

the other hand when you use the batch

play23:11

count these images going to be rendered

play23:14

one after another and that should help

play23:16

you with a slower compute or an older

play23:19

GPU let me know if I left out any

play23:21

amazing advice that I should put in

play23:23

another video like this leave a like if

play23:25

you've enjoyed this video thanks for

play23:27

watching and see you soon bye

play23:29

oh you're still here so uh This is the

play23:32

End screen there's other stuff you can

play23:34

watch like this or that's really cool

play23:36

and yeah I hope I see you soon uh leave

play23:39

a like if you haven't yet and well um

play23:42

yeah

Rate This

5.0 / 5 (0 votes)

関連タグ
AI ArtModel ChoicePrompt TipsNegative EmbeddingUpscalingImage QualityLorasControl NetExtensionsInpainting
英語で要約が必要ですか?