日本一わかりやすいStableDiffusion WebUI AUTOMATIC1111(ローカル版)のインストール方法と基本的な使い方

テルルとロビン【てるろび】旧やすらぼ
27 Feb 202329:52

Summary

TLDRこのビデオスクリプトでは、AIが生成するキャラクターイラストを作成する方法を紹介しています。特に人気の高い「stable diffusion web UI, Automatic 1111」のインストールと使い方について、2023年2月に関連する手順を解説しています。必要な環境設定からPythonやGitのインストール、モデルの選択とダウンロード、VAEの適用方法、そして実際にAIを使って絵を生成するプロンプトの書き方とテクニックまで詳しく説明されています。さまざまなスタイルと機能を持つ複数のモデルを紹介し、AIイラストの進化と新しいモデルの登場にも触れています。

Takeaways

  • 💻 スクリプトはAI画像生成技術について説明しており、特に人気のある「stable diffusion web UI, Automatic 1111」のインストールと使用方法を紹介している。
  • 📷 AIは手描きイラストとAI生成イラストを組み合わせてキャラクターのイラストを作成することができる。
  • 🖼️ スクリプトでは、AIが生成するさまざまなスタイルのイラストを紹介しており、その中にはスケッチタッチイラストや画像のようなイラスト、ふわふわイラストが含まれている。
  • 🛠️ インストールには比較的高スペックのコンピュータが必要なため、Windows 10以降のOS、NVIDIA製のグラフィックカード、4GB以上のビデオメモリが推奨されている。
  • 🔍 PythonとGitのインストールが前提とされており、Pythonのバージョンは3.10.6が指定されている。
  • 📂 インストールフォルダ名には全角文字を避け、DドライブなどCドライブ以外を推奨している。また、インストールフォルダへのショートカットを作成することが望ましい。
  • 🔗 Gitを使用してStable Diffusion Web UIのリポジトリをクローンし、必要なモデルをダウンロードして使用する。
  • 🎨 モデルは目的に合わせて選択可能で、アニメキャラクタや実写風イラストに長けたモデルが存在する。
  • 🌐 VAE(Variational Auto Encoder)の適用がイラストの質感に影響を与えるとされ、スクリプトではStability AIのVAE 840,000を使用する例が紹介されている。
  • 🔧 プロンプト( Prompt)の書き方と設定変更がAIが生成するイラストに大きな影響を与えるとされ、スタイル、環境、主体部を含めたプロンプトの組成方法が説明されている。
  • 🎭 スクリプトでは、AIが提供する様々な機能についても触れており、テキストから画像生成(Text to Image)、画像から画像(Image to Image)、アップスケール、PNG情報の表示などが紹介されている。

Q & A

  • AIが生成するイラストの半分はどのように作られていると説明されていますか?

    -AIが生成するイラストの半分は手書きのイラストに基づいており、もう半分はAIによって生成されたイラストです。

  • Live2Dとウェブカメラを使って何ができるようになりますか?

    -Live2Dとウェブカメラを使って、絵を描くスキルがなくても自分好みのキャラクターのイラストを作成することができるようになります。

  • 「stable diffusion web UI, Automatic 1111」は何で使われていると説明されていますか?

    -「stable diffusion web UI, Automatic 1111」はローカル環境で構築され、動作する最も人気のある画像生成AIです。

  • インストールに必要なPCのスペックはどの程度必要ですか?

    -インストールに必要なPCスペックは比較的高く、Windows 10以降のOSが推奨され、NVIDIA製のグラフィックカードと少なくとも4GBのビデオメモリが必要なとされています。

  • Pythonのバージョンについてどのような注意が必要ですか?

    -Pythonのバージョンは3.10.6を使用することが推奨されており、異なるバージョンではエラーが発生する可能性があると注意が必要です。

  • Stable Diffusion Web UIをインストールする際に、どのドライブにインストールするのが良いと提案されていますか?

    -Stable Diffusion Web UIはDドライブなどのCドライブ以外にインストールすることが推奨されており、SSDにインストールすることが好ましいと提案されています。

  • モデルファイルの種類についてどのように選ぶべきですか?

    -モデルファイルはprunedとsafetensorsバージョンがバランスが良く、セキュリティと容量の両方において優れているとされています。

  • VAEとは何で、どのような効果がありますか?

    -VAEは「Variational Auto Encoder」の略で、AIイラストにおいては、イラストの仕上がりを微妙に変化させる効果があります。

  • プロンプトの書き方にはどのようなルールがありますか?

    -プロンプトは前方の言葉が強く考慮され、スタイル、環境、メインボディーの3つの要素で品質を上げることが推奨されています。

  • エミュレーションの方法として何が挙げられていますか?

    -エミュレーションの方法として「Hypernetwork」、「Dreambooth」、「LoRA」などが挙げられています。

  • AIが生成するイラストで、人体が正常に表示されない場合、どうすればよいですか?

    -人体が正常に表示されない場合は、「Restore face」機能をオフにし、解像度やアスペクト比を調整することで改善が可能な場合があります。

Outlines

00:00

🧑‍💻 AI画像生成技術の紹介

この段落では、AI画像生成技術が最近のホットトピックであることと、特に注目されている「stable diffusion web UI, Automatic 1111」について紹介しています。AIを使って自分の好きなキャラクターのイラストを簡単に作成できる方法を説明しており、必要なPCスペックやOS要件、PythonやGitなどのインストール手順について詳述されています。また、Windows 10以降での操作方法や、ファイル拡張子の確認方法も説明されています。

05:01

📂 Stable Diffusion Web UIのインストール手順

段落2では、Stable Diffusion Web UIをインストールする具体的な手順が紹介されています。まず、PythonとGitのインストール後に、インストール先のフォルダを作成し、コマンドプロンプトからGitを使用してリポジトリをクローンする方法について説明されています。次に、モデルの選択とダウンロード、モデルファイルを配置する場所について詳しく説明しており、最初の起動時に必要な手順も紹介されています。

10:03

🖼️ AI画像生成の基本操作

この段落では、Stable Diffusion Web UIの基本的な操作方法が紹介されています。ウェブUIの起動方法と、最初の画像生成までの手順について説明されています。また、UI上部のタブ機能や各タブでの操作内容についても詳述しており、テキストから画像を生成する「Text to Image」タブの使い方など、基本的な画像生成プロセスを紹介しています。

15:05

🎨 プロンプトの書き方とVAEの適用

段落4では、AIが理解するプロンプトの書き方と、画像の質感を調整するVAE(Variational Auto Encoder)の適用方法について説明されています。プロンプトの書き方として、スタイル、環境、メインボディーの3つの要素を組み合わせる方法や、品質を向上させるための「品質スペル」について紹介されています。VAEの種類と、それを適用することで得られる効果についても詳述されています。

20:09

🔍 画像生成の詳細設定と強調表現

この段落では、画像生成の詳細設定と強調表現について紹介されています。生成方法の選択肢としてカーラス系列の方法が挙げられ、ステップ数やリゾリューションなどの設定項目について説明されています。また、強調表現として、ブラケットを使用して特定の属性を強調する方法や、その効果についても紹介されています。

25:12

🌟 おすすめAIモデルの紹介

最後の段落では、様々なAIモデルについて紹介されています。異なるモデルが生成する画像の質感の違いや、それぞれが得意とするイラストのスタイルについて説明されています。また、モデルの選択肢として、アニメ風から実写風まで幅広いモデルが紹介されており、AIイラスト技術の進化と新しい機能のリリースについても触れています。

Mindmap

Keywords

💡AI

AIとは「人工知能」の略で、コンピュータが人間のように思考し、判断し、行動する技術を指します。このビデオでは、AIを用いてキャラクターのイラストを生成する技術に焦点が当てられており、AIによって手書きのイラストと組み合わせて新しいアート作品が創造されています。

💡Live2D

Live2Dは、2Dのイラストを3Dのように動かすことができる技術です。ビデオでは、Live2Dを使ってキャラクターの動きを表現することができると説明されています。これは、静的なイラストから動的な表現への一歩として、創造的なプロセスに重要な役割を果たしています。

💡stable diffusion web UI, Automatic 1111

stable diffusion web UI, Automatic 1111は、ビデオの中心となるAI画像生成ソフトウェアです。ローカル環境で動作し、ユーザーが自分の好みに合わせてキャラクターイラストを生成できるツールです。ビデオでは、このソフトウェアのインストールと使用方法が詳しく紹介されています。

💡Python

Pythonは、普及しているプログラミング言語の一つです。ビデオでは、PythonのインストールがAIソフトウェアを実行するための前提条件として挙げられています。Pythonは、AI開発においても重要な言語であり、様々なAIアルゴリズムを実装するのに使われます。

💡Git

Gitは、ソフトウェア開発においてバージョン管理を行うツールです。ビデオでは、GitのインストールがAIソフトウェアのセットアップ手順の一部として必要とされています。Gitを通じて、ソフトウェアの開発者たちは共同でプロジェクトを進め、変更履歴を管理することができます。

💡NVIDIA graphics card

NVIDIA製のグラフィックカードは、ビデオでAIソフトウェアを実行するために推奨されるハードウェアの一つです。グラフィックカードは、画像処理能力が強く、AIの学習や画像生成プロセスで高速な処理が必要な時に特に重要です。

💡SSD

SSDとは「ソリッドステートドライブ」の略で、コンピュータのストレージデバイスです。ビデオでは、SSDへのソフトウェアのインストールが推奨されており、これにより起動や読み込みが高速化されると説明されています。SSDは、データの読み書き速度がHDDよりもはるかに早く、パフォーマンス向上に寄与しています。

💡Hugging Face

Hugging Faceは、AIモデルを共有・配布するプラットフォームです。ビデオでは、Hugging Faceから様々なAIモデルをダウンロードし、Stable Diffusion Web UIで使用することができると紹介されています。このプラットフォームは、AIコミュニティにとって重要なリソースです。

💡VAE

VAEとは「Variational Auto Encoder」の略で、AIで使われるアルゴリズムの一つです。ビデオでは、VAEを適用することでイラストの質感を変化させる方法が紹介されています。VAEは、生成されたアートワークの多様性や表現力を高めるために用いられます。

💡Text to Image

Text to Imageは、テキストを元に画像を生成するAIの機能です。ビデオでは、Text to Imageタブを使ってテキストを入力し、AIがそれをもとにイラストを生成するプロセスが説明されています。この機能は、創造性と柔軟性を持ち、ユーザーが自分の想像を視覚化するのに役立ちます。

💡PNG Info

PNG Infoは、生成されたPNG画像に含まれる情報を見ることができる機能です。ビデオでは、PNG Infoタブを使って過去の生成設定を確認し、再利用することができると紹介されています。この機能は、同じ結果を再現するのにも役立ちます。

Highlights

AI生成插图技术日益普及,可以创建手绘和AI生成图像的混合体。

介绍了如何安装和使用流行的本地环境图像生成AI 'stable diffusion web UI, Automatic 1111'。

安装AI需要较高的电脑配置,推荐使用Windows 10或更高版本的操作系统。

需要NVIDIA显卡和至少4GB的视频内存,理想情况下使用12GB以及SSD硬盘。

展示如何检查操作系统版本和文件扩展名显示设置。

详细步骤教导如何安装Python和Git,Stable Diffusion Web UI的前提条件。

解释了如何通过Git克隆Stable Diffusion Web UI的代码到本地文件夹。

强调了安装文件夹名称应使用单字节字符以避免安装错误。

介绍了如何获取和安装AI模型文件,例如动漫风格的AnythingV4模型。

演示了Stable Diffusion Web UI的首次启动过程及其界面概览。

解释了如何结束Web UI会话以及下次启动的方法。

介绍了“Xformers”选项,可以提高图像生成速度。

界面介绍包括Text to Image、Image to Image、Inpaint、Extra和PNG Info等标签页的功能。

高级用户功能包括Checkpoint Merger和Train标签页,用于模型合成和学习。

详细说明了Text to Image标签页的基本操作和如何生成图像。

解释了VAE(变分自编码器)的作用以及如何安装和应用VAE文件以改善图像质量。

讨论了编写提示(prompts)的技巧,包括质量、风格、环境和主体的组合。

介绍了如何使用强调提示(emphasized prompts)来控制AI生成图像的特定特征。

分享了多个作者推荐的不同风格的AI模型,如Anything V3、ACertainthing、Seventh Anime V3等。

强调AI插图技术正在快速发展,不断有新模型和功能被发布。

Transcripts

play00:00

What will come out?

play00:02

Soooo Cute!

play00:04

Hello, I'm Robin.

play00:05

I'm Teruru.

play00:06

Recently, I've been hearing the word AI almost every day.

play00:11

AI is everything in the world.

play00:13

In fact, half of our bodies are hand-drawn illustrations, and the other half is based on AI-generated illustrations.

play00:19

And motioned with Live2D and a web camera.

play00:22

Even if you don't have any drawing skills, you can make your own favorite character illustrations.

play00:27

This is the image-generating AI that is now a hot topic.

play00:29

For example, such a sketch-touch illustration and a picture-like illustration are all generated by AI.

play00:36

And you can make such a fluffy illustration.

play00:39

This time, among the many image-generating AI,

play00:42

I would like to introduce how to install and use the most popular "stable diffusion web UI, Automatic 1111", which is built and runs in a local environment.

play00:55

It's a way of using it in February of 2023, so please understand that it may change in the future.

play01:04

First of all, what you need to install.

play01:06

Of course computer!

play01:07

However, the required specs are quite high, and the OS is after Windows 10.

play01:12

It seems that it can be built in Win7, Mac, and Linux environments, but it seems that many of the expansion functions are made in the 10or11 environment, so I recommend Windows after 10.

play01:22

What do you do when you don't know the OS?

play01:25

Press Windows and R keys at the same time on the desktop.

play01:28

Select the file name and enter DXDIAG to see the dialog called execution.

play01:35

Then the DirectX diagnosis tool will start, so you can see it in the operating system section in the system tab on page 1.

play01:43

In this video, we will talk about the Windows 11 64-bit version.

play01:50

And you need to have an NVIDIA-made graphics card.

play01:54

In addition, it is difficult if you do not have at least 4GB of video memory.

play02:02

If you just generate an illustration, it will work with 4GB, but it is ideal to have about 12GB when using the learning function.

play02:12

I think it's better to have about 30GB of drive space.

play02:18

This is to eat a lot of capacity, with a capacity of about 5GB per model.

play02:24

Also, if it is a hard disk, it will be too late and your head will be bald, so I recommend installing it on SSD.

play02:30

The last thing you should not forget is to display the extension.

play02:34

Press Windows and E keys at the same time to display the Explorer.

play02:38

Check the extension of the file name from the display.

play02:42

If this is an image file, .jpeg and .png.

play02:46

If it is a music file, .wav, .mp3, etc. will be displayed at the end.

play02:51

You can see the type of file at a glance.

play02:54

Here are the prerequisites.

play02:56

If it is a gaming PC purchased within 3 years, I think it is roughly satisfied.

play03:04

Access the page of Github's Automatic1111.

play03:09

There is an item called Installation on Windows in the middle of the page.

play03:14

There are three things you need to install in advance.

play03:17

One is Python, a programming language.

play03:25

Click on the link and jump to the download page of Python.

play03:32

Press and hold the Ctrl key and F key at the same time to display the search dialog on the page.

play03:37

Enter 3.10.6 in the dialog.

play03:41

Then you will be able to jump to 3.10.6, so click on the Installer 64-bit at the bottom and download it.

play03:48

If the version is different, errors may occur, so be careful.

play03:51

If you can download it, run the file.

play03:54

Then you will see a screen like this, so check the bottom "Add Python to PATH".

play03:59

Don't forget to mark it.

play04:02

This is to remember the location of the installed Python, so you don't have to write a long address during the program.

play04:09

Even if you don't write a few lines and numbers on the letter, if you write Python, it will be delivered to you.

play04:16

Click "Install Now" and Success.

play04:19

Yay!

play04:20

It's a good opportunity, so let's use the command prompt to check the version of the installed Python.

play04:26

What is a command prompt?

play04:29

Press the Windows key and R key at the same time on the desktop and enter CMD there.

play04:34

It's what Super computer-geek uses.

play04:36

You can check it by entering "Python space, hyphen, hyphen, and version" there.

play04:42

If display 3.10.6, you can install it properly.

play04:46

When you close it, either enter an exit or click the X in the upper right.

play04:50

This is called a command prompt, and I'll use it again later, so remember it.

play04:55

And the second thing you need is Git.

play04:58

This is also from Automatic 1111.

play05:01

Click the Click Here To Download at the top.

play05:04

The latest stable version for 64-bit Windows will be downloaded.

play05:09

If you can download it, run it, and you can press Next.

play05:13

Once you have installed Python and Git, decide where to install the Stable Diffusion Web UI.

play05:19

I don't want to install it on C drive with my habit, so this time I'm going to create a new folder right below the D drive and install it.

play05:27

Open the D drive, right-click, and create a new folder.

play05:31

I put in Stable Diffusion and Today's Date.

play05:34

At this time, please do not use Double-Byte-Characters such as Japanese in the folder name.

play05:39

This is a tradition since time immemorial when installing foreign software.

play05:43

If the root folder contains DBC such as Japanese, it may fail to install.

play05:48

Even if it goes well, there may be an error when installing another extensions later.

play05:53

When installing software, remember that the folder name should be with Single-byte-Characters.

play05:58

By the way, this install folder will be accessed frequently in the future, so it is recommended to create a shortcut around the desktop.

play06:07

Once you have a folder to install, click the empty part of the Explorer address bar in that folder.

play06:13

Then enter CMD.

play06:16

Now the command prompt has opened in the position of this folder.

play06:19

It's the one we used to check the version of Python earlier.

play06:22

Select "Git clone ~ .git" on the page of Automatic1111 and copy it.

play06:29

Go back to the command prompt, press the Ctrl key and V key to paste.

play06:34

This means duplicating the file from the page of Automatic1111 in the folder I just created in the D drive.

play06:43

Press Enter to run.

play06:45

When the cursor pops up and pops up in the same way as the first time, it's done.

play06:52

If you go back to the folder, you can see that a lot of files have been generated.

play06:56

This is the essensial files of Stable Diffusion Web UI Automatic1111.

play07:05

Actually, we can't use it yet.

play07:07

It's not written on the page of Automatic1111, but you need to get at least one model.

play07:14

Models are based on the taste of the generated illustration.

play07:18

There are more and more models, such as those that are good at anime characters and those that are good at live-action illustrations, so prepare the one that suits your purpose.

play07:26

Which model do you use?

play07:28

This time, I would like to install AnythingV4, a standard model of anime illustrations.

play07:36

Let's get a model!

play07:38

Search for Hugging Face in the browser and access.

play07:42

Enter Anything in the search window of Hugging Face.

play07:46

Then you will see some candidates, so click AnythingV4 by Andite.

play07:51

If you scroll through the page, you will see a sample, and it seems that you can generate illustrations like this.

play07:57

Cute!

play07:58

It seems that you are also good at boys and landscape paintings.

play08:01

Click "Files and Versions" above.

play08:04

The model is the one with the extension .checkpoint or .safetensors in the file list that is being released.

play08:12

The capacity is about the size of a gigabyte file.

play08:15

There is a reason why it is better to have an empty capacity.

play08:19

I made it possible to display the extension earlier so that I could identify the dropped file.

play08:25

This person seems to be uploading a lot of models, and there seems to be Anything V3, V4, and V4.5.

play08:32

Which one should I choose?

play08:34

This time, I will choose the pruned and safetensors version from V4.

play08:39

Many models have a float.16 version, float.32 version, pruned version, safe tensors version, and a full-size version, but to be honest, I don't feel much difference.

play08:53

Therefore, it seems that the pruned and safetensors version is preferred for the balance of security and capacity.

play08:59

It would be good to select the one with these labels as a priority.

play09:03

If you do this, it will be downloaded automatically.

play09:06

Once you have downloaded it, put the file in the model folder of the web UI that you deployed earlier.

play09:12

D drive, the folder I made earlier.

play09:15

Open it because there is a folder called Stable Diffusion Web UI in it.

play09:20

Open it because there is a folder called Models in it.

play09:23

Open the folder called Stable Diffusion.

play09:27

Put checkpoint here!

play09:29

It's easy to understand because there is a memo saying that the model is OK here.

play09:33

If it looks like this, it's OK.

play09:35

This is a folder for storing models, and you will be able to access it again when installing a new model, so keep it in mind.

play09:44

Launch for the first time!

play09:45

I'm ready to launch, so I'll launch for the first time.

play09:48

There is a batch file called "webui-user.bat" in the deployed file.

play09:55

This is the file to be used when launching.

play09:58

Double-click to run.

play10:00

Then a suspicious screen will appear and automatically start preparing to launch.

play10:03

It takes a while to complete.

play10:06

It takes time to launch for the first time.

play10:08

The waiting time depends on the machine performance, but it may take about 10 minutes to 1 hour.

play10:14

After waiting for a while, if it says "Running on Local URL" "To Create Public Link", the preparation is complete.

play10:22

Yay!

play10:23

Don't close this screen, leave it as it is.

play10:25

Select from "http ~ 7860" by mouse.

play10:29

Then press the Ctrl key and C key to copy.

play10:33

Paste it in the URL column of the browser with Ctrl V, press Enter to access it.

play10:38

Wow!

play10:39

This is the web UI operating screen.

play10:41

Let's register this page as you like.

play10:44

Registering will make it easier to start the next time.

play10:48

I'll mass-produce the HOT image right away.

play10:51

Leave it as it is and check the operation first.

play10:55

I think there is a command called Stable Diffusion Checkpoint on the top left of the screen.

play11:00

This is the model currently in use.

play11:02

If you confirm that the model name you just installed is displayed there,

play11:07

press the orange Generate button on the right.

play11:10

After a while, an image will be generated.

play11:14

Something is coming out.

play11:16

Once the blur is output, the operation check is successful.

play11:19

The installation is going well and it is in a state where it can be used.

play11:23

Yay!

play11:24

Let's HOT images right away!

play11:25

Now that the operation has been confirmed, I will explain how to end the web UI and how to start it the next time.

play11:31

When you finish, just close the browser and the prompt screen.

play11:35

Well, according to theory.

play11:36

To start, double-click the "webUI-user.bat" file as before.

play11:42

Wait a while for the prompt screen to start.

play11:45

After confirming that it came out with Create Public Link in the same way,

play11:49

call the web UI from browser.

play11:52

When you start, you need to do it in the order of batch file, standby, and access in the browser.

play11:59

It's not a must, but as a bonus, I'll put in "Xformers".

play12:04

After finishing the web UI, right-click the batch file and select Edit from the other options.

play12:11

Then you can open the batch file in a notepad.

play12:14

Enter "--Xformers" at the end of the set command line arguments in it.

play12:21

If you write this down, the image generation will be a little faster.

play12:25

If you can, save it and finish it.

play12:27

With this, "Xformers" will be automatically activated when it is activated after the next time.

play12:35

Let's use it!

play12:36

I was able to install it safely, so let's actually use it.

play12:39

First of all, I will briefly explain the interface.

play12:42

I think you can see that there are several tabs under the model name on the top left.

play12:47

What you can do with this tab is different.

play12:50

The default tab is the Text to Image tab.

play12:54

The number "2" is the English "TO".

play12:56

In other words, it means to generate images from text and sentences.

play13:00

In short, T2i is also written.

play13:03

Next to it is Image to Image, a function that generates images from images.

play13:07

In short, this is also written as I2i.

play13:10

The function called Inpaint, which corrects a part, is also included here.

play13:14

Next is the Extra tab.

play13:16

It is mainly used for upscaling images.

play13:18

It is easy to use because you can just specify the magnification method and magnification ratio.

play13:22

Next is the PNG Info tab.

play13:25

This is a function that displays the information engraved on the PNG image.

play13:29

In the image output in the Stable Diffusion Web UI,

play13:32

information on what kind of settings were output is engraved.

play13:36

You can view that information and use it again when you re-output it in the Text to Image.

play13:41

You can also use it to look at the settings of the AI images created by other people.

play13:46

The following Checkpoint Merger tab and Train tab are for advanced users.

play13:50

Checkpoint Merger is a synthetic function of the model as its name suggests.

play13:54

For example, you can combine a model that is good at anime illustrations and a model that is good at live-action styles

play13:59

to create a model that matches the characteristics of both companies.

play14:03

I sometimes call the combined model a "Merged model".

play14:07

The Train tab is a learning function.

play14:09

You can use it when you want to learn more about your favorite works.

play14:13

There are ways to learn, such as "Hypernetwork", "Dreambooth", and "LoRA", which is a hot topic recently.

play14:19

It's a little difficult, so you don't have to worry about it at first.

play14:23

The setting tab next to it is set as its name suggests.

play14:26

Extensions are used to add and install expansion functions.

play14:30

This time, I will explain how to use the basic Text to Image.

play14:36

In the Text to Image tab, I will introduce the basic operation while actually generating the image.

play14:41

Since it is an interface at the time of February 23,

play14:44

it may look a little different when the installation time is different.

play14:48

First, we will operate from the default state right after launching.

play14:52

The prompt and negative prompt column in the Text to Image tab.

play14:56

This is where you write what you want to generate by AI.

play14:59

I'm going to ask AI to draw a picture like this.

play15:02

First, I'll type "Girl" in the prompt column above.

play15:05

It means to draw a girl.

play15:07

Press the Generate button on the right to wait for the output.

play15:10

A girl came out.

play15:12

You can see that it became a girl more clearly than the blur that was generated by the start-up confirmation.

play15:16

Next, let's make it a smile.

play15:18

I'll add a "smile" to the prompt column.

play15:21

Separate each prompt by comma.

play15:23

When I generated it again,

play15:25

a picture of a smile was generated.

play15:28

I think you already understand.

play15:30

All you have to do is write down the attribute of the illustration you want to be drawn in the prompt column.

play15:34

It's good to remember that prompt equals to draw everyone's personality.

play15:38

By the way, the output image is automatically saved in the Outputs folder in the installed folder.

play15:44

Does this mean that i can draw a tight skirt hot lady or a little loli girl?

play15:49

Of course you can draw.

play15:51

For example, let's write it like this.

play15:53

Change the girl to a woman and add an office lady and a tight skirt.

play15:57

Then...

play15:58

Wow!

play15:59

By the way, why is the image smoked?

play16:03

Good question.

play16:04

This is because "VAE" is not applicable.

play16:07

"United Arab Emirates"??

play16:09

Think of it as a seasoning that affects the texture of the illustration.

play16:13

Let's get VAE!

play16:15

There are two types of VAE.

play16:17

One is designed by the creator specifically for that model,

play16:21

and another is a "common VAE".

play16:24

For example, if it is the current Anything V4,

play16:27

a small file called "VAE.py.touch" on the page of the download-page on Hugging Face is VAE file.

play16:34

It's easy to understand because it often has the same name as the model name.

play16:38

There are several types of VAE called standard,

play16:41

but this time I will install the one from the Stability AI.

play16:45

Access Hugging Face again and enter Stability AI in the search window,

play16:50

and you will see VAE Original in the top right corner, so click on it.

play16:54

There is a file called VAE 840,000 in the Files and Versions where you click.

play17:00

This is a VAE file.

play17:02

It's 840,000, so I'll call it "84VAE".

play17:06

I will download the SafeTensors version again this time.

play17:09

Where should I install it?

play17:11

There are two ways to apply VAE.

play17:14

There are automatic and manual applications.

play17:17

However, if possible, it is better to learn how to apply it manually,

play17:21

so I will explain in that way this time.

play17:23

The place to put the VAE file is in the Install folder.

play17:27

There is a folder called Models, and there is a folder called VAE in it.

play17:31

Put it in this folder.

play17:33

If the web UI is running, please close and restart browser and batch.

play17:39

Open the Settings tab after restarting.

play17:42

Click Stable Diffusion from the column on the left.

play17:45

There is an item called SD VAE.

play17:47

I think it's "Automatic" by default.

play17:50

If you click it, it will pull down, so change it to the 840,000 you put in earlier.

play17:55

If you set it to 840,000, press the Apply Settings above to apply it.

play17:58

Will the smoke clear with this?

play18:00

Let's compare the difference wuth the tight skirt lady.

play18:04

First, open the PNG Info tab of the web UI.

play18:08

Open the Outputs folder in the Install folder.

play18:11

Then, the image you created earlier is saved in it.

play18:15

The lady from earlier!

play18:16

Drag and drop this lady to PNG Info.

play18:20

Then, the prompt, seed value, and model name used for creation will be displayed in this way.

play18:26

Click Send to T2i.

play18:28

This will copy the prompt and seed value to the T2i screen.

play18:32

You re-summon the same lady using the same spell and seed value.

play18:37

Press the Generate button to generate the image.

play18:40

Oh, it's a vivid!

play18:42

The smoke cleared with VAE applied.

play18:45

Let's compare them side by side.

play18:47

It's totally different!

play18:49

As a side note, VAE stands for "Variational Auto Encoder".

play18:53

In AI illustrations, it has the effect of slightly changing the finish like this comparison.

play19:00

I think you've learned how to use VAE and PNG Info, so let's continue with T2i.

play19:05

You're going to teach me how to spell prompts, right?

play19:08

The way of writing prompts, are different for each person.

play19:11

If there are 100 users, there are 100 ways to write.

play19:15

Also, depending on the model being used, the way of writing that is recommended is different,

play19:19

so if you write it like this, there is no perfect way of writing.

play19:22

However, in the creation of a 2D character illustration, a certain degree of theory has been decided.

play19:27

Let's learn how to spell!

play19:29

First of all, prompts are taken into account as strongly as the one in front, that is, the one on the left.

play19:35

This is the basics of the basics, so don't forget it!

play19:38

Based on those characteristics, quality spellings are composed in the first place by style, environment, and main-body.

play19:45

You can change the order of the three in the back, or you can divide them or mix them.

play19:50

However, i recommend to keep quality in front.

play19:54

What is "quality" prompts?

play19:56

Let's take a look at the tight skirt I mentioned earlier.

play19:59

As before, it starts with T2i sent from PNG Info.

play20:04

Quality pts are intended to improve the overall quality, as the name suggests.

play20:09

As an example, there is such pts.

play20:11

In the part of "~", there are adjectives such as High, Ultra, and Best.

play20:16

I'll add Masterpiece to my lady's head.

play20:19

If you generate it with this,

play20:21

Face is a little clearer now.

play20:23

When compared, the impression has changed a lot.

play20:26

I feel a little goofy.

play20:28

Next, I'll add "Best Quality, Ultra Detailed".

play20:33

It looks like an erotic ad.

play20:35

And these quality-spells are also effective for negative prompts.

play20:39

Don't you draw elements you don't want to be negative?

play20:42

That's right.

play20:43

In other words, if you write a crappy quality-spell that you don't want,

play20:46

It's a way of thinking that the overall quality will increase.

play20:50

I'll add "Worst Quality and Low Quality" to the negative column.

play20:54

It's a fluffy character.

play20:56

But I feel like I'm getting closer to a pretty girl illustration.

play21:00

If you compare them, you'll see the importance of quality-spells.

play21:03

First of all, for the time being,

play21:05

It's okay to recognize that it's okay if you mark Masterpiece and Quality.

play21:12

Now that I've drawn the quality, I'm going to add the style and environment.

play21:15

The style is an illustration, a sketch, a watercolor, a watercolor, etc.

play21:20

You can also add the name of a famous painter in history to bring the style closer.

play21:25

The environment includes the whole picture, bust-up, and other composition information,

play21:29

In addition to the time zone, season, weather, Japanese style, medieval European style, etc.

play21:34

It includes the background as a whole.

play21:46

For example, I'll draw it like this.

play21:48

Add a sketch and a watercolor to the back of the quality.

play21:52

For the "main-body" part, I'm going to change the woman to One Girl Haz Silver Long Hair and make it silver hair long.

play21:58

In addition, to emphasize the eyes, add eyelash and jewelry eye, etc.

play22:02

Look at the viewer's eyes.

play22:05

Finally, as an environmental spelling, it's outdoor, street, and night.

play22:10

It's like an after-five of a silver-haired office lady.

play22:13

The point here is that you can write not only words, but also sentences.

play22:18

In addition, I'm writing a weather warning spell on the negative prompt.

play22:22

I'll write it in the summary section of this video.

play22:25

I'll try to output it with this.

play22:27

Wow!

play22:28

I'll compare it with the previous one.

play22:30

The eyes are beautiful and the amount of information around them has increased.

play22:33

Next, let's touch on the setting part.

play22:36

The sampling method on the top left is default, and it's an Euler.

play22:40

It's fast to generate, but it's a bit of a taste, so I'll change it.

play22:43

Which one should I use?

play22:45

It seems that the lower half of the "Karras series" is preferred for illustration generation.

play22:49

I'll write down most of the features in the summary section.

play22:52

This time, I'll use the "2M Karras".

play22:54

This is what happens when you try to change only the method and output it.

play22:57

We're going to wake up together in tomorrow morning.

play22:59

If you increase the number of steps on the right, the details will increase, but the generation will be slower.

play23:04

Moreover, even if you increase it too much, there are many things that don't change much, so it seems that 20 to 40 are preferred.

play23:10

This time, I'll use 40.

play23:12

The "Restore face" is the one that corrects the face as its name suggests, but it's a scary face in a 2D illustration, so I won't use it.

play23:19

The "Tiling" is also not used when drawing person.

play23:21

"Hi-Resfix" functions as an upscaler.

play23:24

It highlights what is generated at the resolution of the lower size, taking into account the prompts.

play23:30

The settings are really dull, so it's for advanced users.

play23:33

Width and Height are the resolution of the image.

play23:35

It seems that AI is basically thinking in 512.

play23:38

If you make it too large, body will split and proning to deformity, so be careful.

play23:43

If it's a single person, humans are vertically long, so it's good to keep the resolution a little vertical.

play23:50

Batch counts means how many times generate.

play23:53

Batch size means how many draw images per single generate.

play23:57

CFG scale means how much to follow prompts.

play24:01

Step 40, Vidos 768, Height 960.

play24:07

Legs are a little weird, but I think you can see the change when you compare it with the first one.

play24:13

She've re-trained a lot.

play24:15

If the pts and sampling settings change in the same model, even in the same seed value, it changes this much.

play24:21

This is what I output by randomly returning only the seed value.

play24:25

You keep looking for a miracle while changing the spell and sampling settings like this.

play24:32

One more, one more.

play24:34

This "generation-swamp" is the interesting part of the AI illustration.

play24:38

One more, one more.

play24:42

What's wrong?

play24:44

I put a spell in it, but it doesn't come out as I thought it would. What should I do?

play24:49

Oh, you want to generate a flat-chest character, but AI come up with big boob?

play24:53

In that case, let's use an emphasized spell.

play24:56

Emphasized spell?

play24:58

Emphasized spell, that is, an emphasized spellings.

play25:01

For example, if you want to output a new character, you can put it in a flat chest, but sometimes it doesn't work as you think.

play25:09

In that case, circle it in brackets like this.

play25:12

This will determine the flat chest to be 1.1 times.

play25:15

What did you say?

play25:16

In addition, the brackets can be layered.

play25:19

In the case of double brackets, 1.1 times 2 is 1.21 times 2, and 30 is 1.33 times 2.

play25:27

You can also put multiple words together in brackets.

play25:31

For example, if you want to emphasize skinny in addition to the flat chest.

play25:35

In this case, circle it in brackets like this.

play25:38

This will emphasize both words by 1.1 times.

play25:41

And I don't think there are many opportunities to use it, but if you circle it in brackets, it will be weak against it.

play25:47

It is 0.91 times 0.83 times.

play25:51

This can be used if you want to add just a little element in terms of hidden taste.

play25:55

And if it's a hassle to hit a lot of brackets, you can also use the method of specifying by numbers.

play26:01

For example, if you want to emphasize the flat chest by 1.3 times,

play26:05

In the case of multiple cases, if you circle it in brackets with 1.3 in the end,

play26:13

All the elements in the brackets will be emphasized.

play26:16

In addition, if you find it troublesome to type,

play26:19

After tracing the words you want to emphasize with the mouse,

play26:22

You can set the strength and weakness with the up and down arrows while pressing the control key.

play26:27

Let's take a look at the changes.

play26:30

For example, let's emphasize the older sister on board again.

play26:33

It's a status.

play26:35

I'll try to sandwich it with a flat chest.

play26:38

Then, I'll try to double the flat chest by 1.1.

play26:43

I see. Your face and body are a little out of sync.

play26:47

In that case, let's make the girl a little girl.

play26:50

Wow! It's a magical girl outside the office!

play26:53

It's important to match other elements.

play26:56

And what I want you to pay attention to is that

play26:58

Please remember that the emphasis is about 1.5 times more important.

play27:03

Also, depending on the model, it can cause the picture to break down.

play27:06

If you feel that the picture is getting dirty, you can hold it down or scrape it.

play27:10

It's important to arrange everything.

play27:15

If you've seen the video so far, I think you can already use it in a basic way.

play27:20

In that case, I'd like to try another model.

play27:23

So, at the end, I'll introduce you to some of my favorite models.

play27:28

First of all, "Anything V3".

play27:30

It is a standard model of the benchmark existence of a beautiful girl model.

play27:34

In this video, I took up V4 as an example.

play27:37

However, the texture is different, so it is difficult to press and hold.

play27:42

"ACertainthing"

play27:44

It is an image model of V3 with more background information.

play27:48

Personally, when I asked for a sketch prompt,

play27:51

Outline is clearer than V3.

play27:54

It's an easy-to-make model.

play27:59

"Seventh Anime V3"

play28:01

There are multiple versions with different textures.

play28:03

This is a model called V3C.

play28:06

It's delicate, but it's got a solid outline.

play28:09

And the sides are narrow, so it's easy to see the changes in VAE when changing.

play28:16

"Abyss Orange Mix 3"

play28:18

It's a high-quality model with a higher resolution of the face.

play28:23

By default, it's a pale texture.

play28:26

If you like a beautiful girl illustration like a doll, you might like it.

play28:31

"Counterfeit V2.5"

play28:33

It's a delicate touch of a beautiful girl model.

play28:36

It's a model with features that don't have anything in the outline.

play28:39

It's a model of the impression that foreigners think of Moe Bishoujo.

play28:46

"Pastel Mix"

play28:47

This is what the same person as Anything V4 and V4.5 is publishing.

play28:54

It's a model with a characteristic touch like a pastel painting.

play28:58

Anyway, it's a model with primary colors.

play29:00

The scenery is beautiful like a painting.

play29:03

It's a level that can be printed out and displayed.

play29:06

"Basil Mix"

play29:08

It's a live-action model specialized for Asian models.

play29:11

It's a model that can do a lot of things.

play29:14

It's a model that's a little difficult to handle to output beautifully, such as adjusting the resolution to fit the human head and body.

play29:22

Only a small part of the models featured in this video.

play29:25

The AI illustration is making great progress.

play29:28

At this moment, new models and new expansion features are being released all over the world.

play29:38

That's all for today.

play29:40

See you next week!

Rate This

5.0 / 5 (0 votes)

Related Tags
AI画像Stable DiffusionWeb UIインストール使い方画像生成テクニックアニメイラストデジタルアート創作支援
Do you need a summary in English?