【AIグラビア、LoRA、LoCon】画像1枚からLORAを作る方法、完全解説【AI絵師、Stable Diffusion】

22 Apr 202310:01

TLDRThis video script offers a comprehensive guide on creating variations of a single image using WEBUI. It instructs viewers to install WEBUI if not already done, and then use it to generate images with different poses and details. The tutorial covers the use of denoising strength to maintain fidelity to the original image, the process of inpainting to alter clothing, and the assembly of a collection of images for further processing. It also details the customization of WEBUI settings for specific tasks, the use of Google Drive for organization, and the selection of appropriate models and prompts for image generation. The script concludes with a brief mention of future tutorials on learning animations and characters.


  • 🚀 Launch the installed WEBUI from the Google Drive folder named 'Collaboration Notebook'.
  • 🔔 If prompted with warnings, click 'Allow' to ensure images are saved successfully.
  • 🎨 Create a similar image to the original by adjusting the 'Denoising Strength' slider in the WEBUI.
  • 👗 Use the 'Inpaint' feature to alter clothing by selecting the desired area with a pen tool.
  • 📂 Organize collected images in a new folder named 'Input' on Google Drive.
  • 🤖 Choose a checkpoint in the WEBUI that closely resembles the desired output for the model.
  • 🔧 Modify the 'Custom Hyperparameters' in the WEBUI with the provided long English text.
  • 📌 Ensure 'Resumable' and 'Check Enable Console Prompts' are selected for advanced settings.
  • 🌐 Install 'DreamBooth' from the Extensions tab if not already installed, following the video's instructions.
  • 🎭 Create a new model by inputting a name and selecting a base model using the 'Create' button.
  • 📝 Edit 'Instance Prompt' and 'Class Token' to specify the type of images to be used for training the model.

Q & A

  • What is the main topic of the video?

    -The main topic of the video is a step-by-step guide on how to create variations of an image using a web UI interface, specifically focusing on changing poses and clothing.

  • What is the first step in the process described in the video?

    -The first step is to launch the web UI that was installed in a previous tutorial, and access the 'Collaboration Notebook' folder on Google Drive.

  • What should you do if you encounter a warning message during the web UI setup?

    -If a warning message appears, you should click on the 'Allow' button in blue text to proceed, as this is necessary for the images to be saved.

  • How can you adjust the faithfulness of the generated images to the original in the denoising strength slider?

    -By moving the denoising strength slider, you can control the faithfulness of the generated images to the original. A lower number will produce a more faithful image, while a higher number will result in more variation from the original.

  • What is the purpose of the 'Impersonate' tab in the web UI?

    -The 'Impersonate' tab is used to create images with different clothing by specifying the areas to modify using a pen tool and then adjusting the denoising strength to control the faithfulness to the original image.

  • How does the video script instruct the user to select images for the process?

    -The user is instructed to generate about 10 images, then select approximately 5 of them. The selection should include images that are close to the original but also have some variation.

  • What is the significance of the 'ckpt' selection in the web UI?

    -The 'ckpt' selection is crucial as it determines the model used for generating the images, which can significantly influence the quality and relevance of the generated content.

  • What are the steps to upload the collected images for further processing?

    -The user is instructed to open Google Drive, create a new folder named 'Input', and upload all the collected image files into this folder.

  • How long does the learning process for the images take?

    -The learning process typically takes between 1 to 2 hours, but it may vary depending on the number of images and the user's subscription plan.

  • What should the user do if they encounter an error during the patch loading process?

    -If an error occurs during patch loading, the user should revisit the web UI startup settings, specifically checking the custom arguments and ensuring that the 'DreamBooth' extension is installed correctly.

  • What is the final step in creating a model using the web UI?

    -The final step involves clicking the 'Create' button after entering the model name, selecting the race for the model, and filling out the necessary prompts for the learning process.



🎨 Creating Variations from a Single Image

This paragraph provides a comprehensive guide on how to create variations from a single image using WEBUI. It instructs the user to launch WEBUI and navigate to a specific folder in Google Drive if the WEBUI is already installed. The script then details the process of creating images with different poses and variations, such as adjusting the denoising strength to achieve desired fidelity to the original image. It also covers the use of the 'Impersonate' feature to alter clothing and the selection of images for further processing. The user is guided through uploading images to a designated folder and using WEBUI to generate variations, with emphasis on selecting appropriate settings and extensions for the task.


🖌️ Post-Processing and Model Creation

The second paragraph focuses on post-processing the collected images and creating a model using the gathered data. It begins with instructions on how to access and set up the pre-processing image section in WEBUI, including the use of specific directories and extensions. The script then moves on to the installation of 'DreamBooth' for further customization and model creation. The user is guided through the process of selecting a model, setting up fine details like the concept, instance tokens, and class prompts, which are crucial for defining the learning process and the final output. The paragraph concludes with a note on the expected duration for learning and options for users facing time limitations due to the free version restrictions.




WEBUI stands for Web User Interface, which in the context of the video refers to a graphical interface that users interact with through a web browser. It is used to control and manage the process of creating images or models, as indicated by the script where it instructs users to launch WEBUI and make changes through it.

💡Google Drive

Google Drive is a cloud storage service that allows users to store and share files. In the video, it is used as a platform to access and store the 'Collabo Notebook' folder, which contains the necessary files for the image creation process.

💡Denoising Strength

Denoising Strength is a parameter used in image generation models to control the level of noise reduction. A lower value results in a more faithful reproduction of the original image, while a higher value allows for more variation and creativity in the generated image.


Inpainting is a technique used in image editing where parts of an image are filled or reconstructed, often to remove unwanted elements or to add new details. In the video, it is used to change the clothing of a person in an image by painting over the desired areas.


DreamBooth is a term used in the context of the video to refer to a feature or extension that seems to be part of the WEBUI system. It is likely related to creating or customizing models within the interface, as users are instructed to install it and use it to create new models.

💡Custom Hyperparameters

Custom Hyperparameters refer to the adjustable settings in a machine learning model that are not the network weights themselves but determine the quality of and control over the training process. In the video, users are guided to change these parameters to influence the outcome of their image generation.


A checkpoint in the context of machine learning and AI refers to a saved state of the model during the training process. It allows users to load a model at a certain point and continue from there, which can be useful for resuming work or applying the model to new data.


In the context of the video, 'Race' likely refers to the selection of a specific racial or ethnic type for the model being created. The script mentions choosing a model that represents an Asian female, suggesting that 'Race' here is about selecting a model that matches certain demographic characteristics.

💡Instance Prompt

Instance Prompt, as used in the video, refers to a specific type of input that guides the AI in generating an image that matches certain criteria. It is a text prompt that provides detailed information about the desired output, helping the AI to create a more accurate representation.

💡Class Token

Class Token is a term that likely refers to a categorization or classification parameter used in the AI model to specify the type of content or category that the generated images should belong to. It helps the AI understand what kind of images the user wants to generate.


In the context of the video, 'Learning' refers to the process of training the AI model with a set of images or data. This process allows the AI to understand and generate new content based on the patterns and characteristics it has learned from the provided data.


A 'Caption' in the context of the video refers to a descriptive text that accompanies an image. The script mentions an automatic captioning feature that generates a description for the images, which, while not always accurate, can be mostly reliable for human-like or complex subjects.


The video provides a comprehensive guide on creating variations from a single image.

Ensure that the WEBUI is installed and launched from the Collaborative Notebooks folder in Google Drive.

If a warning appears, click 'Allow' to ensure images are saved correctly.

Create a similar image to the original by adjusting the Denoising Strength slider.

Use the Impaint feature to alter clothing by selecting the desired areas with the pen tool.

Generate a collection of images with slight variations for a higher success rate in replacements.

Upload the edited images to a new folder named 'Input' in Google Drive.

Select a checkpoint in the WEBUI that closely matches the desired output.

Modify the Custom Hyperparameters in the WEBUI for specific settings.

Ensure the WEBUI is installed by following the instructions in the Basic Edition if not already installed.

Process the images by selecting the 'Preprocess Images' tab in the WEBUI.

The WEBUI can automatically generate captions for images, which can be helpful despite potential inaccuracies.

Click on the 'Extensions' tab and follow the instructions to load additional patches.

DreamBooth installation is required for further customization and editing.

Create a new model by inputting a name and selecting a base model in the WEBUI.

Adjust the concept settings by pasting specific URLs and editing prompts for the learning process.

The learning process may take 1-2 hours, and adjusting the number of images or opting for a paid course can help avoid time limits.

The next tutorial will focus on learning animations and non-human characters.