Consistent Characters In Stable Diffusion

AI Protagonist
11 Sept 202308:53

TLDRIn this informative video, Naquan Jordan demonstrates two methods for creating consistent characters in stable diffusion. The first method involves crafting a detailed prompt, specifying characteristics like age, ethnicity, and physical features. The second method leverages control net, using an existing image as a reference to generate similar images with adjustable settings for style fidelity and control weight. Both methods aim to maintain character consistency across different images, with the latter offering a more efficient approach when control net is available.


  • 🎨 Creating consistent characters in stable diffusion involves using detailed prompts and specific character descriptions.
  • 🖌️ The first method for character consistency is crafting a prompt with intricate details such as age, ethnicity, hair and eye color, and clothing.
  • 👩‍🦱 Including a character's name, preferably a fictional one, can help in recreating them more accurately.
  • 📸 The second method for consistent character creation is using a tool called control net, which requires selecting an image and sending it to control net for variations.
  • 🔍 Ensure intelligent analysis is turned off in control net to avoid the tool creating the image with the best models or prompts it can think of.
  • 🔄 Reference generation in control net allows uploading images of characters, objects, or items to generate similar new images.
  • 🎚️ Style fidelity and control weight are adjustable settings in control net that determine how closely the generated image follows the reference.
  • 🔄 Experiment with different control modes in control net, such as prioritizing props or preprocessing, to achieve the desired character consistency.
  • 👗 Clothing consistency can be an issue with stable diffusion, so using control net can be more effective in maintaining character consistency across images.
  • 📸 Test different settings in control net to find the right balance between character consistency and variation in poses, lighting, and other elements.
  • 💡 The video provides a tutorial on character recreation, offering solutions for those looking to maintain character consistency across different platforms and models.

Q & A

  • What is the main topic of the video?

    -The main topic of the video is about creating consistent characters in stable diffusion.

  • How can one recreate characters from previous prompts in new models and platforms?

    -One can recreate characters by using very detailed prompts that include specific characteristics of the character such as age, ethnicity, hair and eye color, and other physical features.

  • Why is it important to include intricate details in the character description?

    -Including intricate details in the character description helps stable diffusion to more accurately recreate the character's appearance, ensuring consistency across different images and platforms.

  • What is the role of a name in character recreation?

    -Using a name, especially a unique one, can help stable diffusion to recreate the character more consistently, as it adds another layer of identification to the character's identity.

  • What is the second method for creating consistent characters as mentioned in the video?

    -The second method is using control net. This involves selecting an image of the character, sending it to control net, and adjusting settings like style fidelity and control weight to generate similar new images.

  • What should be considered when using control net for character recreation?

    -When using control net, one should pay attention to the style, fidelity, and control weight settings to determine how closely the generated image follows the reference image.

  • How does the control mode in control net affect the output?

    -The control mode in control net determines whether the focus is on prioritizing props, pre-processing, or maintaining a balanced approach, which in turn affects the final generated image.

  • What is the issue with clothing when using stable diffusion?

    -The issue with clothing in stable diffusion is that it may not always recreate the clothing accurately, as seen in the example where the dress did not always appear as intended.

  • How can one ensure that the character's face is consistently recreated?

    -To ensure the character's face is consistently recreated, one can use the 'restore faces' feature in control net and adjust the settings to achieve the desired level of facial consistency.

  • What is the advice given for users who have questions or want to showcase their work?

    -The video encourages users to leave their questions in the comments section and share any showcases they'd like to present, so the creator can engage with their audience and view their work.



🎨 Character Consistency in Art Creation

This paragraph introduces the topic of creating consistent characters using stable diffusion, a technique for generating images from textual descriptions. The speaker, Naquan Jordan, addresses common questions from viewers about character recreation and announces a tutorial on the subject. The paragraph outlines two methods for achieving consistency: detailed textual prompts and the use of control nets. The first method emphasizes the importance of detailed descriptions, including ethnicity, background, and physical features, to recreate characters accurately. The second method involves using control nets to generate images that closely follow a reference image's style and character, with adjustments for style fidelity and control weight.


🖌️ Enhancing Character Consistency with Control Nets

In this paragraph, the speaker continues the discussion on character consistency, focusing on the use of control nets for more precise recreation of characters. The process involves selecting an image and adjusting control net settings to generate variations that maintain the original character's features. The paragraph explains the importance of style fidelity and control weight in achieving the desired level of similarity to the reference image. The speaker demonstrates the effectiveness of control nets by showing how it can recreate the same character with different poses and lighting, emphasizing the efficiency of this method over the traditional prompt-based approach. The paragraph concludes with an invitation for viewers to ask questions, request further tutorials, and share their own work.




In the context of the video, characters refer to the fictional individuals that the creator aims to depict consistently across different platforms and models. The video focuses on the techniques to recreate these characters with a consistent appearance and attributes. For example, the character 'Stacy Long' is described in detail to ensure that she can be recreated in various images with the same features.


Prompts are the textual descriptions or commands used to guide the creation of images using artificial intelligence, such as stable diffusion. In the video, the creator discusses the importance of detailed prompts to recreate characters accurately. A prompt might include details like age, ethnicity, hair and eye color, and clothing to help the AI generate the desired character image.

💡Stable Diffusion

Stable diffusion is an AI model mentioned in the video that generates images based on prompts. It is used to create and recreate characters with consistent features. The video discusses challenges with clothing consistency in stable diffusion and methods to improve character consistency, such as using detailed prompts and control nets.

💡Control Net

Control net is a tool or feature that allows users to guide the AI generation process more precisely by using reference images. In the video, the creator demonstrates how to use control net to generate images of a consistent character by adjusting settings like style fidelity and control weight. This method helps in maintaining the character's appearance across different images.


Variations refer to the different iterations or versions of an image that can be generated using AI models like stable diffusion or control net. The video shows how variations can be used to create multiple images of the same character with slight differences in pose, lighting, or other elements while maintaining the core characteristics of the character.

💡Image Generation

Image generation is the process of creating visual content using AI models, as discussed in the video. It involves inputting prompts or using control nets with reference images to produce a desired output. The video focuses on generating consistent character images through this process.

💡Reference Generation

Reference generation is a feature within control net that allows users to upload images as references to generate new, similar images. This tool is crucial for maintaining consistency in character design, as it uses the uploaded image's style and features as a basis for creating new images.

💡Style Fidelity

Style fidelity is a parameter in reference generation that adjusts how closely the generated image follows the reference image's style. It is a measure of the adherence to the original image's visual characteristics, allowing for control over the similarity between the reference and the generated images.

💡Control Weight

Control weight is another parameter in reference generation that determines the strength of the control net's influence on the generated image. A higher control weight results in a more pronounced effect of the reference image on the new image, ensuring a closer match to the original character's features.

💡Character Consistency

Character consistency refers to the ability to recreate a character with the same visual attributes across multiple images or platforms. The video provides techniques for achieving character consistency, such as using detailed prompts and control net, which are essential for creators who want their characters to be recognizable and uniform in different contexts.

💡AI Models

AI models, as mentioned in the video, are the underlying algorithms or systems that power artificial intelligence applications, such as image generation. These models are trained to perform specific tasks, like creating images based on textual prompts or reference images, and are crucial for achieving the desired output in character creation and consistency.


The video discusses methods for creating consistent characters in stable diffusion.

Recreating characters from previous prompts is possible by using detailed prompts or control net.

A detailed prompt includes characteristics like age, ethnicity, hair and eye color, and clothing.

Adding intricate details like eye detail and specific clothing items can enhance character consistency.

Using a first and last name for the character can help in recreating the character more accurately.

Control net allows for the generation of similar new images based on a reference image.

Intelligent analysis should be turned off when using control net for consistent results.

Reference generation in control net uses images as references to create new images of the same character.

Adjusting style fidelity and control weight in control net influences how closely the new image follows the reference.

Control modes in control net can prioritize props, preprocessing, or maintain a balance.

Control net can recreate the same character in different poses and camera lengths.

Care should be taken not to set control net settings too high to avoid off-character results.

The video provides a practical guide for artists and designers looking to use stable diffusion for character creation.

The presenter encourages viewers to share their questions and character showcases in the comments.

The video serves as a tutorial for those interested in using stable diffusion for consistent character generation.