The EASIEST way to generate AI Art on your PC for FREE!

2 Sept 202208:28

TLDRDiscover the easiest way to generate AI art on your PC for free using Stable Diffusion. The video introduces the G-Risk GUI, a user-friendly interface for creating AI-generated images with minimal setup. By downloading the software from and using an NVIDIA graphics card, users can run the program and adjust settings like steps, V-scale, and output resolution to create detailed images based on their prompts. The process is simple, allowing for experimentation and the generation of numerous images overnight without any cost.


  • 🚀 Stable Diffusion is a powerful AI art generator that has been made publicly available with open source support.
  • 💻 The easiest way to run Stable Diffusion on a Windows PC is by using the Stable Diffusion G-Risk GUI, which requires minimal setup.
  • 🎮 An NVIDIA graphics card is needed to run the G-Risk GUI due to its use of the CUDA rendering engine.
  • 📂 Download the .rar file from, extract it, and run the .exe file to start using the Stable Diffusion G-Risk GUI.
  • 🖼️ Users can import their own image models or use the default one provided by the software.
  • 📝 Enter a text prompt to guide the AI in creating the image, and choose an output folder for the results.
  • 🌟 Seeds can be used to generate different variations of the same prompt, offering more creative control.
  • 📊 Output resolution can be adjusted based on the VRAM available on the user's graphics card.
  • ⏱️ Steps and V-scale settings affect the detail and adherence to the prompt, with recommended values for optimal results.
  • 🎨 Stable Diffusion is excellent for generating detailed and specific images, outperforming other services for abstract prompts.
  • 💡 Users can queue multiple prompts to generate a batch of images overnight, waking up to a collection of AI art.

Q & A

  • What is the main topic of the video?

    -The main topic of the video is about generating AI art on a PC for free using Stable Diffusion, specifically the G-Risk GUI.

  • What is Stable Diffusion and how does it work?

    -Stable Diffusion is an AI art generation tool that creates images based on text prompts. It works by using a graphics card, preferably NVIDIA, to leverage the CUDA rendering engine and produce accurate results based on the input.

  • Why is an NVIDIA graphics card required for this process?

    -An NVIDIA graphics card is required because Stable Diffusion utilizes the CUDA rendering engine, which is exclusive to NVIDIA. This allows for efficient and powerful computation needed for generating AI art.

  • How can one obtain the Stable Diffusion G-Risk GUI?

    -The Stable Diffusion G-Risk GUI can be obtained by downloading it from It comes as a .rar file that needs to be extracted to a folder on the user's PC.

  • What are the key components of the Stable Diffusion G-Risk GUI interface?

    -The key components include the image model selection, text prompt input, output folder choice, steps (duration and detail of image creation), vscale (adherence to the prompt), and output resolution settings.

  • What is the recommended range for steps and vscale in the Stable Diffusion G-Risk GUI?

    -The recommended range for steps is around 30 to 50 for better results, and for vscale, a default of 7.5 or between 5 and 7 is suggested for the best balance between detail and adherence to the prompt.

  • How does the output resolution setting affect the process?

    -The output resolution determines the size of the generated image. Higher resolutions use more VRAM, so users with lower-end graphics cards should be cautious and not set the resolution too high.

  • What are the benefits of using Stable Diffusion locally on one's own machine?

    -Using Stable Diffusion locally allows for full control over the process, no filters or restrictions, no need for credits or subscriptions, and the ability to generate images at one's own pace without any costs.

  • How can users experiment with different prompts and settings?

    -Users can experiment by adjusting the text prompts, steps, vscale, and output resolution to see how these changes affect the final image. They can also use seeds for more varied results.

  • What is the advantage of generating images while sleeping according to the video?

    -By queuing up multiple image generation tasks before sleeping, users can wake up to a large number of AI-generated images without actively spending their awake time waiting for the images to render.

  • What does the video suggest about the comparison between Stable Diffusion and other AI art tools?

    -The video suggests that Stable Diffusion performs better with specific and detailed prompts, while other services like Mid Journey might handle more abstract and random prompts more effectively.



🚀 Introduction to Stable Diffusion

The video begins with an introduction to Stable Diffusion, an AI-based image generator that produces highly accurate results based on user prompts. The presenter, Addie, explains that the tool has been publicly launched with open-source support, and various tools have emerged, which will be explored on the Analog Dreams YouTube channel. The focus of the video is on the easiest and most accessible way to run Stable Diffusion locally on a Windows machine with minimal setup. The presenter emphasizes the excitement around the tool and its potential to empower art and creativity. The video also mentions the necessity of an NVIDIA graphics card to run the G-Risk GUI version of Stable Diffusion due to its use of the CUDA rendering engine. Instructions for downloading and running the software are provided, along with a brief overview of the user interface and its functionalities.


🎨 Exploring Stable Diffusion's Capabilities

This paragraph delves into the capabilities of Stable Diffusion and how it can be used to generate images based on text prompts. The presenter discusses the importance of the steps and v-scale settings in creating detailed images and adhering closely to the prompt. The video demonstrates the process of generating an image, including setting up the text prompt, choosing the output folder, and selecting the output resolution. The presenter also shares personal experiences with the tool, such as the recommended number of steps for image quality and the impact of v-scale on the final result. The paragraph concludes with a live demonstration of image generation, highlighting the potential for users to create a multitude of images overnight by queuing up multiple prompts.



💡Stable Diffusion

Stable Diffusion is an AI model designed for generating images from text prompts. It uses a machine learning process to create visual content that matches the input text as closely as possible. In the video, Stable Diffusion is the primary tool discussed, which the creator uses to produce AI-generated art on their PC for free, emphasizing its ease of use and open-source availability.

💡Air Generator

Air Generator is mentioned as a tool that can create accurate results based on the user's prompts. It is likely a reference to a type of AI art generator that utilizes algorithms to interpret text prompts and produce corresponding images. The video focuses on the ease of generating high-quality AI art using such tools, with the Air Generator being one of the modules highlighted for its capabilities.

💡Open Source

Open source refers to software or tools whose source code is made available to the public, allowing anyone to view, use, modify, and distribute the software freely. In the context of the video, the presenter is excited about the launch of Stable Diffusion as an open-source project, which means that the community can contribute to its development, create modifications, and use it without restrictions, fostering innovation and accessibility in AI art generation.

💡Analog Dreams

Analog Dreams is the name of the YouTube channel featured in the video. The channel is dedicated to empowering art and creativity, particularly through the exploration of AI art generator tools and techniques. It provides tutorials, demonstrations, and discussions on various software and platforms that enable users to create digital art, with a focus on user-friendly and accessible tools like Stable Diffusion.

💡CUDA Rendering Engine

The CUDA Rendering Engine is a parallel computing platform and programming model developed by NVIDIA that allows developers to use the GPU (Graphics Processing Unit) for general purpose processing. In the video, it is mentioned as a requirement for running Stable Diffusion, as the tool leverages the power of NVIDIA's GPUs to accelerate the image generation process, making it more efficient and capable of producing high-quality AI art.

💡G-Risk GUI

G-Risk GUI is the graphical user interface (GUI) project for Stable Diffusion that is highlighted in the video. It is designed to provide an easy-to-use interface for users to generate AI art without the need for extensive technical knowledge. The presenter emphasizes the simplicity of using G-Risk GUI, which only requires a double-click to run on a Windows machine, making it accessible for beginners and those without prior experience in AI art generation.

💡Text Prompt

A text prompt is a piece of text that serves as input for AI art generation tools like Stable Diffusion. Users enter a description or concept they wish to visualize, and the AI model uses this prompt to create an image that represents the given text. In the video, the text prompt is a crucial element, as it directly influences the output of the AI-generated art, with examples provided such as 'a computer's dreams and imaginations' and 'David Harbour as Thanos'.

💡Output Resolution

Output resolution refers to the dimensions of the generated image, which determines the level of detail and quality. Higher resolutions produce more detailed images but require more VRAM (Video RAM) and processing power. In the video, the presenter discusses adjusting the output resolution to balance between image quality and the capabilities of the user's hardware, with a suggestion to use a resolution of 1920 by 1920 pixels for a good balance.


VScale is a parameter in the Stable Diffusion model that adjusts how closely the generated image adheres to the text prompt. A higher VScale value makes the AI focus more on the specific details of the prompt, potentially leading to more accurate but sometimes over-manipulated results. The video suggests a VScale value of between 5 and 7 for optimal results, as it provides a good balance between adherence to the prompt and naturalness of the image.


Steps in the context of the Stable Diffusion model refers to the number of iterations the AI performs to create the image. More steps usually result in a more detailed image but also increase the time it takes to generate the artwork. The video discusses setting the number of steps to 100 as a balance between quality and generation time, noting that the recommended number of steps is around 150 or less for optimal results.


Stable diffusion is a powerful AI art generator that can produce highly accurate results based on user prompts.

The tool has been made publicly available with open-source components, enabling a wide range of applications.

The video introduces the easiest and most accessible way to generate AI art on a Windows PC with minimal setup.

Stable diffusion can be run locally on your machine without any issues, empowering your creativity and art.

The project 'Stable Diffusion G-Risk GUI' is the focus of the video, available on for download.

To run this tool, you need an NVIDIA graphics card due to its use of the CUDA rendering engine.

The process is straightforward, involving downloading a file, extracting it, and running an executable.

The user interface is intuitive, allowing users to import image models and adjust settings like text prompts and output folders.

The 'steps' setting determines how long the AI takes to create the image and its quality, with 100 steps recommended for a balance.

The 'v scale' setting controls how closely the AI adheres to the prompt, with a default of 7.5 providing the best results.

The 'output resolution' setting affects the image size and the amount of VRAM used, with higher resolutions requiring more VRAM.

Once an image is generated, it is saved in the specified output folder along with a text file detailing the configuration.

The tool allows for experimentation with various prompts, such as 'a computer's dreams and imaginations' or 'David Harbour as Thanos'.

Users can generate multiple images overnight by queuing up prompts and letting the AI work while they sleep.

The video emphasizes the benefits of running AI art tools locally, such as having full control and no need for credits or filters.

The presenter plans to share tutorials on running more advanced Linux-based AI art tools in the future.

The video concludes by encouraging viewers to share their AI art creations and prompts in the comments or on the AI Experiments Discord channel.