Stable Diffusion is an open-source machine learning model that can generate images from text, modify images based on text, or fill in details on low-resolution or low-detail images. It has been trained on billions of images and can produce results that are comparable to the ones you’d get from DALL-E 2 and MidJourney. You can run Stable Diffusion locally with a GUI on Windows. You can also use Stable Diffusion to make AI GIFs and videos.
Stable Diffusion (ai-tool) | |
---|---|
Full Name | Stable Diffusion |
Short Name | Stable Diffusion |
Description | An open-source machine learning model that generates images from text, modifies based on text, and enhances low-resolution images |
Company | Open Source |
Logo | |
Web | https://github.com/CompVis/stable-diffusion |
Category | AI Image, AI Video |
License | Open Source. Thanks to a generous compute donation from Stability AI and support from LAION |
Source: YouTube
Stable Diffusion is a deep learning, text-to-image model released in 2022 based on diffusion techniques. The generative artificial intelligence technology is the premier product of Stability AI and is considered to be a part of the ongoing artificial intelligence boom.
It is primarily used to generate detailed images conditioned on text descriptions, though it can also be applied to other tasks such as inpainting, outpainting, and generating image-to-image translations guided by a text prompt. Its development involved researchers from the CompVis Group at Ludwig Maximilian University of Munich and Runway with a computational donation from Stability and training data from non-profit organizations.
Stable Diffusion is a latent diffusion model, a kind of deep generative artificial neural network. Its code and model weights have been released publicly, and it can run on most consumer hardware equipped with a modest GPU with at least 4 GB VRAM. This marked a departure from previous proprietary text-to-image models such as DALL-E and Midjourney which were accessible only via cloud services.