Skip to main content
This documentation was AI-generated. If you find any errors or have suggestions for improvement, please feel free to contribute! Edit on GitHub
This node generates images synchronously using Stability AI’s Stable Diffusion 3.5 model. It creates images based on text prompts and can also modify existing images when provided as input. The node supports various aspect ratios and style presets to customize the output.

Inputs

ParameterData TypeRequiredRangeDescription
promptSTRINGYes-What you wish to see in the output image. A strong, descriptive prompt that clearly defines elements, colors, and subjects will lead to better results. (default: empty string)
modelCOMBOYesMultiple options availableThe Stable Diffusion 3.5 model to use for generation.
aspect_ratioCOMBOYesMultiple options availableAspect ratio of generated image. (default: 1:1 ratio)
style_presetCOMBONoMultiple options availableOptional desired style of generated image.
cfg_scaleFLOATYes1.0 to 10.0How strictly the diffusion process adheres to the prompt text (higher values keep your image closer to your prompt). (default: 4.0)
seedINTYes0 to 4294967294The random seed used for creating the noise. (default: 0)
imageIMAGENo-Optional input image for image-to-image generation.
negative_promptSTRINGNo-Keywords of what you do not wish to see in the output image. This is an advanced feature. (default: empty string)
image_denoiseFLOATNo0.0 to 1.0Denoise of input image; 0.0 yields image identical to input, 1.0 is as if no image was provided at all. (default: 0.5)
Note: When an image is provided, the node switches to image-to-image generation mode and the aspect_ratio parameter is automatically determined from the input image. When no image is provided, the image_denoise parameter is ignored.

Outputs

Output NameData TypeDescription
imageIMAGEThe generated or modified image.