Skip to main content
This documentation was AI-generated. If you find any errors or have suggestions for improvement, please feel free to contribute! Edit on GitHub
Loads a batch of images and their corresponding text captions from a specified directory for training purposes. The node automatically searches for image files and their associated caption text files, processes the images according to specified resize settings, and encodes the captions using the provided CLIP model.

Inputs

ParameterData TypeRequiredRangeDescription
folderSTRINGYes-The folder to load images from.
clipCLIPYes-The CLIP model used for encoding the text.
resize_methodCOMBONo”None"
"Stretch"
"Crop"
"Pad”
The method used to resize images (default: “None”).
widthINTNo-1 to 10000The width to resize the images to. -1 means use the original width (default: -1).
heightINTNo-1 to 10000The height to resize the images to. -1 means use the original height (default: -1).
Note: The CLIP input must be valid and cannot be None. If the CLIP model comes from a checkpoint loader node, ensure the checkpoint contains a valid CLIP or text encoder model.

Outputs

Output NameData TypeDescription
IMAGEIMAGEThe batch of loaded and processed images.
CONDITIONINGCONDITIONINGThe encoded conditioning data from the text captions.