ComfyUI is a powerful, node-based graphical interface designed to simplify the process of generating images using Stable Diffusion models. This guide is tailored for Mac users who have loaded ComfyUI via a web browser and now need to load models and create one image from scratch quickly. The key components include models, nodes, and the user interface designed for text-to-image tasks.
ComfyUI relies on a visual programming approach where various functionalities are divided into nodes. Each node has a distinct function:
This node is essential as it allows you to load a Stable Diffusion model. Models are not included by default, so you must download them and place them in the appropriate folder structure.
The CLIP (Contrastive Language–Image Pre-training) node processes your text prompt. It helps the model understand and translate your prompt into a meaningful representation in latent space.
Samplers guide the generation process by controlling how the noise is modified. There are various presets and settings to fine-tune the image generation process.
Since ComfyUI does not ship with models, you must download a Stable Diffusion model to proceed. Here are the key steps:
Select a model such as Stable Diffusion “v1.5-pruned-emaonly-fp16.safetensors” or any other trusted variant from a reputable source. You can use popular platforms like Hugging Face or similar repositories.
After downloading the model, navigate to your ComfyUI directory on your Mac. Within the ComfyUI folder, locate the “models” subdirectory. Under “models”, create or use existing subfolders such as:
Folder | Description |
---|---|
checkpoints | Place your main Stable Diffusion models (*.safetensors files) here. |
embeddings | Save any embedding files that enhance text-to-image quality or add specific features. |
VAEs | If using a custom VAE for image quality improvements, store them here. |
LoRA | For any low-rank adaptation models (optional), use this folder. |
Ensure that the downloaded model files have the correct file extension (typically .safetensors
) and are in the appropriate folders. Improper placement can result in the nodes failing to detect it.
Once ComfyUI is loaded, you should see it running within your web browser on your Mac. Typically, this will be at a URL such as localhost:xxxx
. The interface should display the default text-to-image workflow with various nodes arranged visually.
If you are new to this node-based layout, consider clicking on a “Load Default” button found on the right panel. This resets the workspace to the default text-to-image arrangement and ensures that every node is in place.
The next critical step is to add your chosen model into the ComfyUI workflow. Follow these instructions:
Find the “Load Checkpoint” node within the default workflow. This node is typically represented as a rectangular block, and you might see a reference to the model’s name if one is already loaded.
If it is not visible, you may need to add it from the node library. Simply drag and drop it into the workspace and connect it to the appropriate input pipes.
Click on the “Load Checkpoint” node to open up the model selection dialog. Here, you should see a dropdown list of models that have been previously downloaded and placed in the “models/checkpoints” folder.
Select the desired model. If the list appears empty, double-check that the model file is correctly placed in the directory and that the file extension is supported.
Once your model is loaded, it is time to set up the workflow to generate an image. The default setup usually includes nodes for text processing and image sampling.
Apart from the “Load Checkpoint” node, there are several other crucial nodes:
For beginners, it is recommended to stick with default settings to streamline the process. However, you can experiment with the following options:
With your model loaded and workflow configured, you can now focus on generating your image through prompt entry.
Find the text input fields typically appended to the CLIP node. Here’s how to do it:
After entering your prompt, locate the “Queue Prompt” or “Generate” button on the interface. You can often find this button near the prompt input field or as part of the node configuration.
Clicking this button initiates the image generation process. On a Mac, you might also use a shortcut key (such as Cmd+Enter
) to expedite the process.
Once the process starts, ComfyUI will process your settings by passing data through the nodes—first encoding your prompt into latent space, then sampling it through the diffusion process, and finally decoding it back into an image via the VAE node.
You will see a progress indicator or status message in the interface. After a short wait, your generated image will appear on the screen in the output panel.
Exploring and experimenting is part of using ComfyUI. Below are some further tips that can help you if you encounter any problems or want to optimize your experience:
The interface is designed for interactive exploration. Use your mouse wheel or a two-finger pinch gesture on your Mac trackpad to zoom in and out of the node graph. Drag the workspace with your mouse to navigate between nodes.
If the “Load Checkpoint” node shows no models or doesn’t function as expected, verify the path of your downloaded model files. Sometimes, refreshing the application or using the command Cmd+0
to reload the default panel can resolve interface issues.
Advanced users can integrate models from other platforms, such as AUTOMATIC1111. To do this, adjust the extra_model_paths.yaml
configuration file within your ComfyUI directory. This allows more diverse model sharing and experimentation.
Step | Action | Description |
---|---|---|
Step 1 | Download & Organize Models | Obtain a Stable Diffusion model and correctly place *.safetensors files in models/checkpoints folder. |
Step 2 | Launch ComfyUI | Access the ComfyUI web interface via localhost:xxxx and load default workflow if necessary. |
Step 3 | Load a Model | Use the "Load Checkpoint" node to select the appropriate model from your organized folder. |
Step 4 | Workflow Setup | Confirm essential nodes like CLIP, Sampler, and VAE are part of the workflow and adjust settings if desired. |
Step 5 | Enter Prompt & Generate | Input positive/negative prompts, queue the generation process, monitor progress, and view the output image. |