Running Qwen Image Edit in ComfyUI for low VRAM (GGUF) | Alpha | PandaiTech

Running Qwen Image Edit in ComfyUI for low VRAM (GGUF)

How to optimize VRAM usage by using quantized (GGUF) models so that even entry-level GPUs can run Qwen Image Edit smoothly.

Learning Timeline
Key Insights

Model Selection Guide by VRAM

Choose a GGUF variant based on your GPU capacity: Use Q2 (lightest) for 8GB VRAM, medium variants for 12GB VRAM, and any variant if you have 16GB VRAM or more.

Quality vs Performance

The Q2 version is the smallest and most VRAM-efficient, but it may produce noise or image quality degradation compared to the larger original models.
Step by Step

Setting Up GGUF Custom Nodes

  1. Open your ComfyUI interface.
  2. Click the 'Manager' button on the main menu panel.
  3. Select the 'Custom Nodes Manager' menu from the list of options.
  4. In the search bar, type 'gguf'.
  5. Look for the node named 'ComfyUI-GGUF' created by 'city96'.
  6. Click the 'Install' button (or 'Update' if you already have it to ensure you have the latest version).
  7. Click the 'Restart' button on the prompt that appears and press 'OK' to restart ComfyUI.

Configuring the Qwen GGUF Workflow

  1. Ensure the GGUF model files downloaded from Quonstack have been placed in the 'ComfyUI/models/unet/' folder.
  2. In your Qwen Image Edit workflow, locate the existing 'Load Diffusion Model' node.
  3. Delete or disconnect that 'Load Diffusion Model' node.
  4. Double-click on any empty space in the ComfyUI interface.
  5. Type and select the 'Unet Loader (GGUF)' node to add it to the canvas.
  6. On the 'Unet Loader (GGUF)' node, click the 'unet_name' dropdown menu and select the Qwen GGUF model you downloaded (e.g., Q2_K).
  7. Connect the 'MODEL' output from the 'Unet Loader (GGUF)' node to the model input of the next node in the workflow.
  8. Click the 'Queue Prompt' button to start generating images using the quantized model.

More from Generate & Edit Professional AI Images

View All