SwarmUI provides a UI that can handle multiple ComfyUI instances as backends at once.
https://github.com/mcmonkeyprojects/SwarmUI

Currently, ComfyUI does not provide a method to execute workflows in parallel.

If you are a developer and want to implement inference functionality for multiple GPUs, I think modifying the KSampler would be the most effective approach.

If I had a multiple GPU environment, I would like to experiment with this, but I'm not sure if PyTorch can properly handle this scenario.

It's important to note that several custom nodes use implementations that hijack the Sampling function. Your modifications might cause issues with these nodes.

🌐
GitHub
github.com › pollockjj › ComfyUI-MultiGPU
GitHub - pollockjj/ComfyUI-MultiGPU: This custom_node for ComfyUI adds one-click "Virtual VRAM" for any UNet and CLIP loader as well MultiGPU integration in WanVideoWrapper, managing the offload/Block Swap of layers to DRAM *or* VRAM to maximize the latent space of your card. Also includes nodes for directly loading entire components (UNet, CLIP, VAE) onto the device you choose
This custom_node for ComfyUI adds one-click "Virtual VRAM" for any UNet and CLIP loader as well MultiGPU integration in WanVideoWrapper, managing the offload/Block Swap of layers to DRAM *or* VRAM to maximize the latent space of your card. Also includes nodes for directly loading entire components (UNet, CLIP, VAE) onto the device you choose - pollockjj/ComfyUI-MultiGPU
Starred by 687 users
Forked by 48 users
Languages   Python 99.5% | Shell 0.5%
🌐
Reddit
reddit.com › r/comfyui › sanity check: using multiple gpus in one pc via comfyui-multigpu. will it be a benefit?
r/comfyui on Reddit: Sanity check: Using multiple GPUs in one PC via ComfyUI-MultiGPU. Will it be a benefit?
April 22, 2025 -

I have a potentially bad idea, but I wanted to get all of your expertise to make sure I'm not going down a fruitless rabbit hole.

TLDR: I have a one PC with a 4070 12gb and one PC with a 3060 12gb. I run AI on both separately. I purchased a 5060 Ti 16gb.

My crazy idea is to get a new motherboard that will hold 2 graphics cards and use ComfyUI-MultiGPU to set up one of the PCs to run two GPUs (Most likely the 4070 12gb and 3060 12gb) and allow it to offload some things from the VRAM of the first GPU to the second GPU.

From what I've read in the ComfyUI-MultiGPU info it doesn't allow for things like processing on both GPUs at the same time, only swapping things from the memory of one GPU to the other.

It seems (and this is where I could be mistaken) that while this wouldn't give me the equivalent of 24GB of VRAM it might allow for things like GGUF swaps onto and off of the GPU and allow the usage of models over 12GB in the right circumstances.

The multi-GPU motherboards I am looking at are around $170-$200 or so and I figured I'd swap everything else from my old motherboard.

Has anyone had experience with a set up like this and was it worth it? did it help in enough cases that it was a benefit?

As it is I run two pcs and this allows me to do separate things simultaneously.

However, with many things like GGUF and block swapping allowing things to be run on cards with 12GB this might be a bit of a wild goose chance.

What would the biggest benefit of a set up like this be if any?

Discussions

Multi-GPU Support
Hi! I'm interested in using ComfyUI with multiple GPUs for both training and inference. Specifically, I'm planning to utilize H100s. However, I haven't been able to find any specific op... More on github.com
🌐 github.com
9
9
July 30, 2024
Made a ComfyUI extension for using multiple GPUs in a workflow
I wanted to find out what it would take to add proper multi-GPU support to ComfyUI. While this is not it, these custom nodes will allow you to pick which GPU to run a given model on. This is useful if your workflow doesn't completely fit in VRAM on a single GPU. On my testing setup (2x 3090) there is a noticeable improvement when running flux dev by offloading the text encoders & VAE to the 2nd GPU. It's implemented in a very hacky but simple way and I'm surprised it even works. I saw some requests for this on the sub recently so hopefully it's useful to somebody. More on reddit.com
🌐 r/StableDiffusion
37
97
August 4, 2024
Multiple GPU Setup? Toggle
Has anyone figured out to toggle between two nvidea Gpus , GPU0, GPU1 More on forum.comfy.org
🌐 forum.comfy.org
0
August 4, 2025
Anyway to run on dual GPU?
You can spread specific model components across multiple GPUs - so you can load the text encoder into the VRAM of one GPU and the unet model into the VRAM of another machine. Here take a look at this: https://github.com/pollockjj/ComfyUI-MultiGPU More on reddit.com
🌐 r/comfyui
4
2
February 10, 2025
🌐
RunComfy
runcomfy.com › comfyui-nodes › ComfyUI-MultiGPU
ComfyUI-MultiGPU detailed guide | ComfyUI
April 17, 2025 - At its core, ComfyUI-MultiGPU enhances memory management rather than parallel processing. This means that while the steps in your workflow still execute one after the other, the extension allows different components of your models to be loaded ...
🌐
Medium
medium.com › @xditproject › supercharge-your-aigc-experience-leverage-xdit-for-multiple-gpu-parallel-in-comfyui-flux-1-54b34e4bca05
Supercharge Your AIGC Experience: Leverage xDiT for Multiple GPU Parallel in ComfyUI Flux.1 Workflow | by xdit-project | Medium
October 14, 2024 - Without compromising the quality of image generation, XDiT significantly accelerates the generation efficiency of the FLUX.1 workflow in ComfyUI by leveraging multi-GPU parallelism.
🌐
ComfyDeploy
comfydeploy.com
ComfyDeploy - ComfyUI for teams
Empower your team with A100s, H100s GPUs. Generate at the same time with auto-scaling. ... Define Inputs and outputs in ComfyUI workflows to create simple UIs.
Find elsewhere
🌐
GitHub
github.com › neuratech-ai › ComfyUI-MultiGPU
GitHub - neuratech-ai/ComfyUI-MultiGPU: Rudimentary support for using multiple GPUs in a ComfyUI workflow
This extension adds new nodes for model loading that allow you to specify the GPU to use for each model. It monkey patches the memory management of ComfyUI in a hacky way and is neither a comprehensive solution nor a well-tested one.
Starred by 314 users
Forked by 27 users
Languages   Python
🌐
ComfyAI.run
comfyai.run › custom_node › ComfyUI-MultiGPU
ComfyUI-MultiGPU Custom Node
ComfyUI-MultiGPU is a powerful tool for optimizing VRAM usage by leveraging additional GPUs or system resources, allowing users to make the most out of their hardware. Instead of parallel processing, it distributes workload components across different devices, ensuring efficient memory management ...
🌐
Reddit
reddit.com › r/stablediffusion › made a comfyui extension for using multiple gpus in a workflow
r/StableDiffusion on Reddit: Made a ComfyUI extension for using multiple GPUs in a workflow
August 4, 2024 - However, would AI even be able to utilize two GPU:s? ... XDiT finally release their ComfyUI node for Parallel Multi GPU worker.
🌐
ComfyAI
comfyai.run › documentation › DeviceSelectorMultiGPU
Device Selector Multi GPU ComfyUI Node
Imagine you have a task that requires rendering complex graphics or processing data that can significantly benefit from GPU acceleration. The Device Selector Multi GPU node aids by allowing you to choose which GPU to run the process on, potentially speeding up the task and optimizing resource ...
🌐
ComfyUI
forum.comfy.org › desktop feedback
Multiple GPU Setup? Toggle - Desktop Feedback - ComfyUI
August 4, 2025 - Has anyone figured out to toggle between two nvidea Gpus , GPU0, GPU1
🌐
HOSTKEY
hostkey.com › apps › machine-learning › comfyui
ComfyUI Multi-GPU Hosting – NVIDIA and AMD | HOSTKEY
2 weeks ago - Run ComfyUI on multiple GPUs for faster Stable Diffusion workflows. NVIDIA & AMD options, quick activation, 24/7 support, global locations.
🌐
NVIDIA Blog
blogs.nvidia.com › home › get started using generative ai for content creation with comfyui and nvidia rtx ai pcs
Get Started Using Generative AI for Content Creation With ComfyUI and NVIDIA RTX AI PCs
September 9, 2025 - ComfyUI — an open-source, node-based graphical interface for running and building generative AI workflows for content creation — published major updates this past month, including up to 40% performance improvements for NVIDIA RTX GPUs, and ...
🌐
Comfy
comfy.icu › node › CLIPLoaderMultiGPU
CLIPLoaderMultiGPU - ComfyUI Cloud
December 18, 2024 - This extension adds CUDA device selection to supported loader nodes in ComfyUI. By monkey-patching ComfyUI’s memory management, each model component (like UNet, Clip, or VAE) can be loaded on a specific GPU. Examples included are multi-GPU workflows for SDXL, FLUX, LTXVideo, and Hunyuan Video ...
🌐
OpenArt
openart.ai › workflows › zentrocdot › flux-full-multi-gpu-gguf-workflow › RXSpq1w087JY2IfH1pTx
Flux Full Multi-GPU GGUF Workflow | ComfyUI Workflow | OpenArt
Downscaling is also possible. This workflow offers now full multi-GPU support. To run it on under-dimensioned computer systems it is using GGUF. LoRA can be used if needed. The images are saved as JPEG image to save space.
🌐
ComfyUI
comfy.icu › extension › pollockjj__ComfyUI-MultiGPU
ComfyUI-MultiGPU - ComfyUI Cloud
This extension adds CUDA device selection to supported loader nodes in ComfyUI. By monkey-patching ComfyUI’s memory management, each model component (like UNet, Clip, or VAE) can be loaded on a specific GPU. Examples included are multi-GPU workflows for SDXL, FLUX, LTXVideo, and Hunyuan Video ...
🌐
GitHub
github.com › robertvoy › ComfyUI-Distributed
GitHub - robertvoy/ComfyUI-Distributed: ComfyUI extension that enables multi-GPU processing locally, remotely and in the cloud
A powerful extension for ComfyUI that enables distributed and parallel processing across multiple GPUs and machines.
Starred by 442 users
Forked by 44 users
Languages   Python 59.0% | JavaScript 41.0%
🌐
YouTube
youtube.com › watch
Stable Swarm UI - GPU-Network Rendering
Enjoy the videos and music you love, upload original content, and share it all with friends, family, and the world on YouTube.