SwarmUI provides a UI that can handle multiple ComfyUI instances as backends at once.
https://github.com/mcmonkeyprojects/SwarmUI

Currently, ComfyUI does not provide a method to execute workflows in parallel.

If you are a developer and want to implement inference functionality for multiple GPUs, I think modifying the KSampler would be the most effective approach.

If I had a multiple GPU environment, I would like to experiment with this, but I'm not sure if PyTorch can properly handle this scenario.

It's important to note that several custom nodes use implementations that hijack the Sampling function. Your modifications might cause issues with these nodes.

🌐
Reddit
reddit.com › r/comfyui › sanity check: using multiple gpus in one pc via comfyui-multigpu. will it be a benefit?
r/comfyui on Reddit: Sanity check: Using multiple GPUs in one PC via ComfyUI-MultiGPU. Will it be a benefit?
April 22, 2025 -

I have a potentially bad idea, but I wanted to get all of your expertise to make sure I'm not going down a fruitless rabbit hole.

TLDR: I have a one PC with a 4070 12gb and one PC with a 3060 12gb. I run AI on both separately. I purchased a 5060 Ti 16gb.

My crazy idea is to get a new motherboard that will hold 2 graphics cards and use ComfyUI-MultiGPU to set up one of the PCs to run two GPUs (Most likely the 4070 12gb and 3060 12gb) and allow it to offload some things from the VRAM of the first GPU to the second GPU.

From what I've read in the ComfyUI-MultiGPU info it doesn't allow for things like processing on both GPUs at the same time, only swapping things from the memory of one GPU to the other.

It seems (and this is where I could be mistaken) that while this wouldn't give me the equivalent of 24GB of VRAM it might allow for things like GGUF swaps onto and off of the GPU and allow the usage of models over 12GB in the right circumstances.

The multi-GPU motherboards I am looking at are around $170-$200 or so and I figured I'd swap everything else from my old motherboard.

Has anyone had experience with a set up like this and was it worth it? did it help in enough cases that it was a benefit?

As it is I run two pcs and this allows me to do separate things simultaneously.

However, with many things like GGUF and block swapping allowing things to be run on cards with 12GB this might be a bit of a wild goose chance.

What would the biggest benefit of a set up like this be if any?

🌐
GitHub
github.com › pollockjj › ComfyUI-MultiGPU
GitHub - pollockjj/ComfyUI-MultiGPU: This custom_node for ComfyUI adds one-click "Virtual VRAM" for any UNet and CLIP loader as well MultiGPU integration in WanVideoWrapper, managing the offload/Block Swap of layers to DRAM *or* VRAM to maximize the latent space of your card. Also includes nodes for directly loading entire components (UNet, CLIP, VAE) onto the device you choose
This custom_node for ComfyUI adds one-click "Virtual VRAM" for any UNet and CLIP loader as well MultiGPU integration in WanVideoWrapper, managing the offload/Block Swap of layers to DRAM *or* VRAM to maximize the latent space of your card. Also includes nodes for directly loading entire components (UNet, CLIP, VAE) onto the device you choose - pollockjj/ComfyUI-MultiGPU
Starred by 777 users
Forked by 61 users
Languages   Python 99.6% | Shell 0.4%
Discussions

Multi-GPU Support
Hi! I'm interested in using ComfyUI with multiple GPUs for both training and inference. Specifically, I'm planning to utilize H100s. However, I haven't been able to find any specific op... More on github.com
🌐 github.com
10
9
July 30, 2024
Made a ComfyUI extension for using multiple GPUs in a workflow
I wanted to find out what it would take to add proper multi-GPU support to ComfyUI. While this is not it, these custom nodes will allow you to pick which GPU to run a given model on. This is useful if your workflow doesn't completely fit in VRAM on a single GPU. On my testing setup (2x 3090) there is a noticeable improvement when running flux dev by offloading the text encoders & VAE to the 2nd GPU. It's implemented in a very hacky but simple way and I'm surprised it even works. I saw some requests for this on the sub recently so hopefully it's useful to somebody. More on reddit.com
🌐 r/StableDiffusion
37
97
August 4, 2024
Multiple GPU Setup? Toggle
Has anyone figured out to toggle between two nvidea Gpus , GPU0, GPU1 More on forum.comfy.org
🌐 forum.comfy.org
0
August 4, 2025
Anyway to run on dual GPU?
You can spread specific model components across multiple GPUs - so you can load the text encoder into the VRAM of one GPU and the unet model into the VRAM of another machine. Here take a look at this: https://github.com/pollockjj/ComfyUI-MultiGPU More on reddit.com
🌐 r/comfyui
4
2
November 29, 2024
🌐
RunComfy
runcomfy.com › comfyui-nodes › ComfyUI-MultiGPU
ComfyUI-MultiGPU detailed guide | ComfyUI
April 17, 2025 - At its core, ComfyUI-MultiGPU enhances memory management rather than parallel processing. This means that while the steps in your workflow still execute one after the other, the extension allows different components of your models to be loaded ...
🌐
GitHub
github.com › neuratech-ai › ComfyUI-MultiGPU
GitHub - neuratech-ai/ComfyUI-MultiGPU: Rudimentary support for using multiple GPUs in a ComfyUI workflow
This extension adds new nodes for model loading that allow you to specify the GPU to use for each model. It monkey patches the memory management of ComfyUI in a hacky way and is neither a comprehensive solution nor a well-tested one.
Starred by 316 users
Forked by 27 users
Languages   Python
🌐
Medium
medium.com › @xditproject › supercharge-your-aigc-experience-leverage-xdit-for-multiple-gpu-parallel-in-comfyui-flux-1-54b34e4bca05
Supercharge Your AIGC Experience: Leverage xDiT for Multiple GPU Parallel in ComfyUI Flux.1 Workflow | by xdit-project | Medium
October 14, 2024 - ComfyUI, is the most popular web-based Diffusion Model interface optimized for workflow. Yet, its design for native single-GPU usage leaves it struggling with the demands of today’s large DiTs, resulting in unacceptably high latency for users ...
🌐
ComfyDeploy
comfydeploy.com
ComfyDeploy - ComfyUI for teams
Empower your team with A100s, H100s GPUs. Generate at the same time with auto-scaling. ... Define Inputs and outputs in ComfyUI workflows to create simple UIs.
Find elsewhere
🌐
GitHub
github.com › robertvoy › ComfyUI-Distributed
GitHub - robertvoy/ComfyUI-Distributed: ComfyUI extension that enables multi-GPU processing locally, remotely and in the cloud
2 weeks ago - A powerful extension for ComfyUI that enables distributed and parallel processing across multiple GPUs and machines.
Starred by 493 users
Forked by 47 users
Languages   Python 62.0% | JavaScript 38.0%
🌐
ComfyAI.run
comfyai.run › custom_node › ComfyUI-MultiGPU
ComfyUI-MultiGPU Custom Node
ComfyUI-MultiGPU is a powerful tool for optimizing VRAM usage by leveraging additional GPUs or system resources, allowing users to make the most out of their hardware. Instead of parallel processing, it distributes workload components across different devices, ensuring efficient memory management ...
🌐
ComfyAI
comfyai.run › documentation › DeviceSelectorMultiGPU
Device Selector Multi GPU ComfyUI Node
Imagine you have a task that requires rendering complex graphics or processing data that can significantly benefit from GPU acceleration. The Device Selector Multi GPU node aids by allowing you to choose which GPU to run the process on, potentially speeding up the task and optimizing resource ...
🌐
ComfyUI
forum.comfy.org › desktop feedback
Multiple GPU Setup? Toggle - Desktop Feedback - ComfyUI
August 4, 2025 - Has anyone figured out to toggle between two nvidea Gpus , GPU0, GPU1
🌐
ComfyUI
comfy.icu › extension › pollockjj__ComfyUI-MultiGPU
ComfyUI-MultiGPU - ComfyUI Cloud
This extension adds CUDA device selection to supported loader nodes in ComfyUI. By monkey-patching ComfyUI’s memory management, each model component (like UNet, Clip, or VAE) can be loaded on a specific GPU. Examples included are multi-GPU workflows for SDXL, FLUX, LTXVideo, and Hunyuan Video ...
🌐
HOSTKEY
hostkey.com › apps › machine-learning › comfyui
ComfyUI Multi-GPU Hosting – NVIDIA and AMD | HOSTKEY
2 weeks ago - Run ComfyUI on multiple GPUs for faster Stable Diffusion workflows. NVIDIA & AMD options, quick activation, 24/7 support, global locations.
🌐
NVIDIA Blog
blogs.nvidia.com › home › get started using generative ai for content creation with comfyui and nvidia rtx ai pcs
Get Started Using Generative AI for Content Creation With ComfyUI and NVIDIA RTX AI PCs
September 9, 2025 - ComfyUI — an open-source, node-based graphical interface for running and building generative AI workflows for content creation — published major updates this past month, including up to 40% performance improvements for NVIDIA RTX GPUs, and ...
🌐
Comfy
comfy.icu › node › CLIPLoaderMultiGPU
CLIPLoaderMultiGPU - ComfyUI Cloud
December 18, 2024 - This extension adds CUDA device selection to supported loader nodes in ComfyUI. By monkey-patching ComfyUI’s memory management, each model component (like UNet, Clip, or VAE) can be loaded on a specific GPU. Examples included are multi-GPU workflows for SDXL, FLUX, LTXVideo, and Hunyuan Video ...
🌐
OpenArt
openart.ai › workflows › zentrocdot › flux-full-multi-gpu-gguf-workflow › RXSpq1w087JY2IfH1pTx
Flux Full Multi-GPU GGUF Workflow | ComfyUI Workflow | OpenArt
Created by: zentrocdot: With its settings the workflow generates square images, which can be upscaled. Downscaling is also possible. This workflow offers now full multi-GPU support. To run it on under-dimensioned computer systems it is using ...
🌐
YouTube
youtube.com › watch
- YouTube
Enjoy the videos and music you love, upload original content, and share it all with friends, family, and the world on YouTube.