You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+24-5Lines changed: 24 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,20 @@
1
-
# ComfyUI-MultiGPU: Tools to free up your primary GPU’s VRAM by using your CPU or additional GPUs[^1]
1
+
# ComfyUI-MultiGPU: Tools to free up your primary GPU’s VRAM by using your CPU or additional GPUs, now with tighter integration into kijai's WanVideoWrapper[^1]
<em>Free almost all of your GPU for what matters: Maximum latent space processing</em>
6
6
</p>
7
7
8
+
### WanVideoWrapper Integration
9
+
10
+
ComfyUI-MultiGPU now includes a custom, tightly integrated implementation for WanVideoWrapper, providing true multi-GPU support for advanced video generation workflows. Unlike the standard override nodes, these are bespoke wrappers that offer a more stable and feature-rich experience.
11
+
12
+
**Key Capabilities:**
13
+
14
+
***Load Models Anywhere:** Load different WanVideo models onto separate GPUs within the same workflow (e.g., Model 1 on `cuda:0`, Model 2 on `cuda:1`).
15
+
***Swap Blocks Anywhere:** Utilize any GPU as a target for offloading transformer blocks via the `WanVideoBlockSwapMultiGPU` node (e.g., swap to `cuda:2`).
16
+
***Reliable Device Tracking:** The device context is correctly managed even in complex dual-model workflows, ensuring the right model runs on the right GPU.
17
+
8
18
## The Core of ComfyUI-MultiGPU:
9
19
[^1]: This **enhances memory management,** not parallel processing. Workflow steps still execute sequentially, but with components (in full or in part) loaded across your specified devices. *Performance gains* come from avoiding repeated model loading/unloading when VRAM is constrained. *Capability gains* come from offloading as much of the model (VAE/CLIP/UNet) off of your main **compute** device as possible—allowing you to maximize latent space for actual computation.
10
20
@@ -125,9 +135,13 @@ Currently supported nodes (automatically detected if available):
This workflow extends Kijai's I2V example for WanVideo 2.2, demonstrating a dual-model setup. The HIGH model is loaded on `cuda:0`, the LOW model on `cuda:1`, with block-swapping offloaded to `cuda:2`.
140
159
141
160
### Split GGUF-quantized UNet and CLIP models across multiple devices using DisTorch
Copy file name to clipboardExpand all lines: pyproject.toml
+2-2Lines changed: 2 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
[project]
2
2
name = "comfyui-multigpu"
3
-
description = "This custom_node for ComfyUI adds one-click 'Virtual VRAM' for any GGUF UNet and CLIP loader, managing the offload of layers to DRAM or VRAM to maximize the latent space of your card. Also includes nodes for directly loading entire components (UNet, CLIP, VAE) onto the device you choose. Includes 16 examples covering common use cases. Includes support for kijai's ComfyUI-WanVideoWrapper and ComfyUI-HunyuanVideoWrapper, among other popular loaders."
4
-
version = "1.7.3"
3
+
description = "Adds full multi-GPU support for WanVideoWrapper, enabling model loading and block-swapping on any device. Provides a suite of custom nodes to manage multiple GPUs for ComfyUI, including advanced GGUF offloading with DisTorch and device overrides for core nodes."
0 commit comments