You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
feat(distorch): Add universal .safetensors support & memory-based distribution
This commit introduces DisTorch v2.0.0, a major overhaul that extends multi-device model distribution to standard `.safetensors` models.
Key changes include:
- **Universal `.safetensors` Support:** The core distribution logic is no longer limited to GGUF models. It now fully supports `.safetensors`, allowing any UNet supported by native Comfy loaders to have its layers distributed across multiple devices (GPUs and CPU/RAM).
Copy file name to clipboardExpand all lines: README.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,4 +1,4 @@
1
-
# ComfyUI-MultiGPU: Tools to free up your primary GPU’s VRAM by using your CPU or additional GPUs, now with tighter integration into kijai's WanVideoWrapper[^1]
1
+
# ComfyUI-MultiGPU v2.0.0: Universal `.safetensors` and GGUF Multi-GPU Distribution with DisTorch
@@ -18,10 +18,10 @@ ComfyUI-MultiGPU now includes a custom, tightly integrated implementation for Wa
18
18
## The Core of ComfyUI-MultiGPU:
19
19
[^1]: This **enhances memory management,** not parallel processing. Workflow steps still execute sequentially, but with components (in full or in part) loaded across your specified devices. *Performance gains* come from avoiding repeated model loading/unloading when VRAM is constrained. *Capability gains* come from offloading as much of the model (VAE/CLIP/UNet) off of your main **compute** device as possible—allowing you to maximize latent space for actual computation.
20
20
21
-
1.**DisTorch Virtual VRAM for UNet Loaders**: Move UNet layers off your compute GPU
22
-
- Automaticdistribution to RAM or other GPUs
23
-
- One-number control of VRAM usage
24
-
-Support for all GGUF models
21
+
1.**DisTorch Virtual VRAM for `.safetensors` and GGUF Models**: Move model layers off your compute GPU
22
+
- Automatic, memory-size based distribution to RAM or other GPUs
23
+
- One-number control of VRAM usage
24
+
-Universal support for `.safetensors` and GGUF models
25
25
26
26
2.**CLIP Offloading**: Two solutions for LLM-based and standard CLIP models:
27
27
-**MultiGPU CLIP**: Full offload to CPU or secondary GPU
@@ -83,7 +83,7 @@ With a 12GB GPU running an 8GB model:
83
83
- Your GPU now has extra VRAM for larger batches, higher resolutions, or longer video
84
84
85
85
## 🚀 Compatibility
86
-
Works with all GGUF-quantized ComfyUI/ComfyUI-GGUF-supported UNet/CLIP models.
86
+
Works with all `.safetensors` and GGUF-quantized models.
87
87
88
88
⚙️ Expert users: For those of you who were here for the 1.0 release of DisTorch, manual allocation strings still available for advanced configurations. Each log will contain the allocation string for the run so it can be easily recreated and/or manipulated for more sophisticated setups.
logging.info(f"\nWarning: Model size is greater than 90% of recipient VRAM. {on_virtuals:.2f} GB of GGML Layers Offloaded Automatically to Virtual VRAM.\n")
267
+
logger.info(f"\nWarning: Model size is greater than 90% of recipient VRAM. {on_virtuals:.2f} GB of GGML Layers Offloaded Automatically to Virtual VRAM.\n")
0 commit comments