Skip to content

Commit ddd159e

Browse files
committed
docs: Clarify GGUF performance gain comparison in README
Update the README to specify that the "up to 10% faster GGUF inference" claim for DisTorch2 is a direct comparison against the previous DisTorch V1 implementation. This clarification helps manage user expectations and provides a more accurate performance context.
1 parent 9838f2c commit ddd159e

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
[^1]: This **enhances memory management,** not parallel processing. Workflow steps still execute sequentially, but with components (in full or in part) loaded across your specified devices. *Performance gains* come from avoiding repeated model loading/unloading when VRAM is constrained. *Capability gains* come from offloading as much of the model (VAE/CLIP/UNet) off of your main **compute** device as possible—allowing you to maximize latent space for actual computation.
1010

1111
1. **Universal .safetensors Support**: Native DisTorch2 distribution for all `.safetensors` models.
12-
2. **Up to 10% Faster GGUF Inference**: The new DisTorch2 logic provides potential speedups for GGUF models.
12+
2. **Up to 10% Faster GGUF Inference versus DisTorch1**: The new DisTorch2 logic provides potential speedups for GGUF models versus the DisTorch V1 method.
1313
3. **Bespoke WanVideoWrapper Integration**: Tightly integrated, stable support for WanVideoWrapper with eight bespoke MultiGPU nodes.
1414

1515
<h1 align="center">DisTorch: How It Works</h1>

0 commit comments

Comments
 (0)