💥
CUDA Out Of Memory
It's a feature, NOT a bug.
Highlights
Pinned Loading
-
abliteration
abliteration PublicMake abliterated models with transformers, easy and fast
-
turboderp-org/exllamav2
turboderp-org/exllamav2 PublicA fast inference library for running LLMs locally on modern consumer-class GPUs
-
hiyouga/LlamaFactory
hiyouga/LlamaFactory PublicUnified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
-
Tiiny-AI/PowerInfer
Tiiny-AI/PowerInfer PublicHigh-speed Large Language Model Serving for Local Deployment
-
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.



