Can somebody define the maximum model size for llama/MLC?
I've tried a lot of models under 2Gb on llama, but when I try to use ~10Gb it fails to download.
I think there is limitation on react-native-blob-util, since I tried both downloading methods: via downloadModel and straight via react-native-blob-util.
Sometimes for >8Gb models isModelDownloaded returns true (but not consistent on different size models).
For example from https://huggingface.co/lmstudio-community/gemma-2-2b-it-GGUF/tree/main here's breakdown:
- (1.92Gb) gemma-2-2b-it-Q5_K_M.gguf -> The largest model I got successfully
- (2.15Gb) gemma-2-2b-it-Q6_K.gguf -> Model where the issue starts
Can somebody define the maximum model size for llama/MLC?
I've tried a lot of models under 2Gb on llama, but when I try to use ~10Gb it fails to download.
I think there is limitation on
react-native-blob-util, since I tried both downloading methods: viadownloadModeland straight viareact-native-blob-util.Sometimes for >8Gb models
isModelDownloadedreturns true (but not consistent on different size models).For example from https://huggingface.co/lmstudio-community/gemma-2-2b-it-GGUF/tree/main here's breakdown: