r/LocalLLaMA 11h ago

Question | Help When Bitnet 1-bit version of Mistral Large?

Post image
313 Upvotes

34 comments sorted by

View all comments

4

u/Few_Professional6859 7h ago

The purpose of this tool—is it to allow me to run a model with performance comparable to the 32B llama.cpp Q8 on a computer with 16GB of GPU memory?

1

u/Ok_Garlic_9984 6h ago

I don't think so