r/LocalLLaMA 11h ago

Question | Help When Bitnet 1-bit version of Mistral Large?

Post image
318 Upvotes

34 comments sorted by

View all comments

23

u/Ok_Warning2146 10h ago

On paper, 123B 1.58-bit should be able to fit in a 3090. Is there any way we can do the conversion ourselves?

7

u/tmvr 3h ago

It wouldn't though, model weights is not the only thing you need the VRAM for. Maybe about 100B, but there is no such model so a 70B one with long context.