a few others:
Awesome to see a quantized model released. I’ll have to test the 4 but version later… But I’m really waiting for are smaller and quantized models of DeepSeek v3.
bullerwins/DeepSeek-V3-GGUF
Q3, Q4, Q5, Q8, BF16
Awesome but damn, like 400GB even with 4bit quantization. No way I’ll be running that but still pretty cool. Might make the model more available and keep the cost down.
And I hadn’t even look in the last few days but these all seem pretty recent.
Thanks.
Yeah DeepSeek V3 is just too big haha
I wish they had a 32b/70b parameter version of it or something
Thanks for sharing Max! Looks awesome
Blockquote
Yeah DeepSeek V3 is just too big haha
I wish they had a 32b/70b parameter version of it or something
same here
the biggest problem imo is that LLMs are trained on too many programming languages, there should be a deepseek-v3-typescript, deepseek-v3-python, …
because i’m quite certain then it should be possible to get to a 70b or even 32b parameter size
Great opportunity for a game changer fine tuning project…