Local LLM Configuration

From my testing, local “Distilled” models of DeepSeek-R1 don’t work with Artifacts (the bolt terminal and etc.) because they are not “Instruct” fine-tuned. You’ll want to try a different model. If you have a massive hardware constraints (or want to run an LLM on an RPI for example), you can try the smallest “Instruct” model (3B Parameters) I have seen work with bolt (and useable):

And for the size, the current best 7B Parameter LLM on the Hugging Face Open LLM Leader Board:

Rank Type Model Average IFEval BBH MATH GPQA MURS MMLU-Pro CO2 Cost
#708 prithivMLmods QwQ-LCoT2-7B-Instruct 28.57% 55.61% 34.37% 22.21% 6.38% 15.75% 37.13% 1.37 kg

You can install them for Ollama with:
ollama run hf.co/mradermacher/QwQ-LCoT-3B-Instruct-GGUF:Q4_K_M
ollama run hf.co/mradermacher/QwQ-LCoT2-7B-Instruct-GGUF:Q4_K_M

P.S. I do personally plan on creating a “Instruct” fine-tune set of the DeepSeek-R1 “Distilled” (1.5B - 32B) models to support Bolt.diy, but I’m still figuring out a few things and I will release them on both Hugging Face and Ollama. Maybe also write a how-to doc.

Lastly, I’d also go through your Bolt.diy settings to turn off all the extra stuff (like providers) that you aren’t going to use, and enable optimized prompts, etc.

Hope that helps!

2 Likes