Replies: 1 comment 2 replies
-
Hello, thank you for the feedback! Most tasks seem manageable with a 14B parameter model. I’ve done all the development so far by running the ollama_server.py script on my gaming machine (RTX 3060 with 12GB VRAM, 16GB RAM) and connecting via SSH from my MacBook, and it’s worked quite well. But you raise an important issue deepseek 7B does perform poorly and agenticSeek load other hugging face model (router, tts, stt, summarization) which can crash a lot of users. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Nice work, but running full deepseek r1 will not be possible for 90-95%+ users so I wanted to know if you have tested some great local llm models under 32B param which can perform small to mid level task easily without much issues ?
It will be best if you can tell both reasoning and non reasoning ones because we don't need unnecessary reasoning for straightforward tasks.
Beta Was this translation helpful? Give feedback.
All reactions