Is there any development environment or plugin that you're using for local LLM?
The result is pretty good in my opinion, probably depends on your development cases.
For running inference... Ollama is the easiest, but generally the slowest. Llama.cpp is a tad more involved, but faster. vLLM is even more involved, but even faster. Sglang can be difficult and sometimes challenging, but tends to be the fastest.