I'm looking for hardware recommendations for running LLMs locally. Specifically, best bang for bucks at the following price points: $2,000 $5,000 $10,000
What more could you do with a 10k budget than with a 5k budget?
Thank you
The most important part to run LLMs might be the VRAM-size. Since modern Macs use unified memory you simplified get RAM=VRAM, so a 64GB MacBook Pro might outperform an NVidia 4090 in LLMs.
However the specialized cards from nvidia have more VRAM and are generally faster.
The LLM of your choice should fit into vram, so if you only need 21gb, one 4090 will do the job
On a different computer running Cinnamon Mint I used GPT4ALL to load a Llama 3 8b model. I use it to read a bunch of local docs for me. It also didn't cost anything.