What are peoples' current suggestions for using Claude Code with a locally hosted LLM running on regular consumer hardware (for the sake of discussion, assume you're spending $US500-ish on a mini PC, which would get you a reasonably decent CPU, 32Gb RAM and a cheapish GPU)?
I get that it's not going to work as well as hosted/subscription services like Claude/Gemini/Codex/..., but sometimes those aren't an option
You can run something like Qwen 2.5 Coder on a regular machine (https://huggingface.co/Qwen/Qwen2.5-Coder-7B) but it's really not in the same universe with Claude Code, it will be slow and generate bad code.
It might make sense to run a small LLM locally for general conversation or very specific tasks, but they're not a serious option for agentic coding.
I get that it's not going to work as well as hosted/subscription services like Claude/Gemini/Codex/..., but sometimes those aren't an option