As you can imagine, maintaining entropy for all these disparate systems takes time, of which I have less now that I have young kids. So when Claude Code was released, I took to it like a fish to water. We mapped my entire network, I created accounts on all the devices so it can SSH into everything and configure everything (including the Ubiquiti Dream Machine Pro!). I have been blow away at how well it troubleshoots and fixes everything.
I have a DGX Spark AI workstation (128gb of memory), and I really want to now hand off the work to a local model, either using Opencode or Claude Code harnesses and simply pointing at a vLLM instantiated model accessable by API (just point Opencode or Claude Code at the local IP and API endpoint).
It works, except I tried Qwen3-coder just now and it's refusing to help due to security concerns. Ugh. I then tried GLM-4.7-Flash, but vLLM doesn't support yet and so before I rebuild (ask Claude Code to rebuild and deploy) to try GLM.4-7-Flash with some other inference provider, does anyone have a model they use for infrastructure maintenance that isn't a little bitch? I will probably eventually go to an abliterated model if none of the open source ones will help.
1 comments