llama.cpp Adds Anthropic API Support, Rendering Cloud API Lock-In Obsolete
The local inference engine’s native Anthropic Messages API support lets you run Claude Code with local models, collapsing the wall between commercial and private AI workflows.