Local AI agents with smolagents and llama.cpp
I’ve been using llama.cpp on this box for
RAG; this post is the same
engine pointed at a different problem — running an actual agent that
plans, writes Python, and uses tools, entirely against a local model.
The whole thing runs in a rootless Docker container talking to
llama-server over --network host, which turns out to collapse a
surprising amount of the safety story that the smolagents docs spend
real effort on.