2025: The Year of Agents. 2026: The Year of Local Agents?
Relying on cloud-hosted LLMs is often overkill. While frontier models still lead in complex coding, local models are now more than capable of handling many agentic workflows—with zero latency and total privacy.
It provides minimal, high-performance building blocks for agents in C++, built directly around the awesome llama.cpp ecosystem. Stop sending your data to a remote API. Start building and running agents on your own hardware.
- 🤏Improvements for Small Language Models We recommend tinyagent to be used as default when working with Small Language Models. - 🧪 Extending and Improving test suite. For example, we now include a cookbook and an integration test running a Small Language Model (Qwen 1.7B) - 📖Extending and Improving docs