v2.0.7

September 18, 2025

Run agents with local and self-hosted models using Llama CPP

Agno now supports Llama CPP as a first-class model option, enabling teams to run agents on local or self-hosted LLMs. This expands deployment flexibility for organizations that need tighter control over cost, latency, data residency, or infrastructure.

With native Llama CPP support, teams can build and operate agentic systems without relying exclusively on hosted model providers. This makes Agno a stronger fit for on-prem, air-gapped, or cost-sensitive environments—without changing how agents are designed or managed.

Why this matters:

  • Greater control over cost and infrastructure
  • Support for privacy-sensitive or regulated deployments
  • More options for production model strategy