Solo.io Open-Sources 'agentevals' at KubeCon — Fixing Production AI Agent Reliability
One of the persistent frustrations with AI agents in production is that nobody agrees on how to know if they’re working correctly. Solo.io is taking a shot at solving that with agentevals, an open-source project launched at KubeCon + CloudNativeCon Europe 2026 in Amsterdam. The premise is straightforward but the execution is non-trivial: continuously score your agents’ behavior against defined benchmarks, using your existing observability data, across any LLM model or framework. Not a one-time evaluation. Not a test suite that only runs before deployment. A live, ongoing signal. ...