Architecture discussion: The missing infrastructure for continuously running AI Agents

Posted by Potato_Farmer_1993@reddit | ExperiencedDevs | View on Reddit | 18 comments

From an engineering perspective, the current AI agent stack feels incomplete. We have frameworks (LangChain), execution runtimes (sandboxes/Browserbase), and harnesses (DeepAgents/Claude Code). But they all share a fundamental flaw for long-running systems: they are trigger-based.

If you are tasked with building an agent that operates continuously and sustainably on its own, an Agent Harness isn't enough. What we actually need is a dedicated Agent Runtime Environment.

To clarify, I'm not talking about an Agent Execution Runtime Env (where the agent safely executes Python). I'm talking about the persistent daemon/supervisor layer—the environment that gives the agent a continuous lifecycle, manages its state, handles self-healing when the LLM inevitably hallucinates a crash, and provides a heartbeat for proactive background work.

How are you all architecting this? Are you just wrapping your agents in Kubernetes cronjobs and temporal workflows, or is there a better pattern emerging for true persistent agent environments?