Open-Source Tool 'Sage' Intercepts AI Agent Commands Before They Hit Your OS
Sage, a new open-source security tool, blocks rogue AI agent actions before they reach the operating system, creating a new class of protection called Agent Detection & Response.

A new open-source project called Sage has emerged to address one of the quieter security gaps in the AI tooling boom: autonomous coding agents that execute shell commands, fetch URLs, and write files on developer machines with virtually no oversight.
Sage inserts an interception layer between an AI agent and those operations, checking each action before it proceeds, according to reporting by Help Net Security. The project targets a growing class of developer tools, including Claude Code, Cursor/VS Code, and a platform called OpenClaw, working through hook systems native to each supported agent to catch potentially harmful actions at the point of execution.
The timing reflects a real and widening risk. As AI coding assistants have grown more capable, they have also grown more autonomous, able to run multi-step tasks that touch the file system, install packages, and make outbound network calls with little or no inspection of what they are doing. A compromised or manipulated agent on a developer workstation could cause significant damage before any human notices.
Sage applies the term Agent Detection & Response, or ADR, to this class of tooling. The name is a deliberate parallel to the endpoint detection and response category that has been standard in enterprise security for over a decade, drawing a direct line between the problem of unchecked agent behavior and the established challenge of unchecked endpoint activity.
Each intercepted action passes through several detection layers. URL reputation checking runs cloud-based malware, phishing, and scam detection. Local heuristics use YAML-based threat definitions for dangerous patterns. Package supply-chain checks cover registry existence, file reputation, and age analysis for npm and PyPI packages. Plugin scanning runs at session start and checks other installed plugins for threats.

The privacy model keeps most data on the local machine. Sage sends URL hashes and package hashes to Gen Digital reputation APIs, but file content, commands, and source code stay local. Both cloud services can be disabled for fully offline operation, a provision likely to matter for enterprise teams with strict data residency requirements.
Several practical details remain undisclosed in the initial reporting. The project's maintainers have not been publicly identified, no repository link or installation instructions have been published through the outlets that covered the launch, and the supported operating systems have not been specified beyond a generic reference to the underlying OS. Independent performance benchmarks, latency overhead from real-time interception, and any analysis of potential bypass techniques are also absent from the available materials.
Those gaps matter. The ADR concept is compelling, particularly as organizations begin deploying AI agents on sensitive internal infrastructure, but the security value of any interception layer depends heavily on the robustness of its hook implementation, the quality and governance of its threat definitions, and whether the interception itself can be evaded. The YAML-based heuristics, for instance, are only as strong as whoever authors and maintains them.
Security teams evaluating Sage will want answers to those questions before treating it as a production-grade control. For now, the project represents an early attempt to bring the detection-and-response discipline that enterprises have applied to endpoints into a domain, autonomous AI agents, that has so far operated largely without it.
Sources:
Know something we missed? Have a correction or additional information?
Submit a Tip

