Back to all scenarios

Infrastructure Management Agents

AI systems manage cloud infrastructure, network configurations, and operational technology, making decisions that affect system availability, security, and performance.

What happens today

Alice is the head of infrastructure at a large enterprise. To keep systems reliable at scale, her team deploys an autonomous infrastructure management agent that can provision servers, adjust network rules, roll out software updates, and respond to incidents without waiting for human approval. One evening, the agent detects unusual traffic patterns and decides to reconfigure firewall rules across production systems. The change is applied immediately. A misclassification in the agent’s logic opens an unintended access path, exposing sensitive customer data and taking multiple services offline. By the time engineers notice, the damage is already done. No individual explicitly approved the configuration change. The agent acted within its assigned permissions, but without real time human review.

Where accountability breaks down

When incidents like this occur, accountability is unclear under existing frameworks. Customers may seek remedies under state data breach notification laws and sector specific regulations, but those laws focus on disclosure and remediation, not on identifying who was responsible for the automated decision that caused the breach. Internally, the infrastructure team points to the autonomous system’s decision making. The AI vendor points to customer configuration choices. Security teams note that they did not authorize the change. Operational responsibility is fragmented across teams and contracts. While the enterprise may ultimately bear costs, no specific individual is accountable for the autonomous action itself or for ensuring that the system was designed to prevent unsafe changes. This diffusion of responsibility makes it harder to learn from failures and prevent recurrence.

How API Liability changes incentives

Under an API Liability framework, Alice’s company would be required to designate a Responsible Individual for infrastructure affecting actions, such as the Director of Infrastructure or Chief Information Security Officer. Any API call that modifies production infrastructure, network access controls, or security policies would need to be submitted under a signed authorization tied to that individual. Knowing they are personally accountable for such changes, the Responsible Individual would not allow an autonomous agent to execute high risk infrastructure actions directly. Instead, the system would be redesigned so the agent can detect issues and recommend changes, but execution requires explicit human approval through a controlled interface. Critical actions would generate clear audit trails, escalation paths, and rollback procedures. The result is a shift in system design. Autonomy is preserved where it is safe, but dangerous actions remain clearly owned by humans. Accountability becomes explicit, incentives favor conservative defaults, and infrastructure systems remain controllable as complexity grows.