The Glass Box Imperative: Governing intent-based automation in AI-Native networks
Connectivity is foundational infrastructure, and the bar for reliability continues to rise. In our previous post, we described how AI is changing network demand, making traffic more interactive, bursty, and non-uniform. As AI experiences evolve, this variability will not stabilize.
Networks for the AI supercycle need autonomy
This shift has a direct operational implication: networks can no longer rely on static planning cycles. Adaptation must become continuous. The network must sense, decide, and execute adjustments through closed loops without per-event human approval, making it truly autonomous. This is unavoidable for three reasons:
- Pace: Change is faster than human operational cycles.
- Scale: Modern network complexity makes manual intervention infeasible.
- Coupling: Effects span domains; siloed interventions can amplify instability rather than contain it.
While network autonomy has been discussed for years, the foundation was missing: a way to coordinate decisions end-to-end and govern behavior so it remains inspectable, bounded, and reversible. As autonomy becomes unavoidable, we must close these gaps with urgency. Network operationalization was already getting harder; autonomy multiplies interacting control loops, shortens decision timescales, and increases the blast radius of mistakes.
Autonomy cannot be siloed
Autonomy cannot be a collection of independent domain optimizers. If the RAN, core, IP access and transport deploy isolated AI loops acting on shared state, latency, capacity, and energy, their interactions can produce globally unstable behavior like oscillations and priority inversions. In a coupled system, the problem shifts from optimizing individual functions to governing end-to-end outcomes.
Safe autonomy requires system-level coordination to arbitrate conflicts and ensure intent is translated consistently across domains. Without it, autonomy fragments into structural risk.
Trustworthy autonomy demands transparency
End-to-end coordination means decisions are no longer local; one system can drive changes across domains to meet a service outcome. This increases the impact of mistakes. For such power to be deployable, operators must see what changed and why, know which models were active, limit the rate of change, and roll back safely.
Consider congestion in the transport network. If traffic is rerouted, the RAN and core may each react to protect performance, changing the pattern again and causing the transport to reroute once more. The network ends up “chasing its tail.” Coordinated autonomy breaks this cycle by reconciling objectives across domains so the network converges instead of oscillating.
The Glass Box imperative
To be deployable at carrier scale, autonomy must remain inspectable, bounded, and reversible. It must explain its actions, operate within explicit safety limits, and support rollback as a standard operational primitive. We refer to this as the Glass Box.
Two broader realities strengthen this imperative:
- Shift to agentic automation: Across industries, organizations are moving from assistive AI to agents that execute multi-step workflows. In networks, agentic automation is the natural model for continuous adaptation, but it demands a higher bar for accountability.
- The intelligence gravity well: The primary source of lock-in is shifting from hardware to the intelligence layer, the models, policies, and context graphs. Monolithic, proprietary platforms risk trading short-term gains for long-term dependency.
In brownfield, autonomy will not arrive via forklift replacement. Multi-generational networks already run a mix of domain controllers, policy systems, and operational processes, often from multiple vendors. The practical path is incremental: start with narrow, high-value closed loops, run them first in advisory mode, then allow bounded actuation only after observability, provenance, and rollback have been proven in live operations. Scope expands as trust is earned, not as architecture diagrams mature.
In practice, this leads to a hybrid placement of agents within the network architecture. A fully centralized “god agent” is brittle in scale and slow at fast timescales, while purely distributed agents can drift into conflict without arbitration. The most workable model is hierarchical or hybrid: domain-local agents handle fast, local decisions within tight bounds, while an end-to-end governance layer coordinates objectives, resolves conflicts, and controls permissions and rollout across domains. The key design principle is logically unified coordination with physically distributed execution.
The architecture: intent, agents, governance, and an operating fabric
The architectural question is how to structure intelligence so autonomy scales end-to-end without creating instability, opacity, or lock-in. One way to frame this is as an operating fabric that connects four things: what the operator wants, what the system knows, what the system is allowed to do, and how it proves it did the right thing. Those elements turn autonomy from a collection of optimizers into a governable system.
In this context, an agent is software that translates intent into action by discovering context and invoking operational tools under explicit permissions and safety limits. Not all agents should act. A useful taxonomy is to distinguish observer agents (detect and summarize), advisor agents (recommend), actuator agents (execute bounded changes), coordinator agents (support cross-domain arbitration), and lifecycle agents (enforce validation gates, rollout, rollback, and audit). With that definition in mind, the four tenets below describe how intent, agents, governance, and actuation fit together as a system.
- Intent is the contract: We must move from “tuning faster” to expressing outcomes, latency, reliability, and energy tradeoffs. Intent becomes the shared language that removes inter-domain guesswork.
- Agentic automation is the execution model: Agents translate intent into coordinated action sequences. They are not free-form optimizers; they must be authorized for specific actions and constrained by policy.
- AI Governance is the coordination plane: This distributed coordination layer arbitrates conflicts and enforces global policy, ensuring local logic does not override service intent.
- Bounded actuation is the operational safety: Autonomy must be designed with explicit limits on scope, rate of change, and blast radius to prevent cascading instability.
The destination state is a unified operating fabric, where automation reasons about coupled outcomes rather than local KPIs.
The Glass Box requirement: autonomy we can inspect
In practice, glass box autonomy rests on five properties:
- Observability: What decision was made, where, and when.
- Provenance: Which model or policy version was active.
- Traceability: The signals and context that drove the action.
- Safety Controls: Guardrails, bounded actuation, and explicit rollback.
- Auditability: Clear accountability for every intervention.
Together, these properties make autonomy responsible, secure, and trustworthy by design. Explainability means having an actionable decision record for incidents, postmortems, and regulatory review. A system that cannot explain its actions cannot be trusted to act again.
CI/CD as lifecycle control
In an autonomous network, CI/CD must govern not only code but also models, policies, and agent workflows. Autonomy also runs into a hard constraint: data. Much of the most useful network data is distributed across domains, inconsistently labeled, and subject to privacy, residency, and commercial boundaries. That reality shapes the implementation: learning and tuning often must happen where the data lives, with governed sharing of models, features, and summaries rather than raw data pooling. Glass box governance must therefore cover data lineage and access as well as model and policy lineage, because what the system learned from, and what it was allowed to see, directly affects whether its actions are explainable and auditable.
Because behaviors are coupled, validation and rollback must be end-to-end. AIOps provides the sensing layer to detect model drift or regressions, while governance defines the safety envelope for recovery.
Programmability and openness
The new frontier of lock-in is the intelligence layer: the context discovery, tool interfaces, and agent workflows. Open, governed interfaces are what keep agentic automation portable and allow innovation at software speed without recreating proprietary control planes. This is why standards-based openness across O-RAN, 3GPP, TM Forum, and the AI-RAN Alliance matters, and why we should avoid binding autonomy to a single proprietary agent context interface.
Conclusion: autonomy Is inevitable, instability Is unacceptable
AI is pushing intelligence into every layer of the network. The glass box is the line that holds, the only way to scale autonomy without sacrificing the determinism and trust required for essential infrastructure. The only real risk is moving too slowly while autonomy arrives faster than our ability to govern it.