Intelligent Networks: Automating Operations with AI Agents

Author: Denis Avetisyan


A new approach leverages the power of artificial intelligence to streamline the complex tasks of managing and maintaining modern optical networks.

Optical network operations and maintenance have transitioned from labor-intensive manual procedures to increasingly autonomous, agentic workflows, reflecting an inevitable shift toward systems that navigate their own decline with greater efficiency.
Optical network operations and maintenance have transitioned from labor-intensive manual procedures to increasingly autonomous, agentic workflows, reflecting an inevitable shift toward systems that navigate their own decline with greater efficiency.

This review explores a multi-agent architecture utilizing large language models for automated fault management, performance optimization, and digital twin integration in optical network O&M.

The increasing complexity of optical networks presents a significant challenge to traditional operation and maintenance (O&M) approaches. This paper, ‘Large language models for optical network O&M: Agent-embedded workflow for automation’, proposes a multi-agent architecture leveraging large language models (LLMs) to address these challenges by automating key O&M tasks. Through agent-embedded workflows and technologies like prompt engineering, the framework facilitates improved executability and integration with existing tools for functions including fault management and performance optimization. Could this approach pave the way for fully autonomous, closed-loop optical network O&M systems capable of real-time adaptation and self-healing?


The Evolving Architecture of Modern Networks

Modern communication infrastructure is fundamentally built upon high-capacity optical networks, yet these networks are experiencing a surge in complexity. Driven by escalating bandwidth demands from data-intensive applications – including streaming video, cloud computing, and the Internet of Things – network operators are constantly increasing channel counts and adopting more sophisticated modulation formats. This progression, while boosting capacity, introduces significant challenges in areas like signal interference, nonlinear effects, and dynamic resource allocation. Furthermore, the deployment of flexible grid networks and coherent optical technologies adds layers of configuration and control, making traditional, static provisioning methods increasingly ineffective. Consequently, maintaining optimal performance and reliability within these rapidly evolving networks requires a paradigm shift towards more intelligent and automated management strategies.

Contemporary optical networks, while capable of transmitting vast amounts of data, are increasingly susceptible to performance issues due to the limitations of conventional management strategies. These established methods, often reactive and manually intensive, struggle to address the dynamic and multifaceted nature of modern network demands. As network complexity grows – with a proliferation of wavelengths, sophisticated modulation formats, and intricate routing protocols – traditional approaches become bottlenecks, hindering the ability to quickly identify and resolve issues. This results in predictable performance degradation, service disruptions for end-users, and increased operational costs for network providers, highlighting the urgent need for more intelligent and automated solutions capable of proactively maintaining optimal network health.

The escalating demand for bandwidth is driving a proliferation of optical channels within modern networks, creating a significant management challenge. Each channel, carrying data as light pulses, is susceptible to impairments like signal attenuation, dispersion, and non-linear effects, all of which degrade performance. Maintaining optimal signal transmission requires precise control over numerous parameters – wavelength, power levels, modulation formats – and a constant adaptation to fluctuating conditions. The sheer number of these channels, coupled with their intricate interdependencies, quickly overwhelms traditional, manual management approaches. Effectively coordinating these resources, identifying and mitigating impairments in real-time, and ensuring seamless data delivery across the network demands innovative solutions capable of handling this inherent complexity and unlocking the full potential of optical infrastructure.

Realizing the full capabilities of modern optical networks demands a shift towards proactive and intelligent management systems. This work introduces a novel framework designed to achieve closed-loop automation, enabling networks to dynamically adapt to changing conditions and optimize performance without manual intervention. By integrating real-time monitoring, predictive analytics, and automated control mechanisms, the proposed system anticipates potential issues – such as signal degradation or congestion – and preemptively adjusts network parameters to maintain optimal signal transmission. This approach moves beyond reactive troubleshooting, fostering a self-optimizing network infrastructure capable of delivering enhanced reliability, increased capacity, and improved quality of service, ultimately unlocking the true potential of these complex communication systems.

This three-layer optical network architecture integrates a physical layer of optical elements and fibers, a control layer managing network resources, and an application layer enabling digital twin functionality for operations, maintenance, and performance optimization.
This three-layer optical network architecture integrates a physical layer of optical elements and fibers, a control layer managing network resources, and an application layer enabling digital twin functionality for operations, maintenance, and performance optimization.

Decentralized Control: The Multi-Agent System Approach

Traditional network management systems typically rely on centralized control planes, creating single points of failure and scalability bottlenecks. These systems often struggle to adapt to dynamic network conditions and require significant manual intervention for configuration and troubleshooting. The proposed Multi-Agent System (MAS) architecture addresses these limitations by distributing control logic across multiple autonomous agents. Each agent operates independently, processing local information and making decisions related to its assigned network function. This distributed approach enhances system resilience, improves scalability to accommodate growing network complexity, and enables faster response times to changing network demands compared to centralized architectures.

The Multi-Agent System employs individual AI Agents, each dedicated to a discrete network function such as bandwidth allocation, intrusion detection, or quality of service monitoring. These agents operate autonomously, processing local network data and making decisions relevant to their assigned task without centralized command. This distributed architecture facilitates adaptability by allowing the system to respond to localized changes and failures without impacting overall network operation. The agents communicate and coordinate through defined protocols, sharing information and adjusting actions to optimize network performance and maintain stability. This decentralized decision-making process increases resilience and scalability compared to traditional, monolithic network management systems.

The Supervisor Agent functions as the central coordinating entity within the Multi-Agent System. It does not directly manage network resources, but instead receives status updates and performance metrics from individual AI Agents responsible for specific tasks. Based on this aggregated data, the Supervisor Agent dynamically adjusts agent priorities, allocates computational resources, and resolves conflicts between competing agent actions. This orchestration ensures that the collective actions of the AI Agents remain aligned with overall network objectives and that system-wide efficiency is maximized. The Supervisor Agent employs a rule-based system and utilizes pre-defined policies to determine optimal agent behavior and facilitate proactive network management.

Dynamic resource allocation within the Multi-Agent System is achieved by continuously monitoring network conditions and adjusting agent task assignments based on real-time demand and available resources. Proactive issue resolution is facilitated by agents autonomously identifying and mitigating potential problems before they impact network performance; this includes tasks like rerouting traffic around failing links or scaling resources in anticipation of peak loads. This combination of dynamic allocation and proactive resolution forms a critical feedback loop, enabling closed-loop autonomous network management and contributing to the overall intelligence of the system by reducing reliance on manual intervention and optimizing network behavior without explicit programming for every possible scenario.

An LLM-based multi-agent system iteratively refines responses by using a Supervisor Agent to coordinate specialized Sub-Agents and external tools like prompt templates, RAG, and APIs to fulfill user intent.
An LLM-based multi-agent system iteratively refines responses by using a Supervisor Agent to coordinate specialized Sub-Agents and external tools like prompt templates, RAG, and APIs to fulfill user intent.

Distributed Intelligence: Intelligent Fault Management

The fault management system utilizes a distributed artificial intelligence (AI) architecture composed of multiple AI Agents. These Agents operate independently, collecting and analyzing telemetry data from network devices. This data is then shared and correlated across the Agent network, creating a collective intelligence that exceeds the capabilities of any single Agent. This distributed approach enables faster fault detection, improved accuracy in diagnosis, and increased scalability compared to traditional, centralized fault management systems. The Agents employ machine learning algorithms to continuously refine their analysis and adapt to changing network conditions, resulting in a self-optimizing fault management solution.

Root Cause Analysis (RCA) within the fault management system utilizes data aggregation and pattern recognition to pinpoint the fundamental origin of network disruptions. This process moves beyond simple symptom remediation – such as restarting a service or replacing a failed component – by analyzing historical data, system logs, and real-time telemetry. The system correlates events across multiple network elements to identify causal relationships, determining the initiating factor that led to the observed issue. By addressing the root cause, the system prevents recurrence of the problem and reduces the overall frequency of incidents, rather than continuously reacting to downstream effects. This approach increases network stability and minimizes Mean Time To Repair (MTTR) through proactive problem resolution.

Alarm correlation within the system functions by aggregating and analyzing multiple, potentially related, alerts to identify genuine faults and reduce false positives. This process utilizes predefined rules and machine learning algorithms to establish relationships between alerts originating from different network elements and management systems. By suppressing redundant or symptom-based alerts, the system significantly decreases the volume of notifications requiring operator attention. Prioritization of remaining, correlated alerts is then achieved through severity assessment and impact analysis, ensuring that critical issues affecting service availability are addressed with the shortest possible response time. This reduction in alert noise and improved prioritization directly contributes to faster mean time to resolution (MTTR) and enhanced network stability.

The fault diagnosis process integrates network topology data to predict fault propagation and potential service impact. This includes analyzing physical and logical connections between network elements to identify affected services and dependent infrastructure. Furthermore, the system models potential impacts of fiber degradation, accounting for factors like signal attenuation, dispersion, and optical power loss. By cross-referencing fiber health metrics – including OTDR traces and real-time power level monitoring – with topological information, the system can accurately pinpoint fault locations and estimate the scope of the outage, enabling proactive remediation and minimizing service disruption. The system’s ability to model fiber characteristics allows it to differentiate between localized breaks and gradual performance degradation, influencing the urgency and type of response required.

This fault management Agent streamlines root cause localization in live networks by translating operational best practices into a detailed, executable workflow encompassing alarm parsing and fault analysis.
This fault management Agent streamlines root cause localization in live networks by translating operational best practices into a detailed, executable workflow encompassing alarm parsing and fault analysis.

Resilient Networks: Optimizing Performance and Ensuring Stability

The implementation of an intelligent Fault Management system is fundamentally linked to advancements in network performance. This system doesn’t simply react to failures; it anticipates and mitigates potential disruptions through continuous monitoring and analysis of network parameters. By proactively identifying and addressing issues – such as signal degradation or component anomalies – before they escalate, the system minimizes latency and packet loss. This predictive capability allows for dynamic resource allocation and optimized signal transmission, ultimately maximizing throughput and ensuring a consistently high quality of service. The result is a network that operates with greater efficiency, improved stability, and a demonstrably enhanced capacity to handle increasing data demands.

The system’s core strength lies in its predictive capabilities, constantly analyzing network conditions to preemptively resolve issues before they impact performance. This proactive approach directly translates to minimized latency – the delay experienced in data transmission – and maximized throughput, or the amount of data successfully delivered over a given period. By dynamically optimizing signal transmission parameters, such as power levels and modulation schemes, the system ensures data packets travel the most efficient path with minimal obstruction. This isn’t merely reactive troubleshooting; it’s a continuous process of refinement, allowing the network to adapt in real-time to changing demands and maintain consistently high performance levels, even under stress.

Maintaining robust signal integrity across extensive fiber optic networks hinges on sophisticated techniques like Optical Power Equalization. As photons traverse long distances, signal strength diminishes due to absorption, scattering, and other inherent fiber characteristics. Optical Power Equalization dynamically adjusts the transmit power of individual wavelengths, compensating for these losses and ensuring each channel arrives at the receiver with sufficient strength. This isn’t simply amplification; it’s a nuanced process that considers the unique attenuation profile of each fiber span, preventing signal distortion and maximizing the signal-to-noise ratio. By meticulously balancing power levels, the system minimizes bit errors, enhances data throughput, and ultimately extends the reach of the network without requiring costly repeaters or regeneration – a critical factor in building scalable and efficient communication infrastructures.

The culmination of these advancements yields a network fundamentally equipped to handle the escalating demands of contemporary communication systems. Beyond simply reacting to failures, the implemented framework establishes a cycle of continuous optimization and self-correction; it proactively anticipates and mitigates potential disruptions, ensuring consistently high performance and minimized downtime. This closed-loop approach, driven by intelligent algorithms and automated responses, moves beyond conventional network management towards a system capable of autonomous operation and adaptation. The resulting network isn’t merely robust; it’s demonstrably resilient, learning from its operational environment to refine performance and preemptively address challenges, setting a new standard for intelligent and self-healing infrastructure.

This agent optimizes optical network performance by intelligently identifying and tuning optical management segments (OMS) through a workflow encompassing task breakdown, tool interfaces, and prompt design.
This agent optimizes optical network performance by intelligently identifying and tuning optical management segments (OMS) through a workflow encompassing task breakdown, tool interfaces, and prompt design.

The pursuit of automated optical network operations, as detailed in this study, inherently acknowledges the inevitable entropy of complex systems. Every component ages, data streams degrade, and unforeseen faults emerge-a reality that necessitates constant adaptation and refinement. This mirrors the sentiment expressed by Igor Tamm: “The most profound scientific problems are always those which are at the boundary of what is known.” The architecture proposed-with its multi-agent framework and reliance on LLMs-isn’t about achieving static perfection, but about building a resilient system capable of gracefully navigating the unpredictable timeline of network operation. The agents, much like scientific inquiry, push at the boundaries of known states, continuously learning and responding to the network’s evolving condition. This is not merely technical innovation; it’s an acceptance of temporal decay and a commitment to proactive management within it.

What’s Next?

The architecture presented here, while a step toward automating the inevitable entropy of optical networks, merely shifts the locus of failure. The system doesn’t prevent incidents; it restructures them as opportunities for refinement within the LLM agents themselves. The true challenge isn’t creating a responsive system, but building one that degrades predictably – one whose errors become increasingly informative, rather than increasingly catastrophic. Current reliance on digital twin fidelity as a validation mechanism is a temporary reprieve; the twin is, after all, a simplification, a controlled decay mirroring the reality it seeks to preempt.

Future work must confront the inherent unreliability of LLMs not as a bug, but as a fundamental characteristic. Systems will inevitably operate with incomplete or inaccurate information, and the ability to gracefully navigate this ambiguity is paramount. Investigations into self-correcting agents, capable of independently verifying outputs and adapting to changing network conditions, are essential. Further, a deeper consideration of the ‘cost’ of automation – the energetic and computational resources consumed in pursuit of efficiency – is necessary.

Ultimately, the field is less concerned with achieving perfect automation and more with understanding the nature of systemic failure. The network doesn’t strive for immortality; it ages. The question isn’t whether the system will fall, but how it will fall, and whether those failures provide the data necessary for a more resilient, and therefore, more mature successor.


Original article: https://arxiv.org/pdf/2603.11828.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-15 14:18