The future of wireless communication is being reshaped by the integration of artificial intelligence into open radio access networks (O-RAN), a shift that promises to make cellular systems more adaptable, cost-effective, and intelligent. As networks evolve toward 5G and beyond, O-RAN's open, disaggregated architecture—which breaks down traditional radio access components into interoperable units—offers flexibility but also introduces complex s in managing spectrum, allocating resources, and ensuring security. Machine learning (ML) has emerged as a critical tool to address these issues, enabling real-time optimization and automation that could redefine how networks operate. This survey from researchers at Polytechnique Montreal highlights how ML techniques are being deployed within O-RAN to enhance performance, with applications ranging from dynamic spectrum sharing to proactive threat detection, setting the stage for next-generation wireless services.
Key from the paper reveal that ML, particularly reinforcement learning (RL), is the dominant approach in O-RAN research, accounting for approximately 63% of studies analyzed. RL's ability to learn and adapt through interaction with the network environment makes it well-suited for dynamic tasks like resource allocation and radio resource management. For example, RL algorithms such as Deep Q-Networks (DQN) and Proximal Policy Optimization (PPO) have been used to optimize the allocation of physical resource blocks and computational resources in O-RAN slices, improving efficiency under varying traffic conditions. The survey notes that supervised learning (SL) and unsupervised learning (UL) are also applied, but less frequently, with SL used for tasks like anomaly detection and traffic prediction, and UL for clustering and handover management. This dominance of RL underscores its effectiveness in handling the real-time, adaptive demands of O-RAN systems, where network conditions can change rapidly.
Ology behind integrating ML into O-RAN involves a structured approach centered on the RAN Intelligent Controllers (RICs), which are key architectural components. The O-RAN architecture includes a Non-Real-Time RIC and a Near-Real-Time RIC, operating on timescales greater than 1 second and between 10 ms to 1 second, respectively. These controllers host applications called rApps and xApps that implement ML algorithms for closed-loop control of network functions. For instance, xApps deployed in the Near-RT RIC can use deep reinforcement learning models to dynamically allocate resources based on real-time telemetry data collected via open interfaces like E2 and O1. The paper describes a case study where actor-critic with experience replay (ACER) and PPO models were used to allocate computational resources in O-DUs, with simulations showing that ACER achieved faster and more stable convergence than PPO, as illustrated in Figure 11. This approach allows ML models to be tested in simulated environments before deployment, ensuring they meet performance requirements without disrupting live networks.
From the survey demonstrate significant improvements in network performance and security through ML applications. In spectrum management, RL-based xApps have enabled dynamic spectrum access, allowing secondary users to share underutilized bands without interfering with primary users, thereby increasing spectral efficiency. For resource allocation, DRL models have optimized radio and computation resources, with simulations showing reductions in energy consumption; for example, Figure 12 and Figure 13 compare DRL techniques like ACER and PPO against a greedy policy, revealing that DRL maintains lower energy consumption over time and under high user loads. In security, ML techniques such as random forests and support vector machines have achieved high accuracy (above 99.9%) in detecting distributed denial-of-service attacks, as shown in Figure 14 and Figure 15. The paper also highlights a taxonomy (Figure 10) that categorizes ML usage in O-RAN into three objectives: service quality enhancement (e.g., resource allocation), communication quality enhancement (e.g., spectrum management), and security quality enhancement (e.g., attack detection), each supported by specific ML techniques tailored to s.
Of ML-enabled O-RAN extend beyond technical optimizations to real-world benefits for network operators and users. By automating resource management and spectrum allocation, ML can reduce operational costs and improve energy efficiency, which is crucial as networks scale to support massive IoT and ultra-reliable low-latency communications. Enhanced security through ML-driven anomaly detection helps protect against cyber threats in O-RAN's open, multi-vendor environment, ensuring data confidentiality and network integrity. The survey points to future research directions, such as integrating millimeter-wave and terahertz technologies with ML to address bandwidth limitations, using federated learning for scalable and privacy-preserving model training, and leveraging digital twins for predictive network management. These advancements could accelerate the deployment of 6G networks, making them more intelligent and responsive to diverse service demands, from autonomous vehicles to smart cities.
Despite the progress, the paper identifies several limitations and s in ML integration for O-RAN. ML models, especially RL, are sensitive to hyperparameter tuning and require large amounts of labeled data, which can be scarce in dynamic network environments. Computational and energy demands are high, increasing operational costs, and models are vulnerable to adversarial attacks that can manipulate inputs and degrade performance. For instance, the survey notes that adversarial attacks can cause up to 100% degradation in model accuracy, compromising security defenses. Additionally, the open nature of O-RAN introduces supply chain risks and data confidentiality issues, necessitating robust encryption and access control mechanisms. Future work must address these constraints by developing lightweight, explainable AI models and rigorous testing frameworks to ensure reliability and security in real-world deployments.
Original Source
Read the complete research paper
About the Author
Guilherme A.
Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.
Connect on LinkedIn