Last updated: 2026-01-17 05:01 UTC
All documents
Number of pages: 154
| Author(s) | Title | Year | Publication | Keywords | ||
|---|---|---|---|---|---|---|
| M. Gharbaoui, F. Sciarrone, M. Fontana, P. Castoldi, B. Martini | Assurance and Conflict Detection in Intent-Based Networking: A Comprehensive Survey and Insights on Standards and Open-Source Tools | 2026 | Early Access | Surveys Translation Bandwidth Real-time systems Runtime Robustness Systematic literature review Monitoring Heuristic algorithms Engines IBN Intent Assurance Conflict detection Standards Open-source IBN | Intent-Based Networking (IBN) enables operators to specify high-level outcomes while the system translates these intents into concrete policies and configurations. As IBN deployments grow in scale, heterogeneity and dynamicity, ensuring continuous alignment between network behavior and user objectives becomes both essential and increasingly difficult. This paper provides a technical survey of assurance and conflict detection techniques in IBN, with the goal of improving reliability, robustness, and policy compliance. We first position our survey with respect to existing work. We then review current assurance mechanisms, including the use of AI, machine learning, and real-time monitoring for validating intent fulfillment. We also examine conflict detection methods across the intent lifecycle, from capture to implementation. In addition, we outline relevant standardization efforts and open-source tools that support IBN adoption. Finally, we discuss key challenges, such as AI/ML integration, generalization, and scalability, and present a roadmap for future research aimed at strengthening robustness of IBN frameworks. | 10.1109/TNSM.2026.3651896 |
| Yeryeong Cho, Sungwon Yi, Soohyun Park | Joint Multi-Agent Reinforcement Learning and Message-Passing for Resilient Multi-UAV Networks | 2026 | Early Access | Servers Heuristic algorithms Autonomous aerial vehicles Training Surveillance Reliability Training data Reinforcement learning Resource management Resilience Multi-Agent System (MAS) Reinforcement Learning (RL) Communication Graph Message Passing Resilient Communication Network Unmanned Aerial Vehicle (UAV) UAVs Networks | This paper introduces a novel resilient algorithm designed for distributed unmanned aerial vehicles (UAVs) in dynamic and unreliable network environments. Initially, the UAVs should be trained via multi-agent reinforcement learning (MARL) for autonomous mission-critical operations and are fundamentally grounded by centralized training and decentralized execution (CTDE) using a centralized MARL server. In this situation, it is crucial to consider the case where several UAVs cannot receive CTDE-based MARL learning parameters for resilient operations in unreliable network conditions. To tackle this issue, a communication graph is used where its edges are established when two UAVs/nodes are communicable. Then, the edge-connected UAVs can share their training data if one of the UAVs cannot be connected to the CTDE-based MARL server under unreliable network conditions. Additionally, the edge cost considers power efficiency. Based on this given communication graph, message-passing is used for electing the UAVs that can provide their MARL learning parameters to their edge-connected peers. Lastly, performance evaluations demonstrate the superiority of our proposed algorithm in terms of power efficiency and resilient UAV task management, outperforming existing benchmark algorithms. | 10.1109/TNSM.2025.3650697 |
| Yilu Chen, Ye Wang, Ruonan Li, Yujia Xiao, Lichen Liu, Jinlong Li, Yan Jia, Zhaoquan Gu | TrafficAudio: Audio Representation for Lightweight Encrypted Traffic Classification in IoT | 2026 | Early Access | Feature extraction Cryptography Telecommunication traffic Accuracy Malware Vectors Spatiotemporal phenomena Security Intrusion detection Computational efficiency Encrypted traffic classification Malicious traffic detection Mel-frequency cepstral coefficients Traffic representation | Encrypted traffic classification has become a crucial task for network management and security with the widespread adoption of encrypted protocols across the Internet and the Internet of Things. However, existing methods often rely on discrete representations and complex models, which leads to incomplete feature extraction, limited fine-grained classification accuracy, and high computational costs. To this end, we propose TrafficAudio, a novel encrypted traffic classification method based on audio representation. TrafficAudio comprises three modules: audio representation generation (ARG), audio feature extraction (AFE), and spatiotemporal traffic classification (STC). Specifically, the ARG module first represents raw network traffic as audio to preserve temporal continuity of traffic. Then, the audio is processed by the AFE module to compute low-dimensional Mel-frequency cepstral coefficients (MFCC), encoding both temporal and spectral characteristics. Finally, spatiotemporal features are extracted from MFCC through a parallel architecture of one-dimensional convolutional neural network and bidirectional gated recurrent unit layers, enabling fine-grained traffic classification. Experiments on five public datasets across six classification tasks demonstrate that TrafficAudio consistently outperforms ten state-of-the-art baselines, achieving accuracies of 99.74%, 98.40%, 99.76%, 99.25%, 99.77%, and 99.74%. Furthermore, TrafficAudio significantly reduces computational complexity, achieving reductions of 86.88% in floating-point operations and 43.15% of model parameters over the best-performing baseline. | 10.1109/TNSM.2026.3651599 |
| Haoran Hu, Huazhi Lun, Ya Wang, Zhifeng Deng, Jiahao Li, Yuexiang Cao, Ying Liu, Heng Zhang, Jie Tang, Huicun Yu, Jiahua Wei, Xingyu Wang, Lei Shi | Effective Resource Scheduling Design for Concurrent Competing Requests in Quantum Networks | 2026 | Early Access | Purification Quantum networks Quantum entanglement Throughput Damping Scheduling Routing Resource management Qubit Noise Quantum networks resource scheduling concurrent competing requests entanglement fidelity | Quantum networks, as a pivotal platform to support numerous quantum applications, have the potential to far exceed traditional communication networks. Establishing end-to-end entanglement connections with guaranteed fidelity is a key prerequisite for realizing the functionality of quantum networks. Entanglement purification techniques are commonly used in the entanglement distribution process to provide end-to-end entanglement connections that meet the fidelity requirements. Since the purification operation sacrifices a certain amount of entanglement resources, it is critical and challenging to efficiently utilize the scarce entanglement resources in quantum networks with concurrent competing requests. To address this problem, we propose a novel demand-oriented resource scheduling (DRS) algorithm. Considering the overall network demand, DRS introduces a congestion factor to evaluate the resource demand of each link, and performs purification operations sequentially based on the congestion level of the links, thus avoiding the excessive consumption of entanglement resources of bottleneck links. Extensive simulation results show that the DRS algorithm can achieve higher network throughput with similar resource conversion rates compared to traditional resource allocation schemes. Our work provides a new scheme for the resource scheduling problem under concurrent competing requests, which can promote the further development of existing entanglement routing techniques. | 10.1109/TNSM.2026.3651862 |
| Jack Wilkie, Hanan Hindy, Craig Michie, Christos Tachtatzis, James Irvine, Robert Atkinson | A Novel Contrastive Loss for Zero-Day Network Intrusion Detection | 2026 | Early Access | Contrastive learning Anomaly detection Training Autoencoders Training data Detectors Data models Vectors Telecommunication traffic Network intrusion detection Internet of Things Network Intrusion Detection Machine Learning Contrastive Learning | Machine learning has achieved state-of-the-art results in network intrusion detection; however, its performance significantly degrades when confronted by a new attack class— a zero-day attack. In simple terms, classical machine learning-based approaches are adept at identifying attack classes on which they have been previously trained, but struggle with those not included in their training data. One approach to addressing this shortcoming is to utilise anomaly detectors which train exclusively on benign data with the goal of generalising to all attack classes— both known and zero-day. However, this comes at the expense of a prohibitively high false positive rate. This work proposes a novel contrastive loss function which is able to maintain the advantages of other contrastive learning-based approaches (robustness to imbalanced data) but can also generalise to zero-day attacks. Unlike anomaly detectors, this model learns the distributions of benign traffic using both benign and known malign samples, i.e. other well-known attack classes (not including the zero-day class), and consequently, achieves significant performance improvements. The proposed approach is experimentally verified on the Lycos2017 dataset where it achieves an AUROC improvement of.000065 and.060883 over previous models in known and zero-day attack detection, respectively. Finally, the proposed method is extended to open-set recognition achieving OpenAUC improvements of.170883 over existing approaches.The implementation and experiments are open-sourced and available at: https://github.com/jackwilkie/CLOSR | 10.1109/TNSM.2026.3652529 |
| Marco Polverini, Andrés García-López, Juan Luis Herrera, Santiago García-Gil, Francesco G. Lavacca, Antonio Cianfrani, Jaime Galán-Jiménez | Avoiding SDN Application Conflicts With Digital Twins: Design, Models and Proof of Concept | 2026 | Early Access | Digital twins Analytical models Routing Delays Data models Reliability Switches Software defined networking Routing protocols Reviews Network Digital Twin SDN Data Plane SLA | Software-Defined Networking (SDN) enables flexible and programmable control over network behavior through the deployment of multiple control applications. However, when these applications operate simultaneously, each pursuing different and potentially conflicting objectives, unexpected interactions may arise, leading to policy violations, performance degradation, or inefficient resource usage. This paper presents a Digital Twin (DT)-based framework for the early detection of such application-level conflicts. The proposed framework is lightweight, modular, and designed to be seamlessly integrated into real SDN controllers. It includes multiple DT models capturing different network aspects, including end-to-end delay, link congestion, reliability, and carbon emissions. A case study in a smart factory scenario demonstrates the framework’s ability to identify conflicts arising from coexisting applications with heterogeneous goals. The solution is validated through both simulation and proof-of-concept implementation tested in an emulated environment using Mininet. The performance evaluation shows that three out of four DT models achieve a precision above 90%, while the minimum recall across all models exceeds 84%. Moreover, the proof of concept confirms that what-if analyses can be executed in a few milliseconds, enabling timely and proactive conflict detection. These results demonstrate that the framework can accurately detect conflicts and deliver feedback fast enough to support timely network adaptation. | 10.1109/TNSM.2026.3652800 |
| Jian Ye, Lisi Mo, Gaolei Fei, Yunpeng Zhou, Ming Xian, Xuemeng Zhai, Guangmin Hu, Ming Liang | TopoKG: Infer Internet AS-Level Topology From Global Perspective | 2026 | Early Access | Business Topology Routing Internet Knowledge graphs Accuracy Network topology Probabilistic logic Inference algorithms Border Gateway Protocol AS-level topology business relationship hierarchical structure knowledge graph global perspective | Internet Autonomous System (AS) level topology includes AS topology structure and AS business relationships, describes the essence of Internet inter-domain routing, and is the basis for Internet operation and management research. Although the latest topology inference methods have made significant progress, those relying solely on local information struggle to eliminate inference errors caused by observation bias and data noise due to their lack of a global perspective. In contrast, we not only leverage local AS link features but also re-examine the hierarchical structure of Internet AS-level topology, proposing a novel inference method called topoKG. TopoKG introduces a knowledge graph to represent the relationships between different elements on a global scale and the business routing strategies of ASes at various tiers, which effectively reduces inference errors resulting from observation bias and data noise by incorporating a global perspective. First, we construct an Internet AS-level topology knowledge graph to represent relevant data, enabling us to better leverage the global perspective and uncover the complex relationships among multiple elements. Next, we employ knowledge graph meta paths to measure the similarity of AS business routing strategies and introduce this global perspective constraint to infer the AS business relationships and hierarchical structure iteratively. Additionally, we embed the entire knowledge graph upon completing the iteration and conduct knowledge inference to derive AS business relationships. This approach captures global features and more intricate relational patterns within the knowledge graph, further enhancing the accuracy of AS-level topology inference. Compared to the state-of-the-art methods, our approach achieves more accurate AS-level topology inference, reducing the average inference error across various AS link types by up to 1.2 to 4.4 times. | 10.1109/TNSM.2026.3652956 |
| Shagufta Henna, Upaka Rathnayake | Hypergraph Representation Learning-Based xApp for Traffic Steering in 6G O-RAN Closed-Loop Control | 2026 | Early Access | Open RAN Resource management Ultra reliable low latency communication Throughput Heuristic algorithms Computer architecture Accuracy 6G mobile communication Seals Real-time systems Open Radio Access Network (O-RAN) Intelligent Traffic Steering Link Prediction for Traffic Management | This paper addresses the challenges in resource allocation within disaggregated Radio Access Networks (RAN), particularly when dealing with Ultra-Reliable Low-Latency Communications (uRLLC), enhanced Mobile Broadband (eMBB), and Massive Machine-Type Communications (mMTC). Traditional traffic steering methods often overlook individual user demands and dynamic network conditions, while multi-connectivity further complicates resource management. To improve traffic steering, we introduce Tri-GNN-Sketch, a novel graph-based deep learning approach employing Tri-subgraph sampling to enhance link prediction in Open RAN (O-RAN) environments. Link prediction refers to accurately forecasting optimal connections between users and network resources using current and historical measurements. Tri-GNN-Sketch is trained on real-world 4G/5G RAN monitoring data. The model demonstrates robust performance across multiple metrics, including precision, recall, F1 score, and ROC-AUC, effectively modeling interfering nodes for accurate traffic steering. We further propose Tri-HyperGNN-Sketch, which extends the approach to hypergraph modeling, capturing higher-order multi-node relationships. Using link-level simulations based on Channel Quality Indicator (CQI)-to-modulation mappings and LTE transport block size specifications, we evaluate throughput and packet delay for Tri-HyperGNN-Sketch. Tri-HyperGNN-Sketch achieves an exceptional link prediction accuracy of 99.99% and improved network-level performance, including higher effective throughput and lower packet delay compared to Tri-GNN-Sketch (95.1%) and other hypergraph-based models such as HyperSAGE (91.6%) and HyperGCN (92.31%) for traffic steering in complex O-RAN deployments. | 10.1109/TNSM.2026.3654534 |
| Apurba Adhikary, Avi Deb Raha, Yu Qiao, Md. Shirajum Munir, Mrityunjoy Gain, Zhu Han, Choong Seon Hong | Age of Sensing Empowered Holographic ISAC Framework for NextG Wireless Networks: A VAE and DRL Approach | 2026 | Early Access | Array signal processing Resource management Integrated sensing and communication Wireless networks Phased arrays Hardware Arrays Real-time systems Metamaterials 6G mobile communication Integrated sensing and communication age of sensing holographic MIMO deep reinforcement learning artificial intelligence framework | This paper proposes an AI framework that leverages integrated sensing and communication (ISAC), aided by the age of sensing (AoS) to ensure the timely location updates of the users for a holographic MIMO (HMIMO)-assisted base station (BS)-enabled wireless network. The AI-driven framework aims to achieve optimized power allocation for efficient beamforming by activating the minimal number of grids from the HMIMO BS for serving the users. An optimization problem is formulated to maximize the sensing utility function, aiming to maximize the communication signal-to-interference-plus-noise ratio (SINRc) of the received signals and beam-pattern gains to improve the sensing SINR of reflected echo signals, which in turn maximizes the achievable rate of users. A novel AI-driven framework is presented to tackle the formulated NP-hard problem that divides it into two problems: a sensing problem and a power allocation problem. The sensing problem is solved by employing a variational autoencoder (VAE)-based mechanism that obtains the sensing information leveraging AoS, which is used for the location update. Subsequently, a deep deterministic policy gradient-based deep reinforcement learning scheme is devised to allocate the desired power by activating the required grids based on the sensing information achieved with the VAE-based mechanism. Simulation results demonstrate the superior performance of the proposed AI framework compared to advantage actor-critic and deep Q-network-based methods, achieving a cumulative average SINRc improvement of 8.5 dB and 10.27 dB, and a cumulative average achievable rate improvement of 21.59 bps/Hz and 4.22 bps/Hz, respectively. Therefore, our proposed AI-driven framework guarantees efficient power allocation for holographic beamforming through ISAC schemes leveraging AoS. | 10.1109/TNSM.2026.3654889 |
| Jing Zhang, Chao Luo, Rui Shao | MTG-GAN: A Masked Temporal Graph Generative Adversarial Network for Cross-Domain System Log Anomaly Detection | 2026 | Early Access | Anomaly detection Adaptation models Generative adversarial networks Feature extraction Data models Load modeling Accuracy Robustness Contrastive learning Chaos Log Anomaly Detection Generative Adversarial Networks (GANs) Temporal Data Analysis | Anomaly detection of system logs is crucial for the service management of large-scale information systems. Nowadays, log anomaly detection faces two main challenges: 1) capturing evolving temporal dependencies between log events to adaptively tackle with emerging anomaly patterns, 2) and maintaining high detection capabilities across varies data distributions. Existing methods rely heavily on domain-specific data features, making it challenging to handle the heterogeneity and temporal dynamics of log data. This limitation restricts the deployment of anomaly detection systems in practical environments. In this article, a novel framework, Masked Temporal Graph Generative Adversarial Network (MTG-GAN), is proposed for both conventional and cross-domain log anomaly detection. The model enhances the detection capability for emerging abnormal patterns in system log data by introducing an adaptive masking mechanism that combines generative adversarial networks with graph contrastive learning. Additionally, MTG-GAN reduces dependency on specific data distribution and improves model generalization by using diffused graph adjacency information deriving from temporal relevance of event sequence, which can be conducive to improve cross-domain detection performance. Experimental results demonstrate that MTG-GAN outperforms existing methods on multiple real-world datasets in both conventional and cross-domain log anomaly detection. | 10.1109/TNSM.2026.3654642 |
| Manjuluri Anil Kumar, Balaprakasa Rao Killi, Eiji Oki | Generative Adversarial Networks Based Low-Rate Denial of Service Attack Detection and Mitigation in Software-Defined Networks | 2026 | Vol. 23, Issue | Protocols Prevention and mitigation Real-time systems Software defined networking Generative adversarial networks Anomaly detection Denial-of-service attack TCP Routing Training LDoS SDN GAN attack detection and mitigation OpenFlow | Low-rate Denial of Service (LDoS) attacks use short, regular bursts of traffic to exploit vulnerabilities in network protocols. They are a major threat to network security, especially in Software-Defined Networking (SDN) frameworks. These attacks are challenging to detect and mitigate because of their low traffic volume, making it impossible to distinguish them from normal traffic. We propose a real-time LDoS attack detection and mitigation framework that can protect SDN. The framework incorporates a detection module that uses a deep learning model, such as a Generative Adversarial Network (GAN), to identify the attack. An efficient mitigation module follows detection, employing mechanisms to identify and filter harmful flows in real time. Deploying the framework into SDN controllers guarantees compliance with OpenFlow standards, thereby avoiding the necessity for additional hardware. Experimental results demonstrate that the proposed system achieves a detection accuracy of over 99.98% with an average response time of 8.58 s, significantly outperforming traditional LDoS detection approaches. This study presents a scalable, real-time methodology to enhance SDN resilience against LDoS attacks. | 10.1109/TNSM.2025.3625278 |
| Haftay Gebreslasie Abreha, Ilora Maity, Youssouf Drif, Christos Politis, Symeon Chatzinotas | Revenue-Aware Seamless Content Distribution in Satellite-Terrestrial Integrated Networks | 2026 | Vol. 23, Issue | Satellites Topology User experience Network topology Delays Real-time systems Optimization Low earth orbit satellites Collaboration Servers Satellite edge computing (SEC) content caching content distribution dynamic ad insertion | With the surging demand for data-intensive applications, ensuring seamless content delivery in Satellite-Terrestrial Integrated Networks (STINs) is crucial, especially for remote users. Dynamic Ad Insertion (DAI) enhances monetization and user experience, while Mobile Edge Computing (MEC) in STINs enables distributed content caching and ad insertion. However, satellite mobility and time-varying topologies cause service disruptions, while excessive or poorly placed ads risk user disengagement, impacting revenue. This paper proposes a novel framework that jointly addresses three challenges: (i) service continuity- and topology-aware content caching to adapt to STIN dynamics, (ii) Distributed DAI (D-DAI) that minimizes feeder link load and storage overhead by avoiding redundant ad-variant content storage through distributed ad stitching, and (iii) revenue-aware content distribution that explicitly models user disengagement due to ad overload to balance monetization and user satisfaction. We formulate the problem as two hierarchical Integer Linear Programming (ILP) optimizations: one content caching that aims to maximize cache hit rate and another optimizing content distribution with DAI to maximize revenue, minimize end-user costs, and enhance user experience. We develop greedy algorithms for fast initialization and a Binary Particle Swarm Optimization (BPSO)–based strategy for enhanced performance. Simulation results demonstrate that the proposed approach achieves over a 4.5% increase in revenue and reduces cache retrieval delay by more than 39% compared to the benchmark algorithms. | 10.1109/TNSM.2025.3629810 |
| Meixia Miao, Peihong Qiang, Siqi Zhao, Jiawei Li, Guohua Tian, Jianghong Wei | Verifiable Data Streaming Protocol Supporting Keyword Queries | 2026 | Vol. 23, Issue | Protocols Cloud computing Trees (botanical) Security Servers Indexes Cryptography Outsourcing Hash functions Databases Verifiable data streaming prefix tree chameleon authentication tree keyword query | The rapid deployment of emerging networks, such as the Internet of Things and cloud computing, has generated massive amounts of data. Data streaming is significant among these various data types due to its widespread use in many critical applications, such as gene sequencing, network intrusion detection, and stock trading. On the other hand, the continuously increased size of data streaming makes it impractical to store and manage the data locally, especially for those resource-constrained devices. Outsourcing the data streaming to cloud servers provides an ideal solution to the above storage issue. However, this raises the problem of how to guarantee the integrity of the outsourced data, as cloud servers may maliciously modify the data. To this end, the primitive of verifiable data streaming (VDS) was introduced to preserve the integrity of the outsourced data streaming, enabling data users to ensure that queried data items, including the contents and corresponding positions, are correct. Despite many proposed VDS protocols, most can only use the position index to query outsourced data streaming. Consequently, they fail to fulfill the requirements of those practical applications that need keyword queries. For example, in the setting of network intrusion detection, the data analyst would like to query all access records from the same IP address. In this paper, we extend the original VDS protocol to support keyword queries, i.e., allowing data users to retrieve outsourced data items with particular keywords. Specifically, we use a prefix tree to maintain keywords and another chameleon authentication tree to store data items. The two trees are bound together with cryptographic query proofs, ensuring the consistency between the position index and keyword queries. The proposed VDS protocol, which supports keyword queries, is proven secure in the standard model and outperforms previous VDS protocols in terms of functionality. The experimental results indicate that our proposal is also efficient and practical. | 10.1109/TNSM.2025.3629071 |
| Hai Anh Tran, Nam-Thang Hoang | Towards Efficient and Adaptive Traffic Classification: A Knowledge Distillation-Based Personalized Federated Learning Framework | 2026 | Vol. 23, Issue | Adaptation models Training Federated learning Data models Telecommunication traffic Computational modeling Accuracy Knowledge engineering Heterogeneous networks Knowledge transfer Personalized federated learning knowledge distillation traffic classification heterogeneous network systems adaptive model personalization | Traffic classification plays a crucial role in optimizing network management, enhancing security, and enabling intelligent resource allocation in distributed network systems. However, traditional Federated Learning (FL) approaches struggle with domain heterogeneity, as network traffic characteristics vary significantly across different domains due to diverse infrastructure, applications, and usage patterns. This results in degraded performance when applying a single global model across all domains. To overcome this challenge, we propose KD-PFL-TC, a Knowledge Distillation-based Personalized Federated Learning framework for Traffic Classification, aimed to balance global knowledge sharing with personalized model adaptation in heterogeneous network environments. Our approach leverages knowledge distillation to enable collaborative learning without directly sharing raw data, preserving privacy while mitigating the negative effects of domain shifts. Each domain refines its local model by integrating insights from a global model and peer domains while maintaining its unique traffic distribution. To further enhance performance, we introduce an adaptive distillation strategy that dynamically adjusts the influence of global, peer, and local knowledge based on the similarity between traffic distributions, ensuring optimal knowledge transfer designed to each domain’s characteristics. Extensive experiments on real-world traffic datasets show that KD–PFL–TC maintains 88.0% accuracy under high heterogeneity (vs. 75.0% for FedAvg) while reducing communication overhead by ~60%, delivering an efficient and robust solution for large-scale, heterogeneous networks. | 10.1109/TNSM.2025.3629241 |
| F. Busacca, L. Galluccio, S. Palazzo, A. Panebianco, R. Raftopoulos | Bandits Under the Waves: A Fully-Distributed Multi-Armed Bandit Framework for Modulation Adaptation in the Internet of Underwater Things | 2026 | Vol. 23, Issue | Throughput Scalability Training Propagation losses Mathematical models Energy consumption Adaptation models Absorption Support vector machines Internet Underwater communications underwater modulation adaptation reinforcement learning multi-player multi-armed bandit | Acoustic communications are the most exploited technology in the so-called Internet of Underwater Things (IoUT). UnderWater (UW) environments are often characterized by harsh propagation features, limited bandwidth, fast-varying channel conditions, and long propagation delay. On the other hand, IoUT nodes are usually battery-powered devices with limited processing capabilities. Accordingly, it is necessary to design optimization algorithms to address the challenging propagation features while balancing them with the limited device capabilities. To address the constraints of the nodes in energy and processing resources, it is crucial to adjust the transmission parameters based on the channel conditions while also developing communication procedures that are both lightweight and energy-efficient. In this work, we introduce a novel Multi-Player Multi-Armed Bandit (MP-MAB) framework for modulation adaptation in Multi-Hop IoUT Acoustic Networks. As opposed to widely used, computation-demanding Deep Reinforcement Learning (DRL) techniques, MP-MAB algorithms are simple and lightweight and allow to iteratively make decisions by selecting one among multiple choices, or arms. The framework is fully-distributed and is able to dynamically select the best modulation technique at each IoUT node by leveraging on high-level statistics (e.g., network throughput), without the need to exploit hard-to-extract channel features (e.g., channel state). We evaluate the performance of the proposed framework using the DESERT UW simulator and compare it with state-of-the-art centralized solutions based on Deep Reinforcement Learning (DRL) for cognitive and heterogeneous networks, namely DRL-MCS, DRL-AM, PPO, SAC, as well as with a multiple-agent, distributed version of the PPO. The results highlight that, despite its simplicity and fully-distributed nature, the proposed framework achieves superior performance in UW networks in terms of throughput, convergence speed, and energy efficiency. Compared to DRL-MCS and DRL-AM, our approach improves network throughput by up to 33% and 20%, respectively, and reduces energy consumption by up to 18% and 16%. When compared to PPO, SAC, and Multi-PPO, the proposed solution achieves up to 11%, 34%, and 38% higher throughput, and up to 7%, 17%, and 33% lower energy consumption, respectively. | 10.1109/TNSM.2025.3629240 |
| Leonardo Lo Schiavo, Genoveva García, Marco Gramaglia, Marco Fiore, Albert Banchs, Xavier Costa-Perez | The TES Framework: Joint Statistical Modeling and Machine Learning for Network KPI Forecasting | 2026 | Vol. 23, Issue | Predictive models Forecasting Time series analysis Adaptation models Load modeling Deep learning Autonomous networks Context modeling Accuracy Transformers Forecasting prediction mobile traffic network KPI network management neural networks statistical modeling | The vision of intelligent networks capable of automatically configuring crucial parameters for tasks such as resource provisioning, anomaly detection or load balancing largely hinges upon efficient AI-based algorithms. Time series forecasting is a fundamental building block for network-oriented AI and current trends lean towards the systematic adoption of models based on deep learning approaches. In this paper, we pave the way for a different strategy for the design of predictors for mobile network environments, and we propose the Thresholded Exponential Smoothing (TES) framework, a hybrid Statistical Modeling and Deep Learning tool that allows for improving the performance of network Key Performance Indicator (KPI) forecasting. We adapt our framework to two state-of-the-art deep learning tools for time series forecasting, based on Recurrent Neural Networks and Transformer architectures. We experiment with TES by showcasing its superior support for three practical network management use cases, i.e., (i) anticipatory allocation of network resources, (ii) mobile traffic anomaly prediction, and (iii) mobile traffic load balancing. Our results, derived from traffic measurements collected in operational mobile networks, demonstrate that the TES framework can yield substantial performance gains over current state-of-the-art predictors in the applications considered. | 10.1109/TNSM.2025.3628788 |
| Hojjat Navidan, Cristian Martín, Vasilis Maglogiannis, Dries Naudts, Manuel Díaz, Ingrid Moerman, Adnan Shahid | An End-to-End Digital Twin Framework for Dynamic Traffic Analytics in O-RAN | 2026 | Vol. 23, Issue | Open RAN Adaptation models Real-time systems Biological system modeling 5G mobile communication Predictive models Traffic control Incremental learning Anomaly detection Data models Digital twin generative AI open radio access networks incremental learning traffic analytics traffic prediction anomaly detection | Dynamic traffic patterns and shifts in traffic distribution in Open Radio Access Networks (O-RAN) pose a significant challenge for real-time network optimization in 5G and beyond. Traditional traffic analytics methods struggle to remain accurate under such non-stationary conditions, where models trained on historical data quickly degrade as traffic evolves. This paper introduces AIDITA, an AI-driven Digital Twin for Traffic Analytics framework designed to solve this problem through autonomous model adaptation. AIDITA creates a digital replica of the live analytics models running in the RAN Intelligent Controller (RIC) and continuously updates them within the digital twin using incremental learning. These updates use real-time Key Performance Metrics (KPMs) from the live network, augmented with synthetic data from a Generative AI (GenAI) component to simulate diverse network scenarios. Combining GenAI-driven augmentation with incremental learning enables traffic analytics models, such as prediction or anomaly detection, to adapt continuously without the need for full retraining, preserving accuracy and efficiency in dynamic environments. Implemented and validated on a real-world 5G testbed, our AIDITA framework demonstrates significant improvements in traffic prediction and anomaly detection use cases under distribution shifts, showcasing its practical effectiveness and adaptability for real-time network optimization in O-RAN deployments. | 10.1109/TNSM.2025.3628756 |
| Guiyun Liu, Hao Li, Lihao Xiong, Zhongwei Liang, Xiaojing Zhong | Attention-Model-Based Multiagent Reinforcement Learning for Combating Malware Propagation in Internet of Underwater Things | 2026 | Vol. 23, Issue | Malware Mathematical models Predictive models Optimal control Prediction algorithms Adaptation models Wireless communication Optimization Network topology Vehicle dynamics Internet of Underwater Things (IoUT) malware fractional-order model model-based reinforcement learning (MBRL) | Malware propagation in Internet of Underwater Things (IoUT) can disrupt stable communications among wireless devices. Timely control over its spread is beneficial for the stable operation of IoUT. Notably, the instability of the underwater environment causes the propagation effects of malware to vary continuously. Traditional control methods cannot quickly adapt to these abrupt changes. In recent years, the rapid development of reinforcement learning (RL) has significantly advanced control schemes. However, previous RL methods relied on long-term interactions to obtain a large amount of interaction data in order to form effective strategy. Given the particularity of underwater communication media, data collection for RL in IoUT is challenging. Therefore, improving sample efficiency has become a critical issue that current RL methods need to address urgently. The algorithm of Attention-Model-Based Multiagent Policy Optimization (AMBMPO) is proposed to achieve efficient use of data samples in this study. First, the algorithm employs an explicit prediction model to reduce the dependence on precise model. Secondly, an attention mechanism network is designed to capture high-dimensional state sequences, thereby reducing the compound errors during policy training. Finally, the proposed method is validated for optimal control problems and compared with verified benchmarks. The experimental results show that, compared with existing advanced RL algorithms, AMBMPO demonstrates significant advantages in sample efficiency and stability. This work effectively controls the spread of malware in underwater systems through an interactive evolution based approach. It provides a new implementation approach for ensuring the safety of underwater systems in deep-sea exploration and environmental monitoring applications. | 10.1109/TNSM.2025.3628881 |
| Md Ibrahim Ibne Alam, Anindo Mahmood, Prasun Kanti Dey, Murat Yuksel, Koushik Kar | Meta-Peering: Automating ISP Peering Decision Process | 2026 | Vol. 23, Issue | Costs Routing Internet Border Gateway Protocol Automation Web and internet services Monitoring Peering Internet service provider Internet exchange point network management traffic engineering | Peering between Internet Service Providers (ISPs) is playing an increasingly critical role in Internet traffic exchange. As content delivery networks continue to expand, major content ISPs are increasingly opting for peering arrangements over transit services to facilitate faster exchange of traffic. The satisfaction of the ISP pair and the longevity of the peering arrangement depend on the stability and performance of these peering relationships. We introduce meta-peering, a term which refers to the set of tools needed to help and automate the ISP peering process – starting with identifying a list of ISPs that are likely to peer, writing router rules to establish BGP sessions with them, and extending the service to monitor all these sessions for notifying any major outages or peering agreement violations. In this paper, we first make a thorough analysis of recent trends in ISP peering and describe how meta-peering can be implemented by integrating some of the existing tools. We mainly focus on instrumenting the automation of the peer selection process with an aim to identifying potential peering partners and peering locations to exchange traffic. Using these direct peering links greatly reduces energy consumption as traffic takes much shorter paths to their destinations, going through reduced number of intermediary devices (e.g., routers, switches) compared to elongated transit routes, consequently reducing the environmental impact. Utilizing PeeringDB and CAIDA datasets to identify possible peering points for ISP pairs, we consider ISPs’ internal policies to generate a list of acceptable peering contracts (APCs). We design two methodologies to rank order each ISP in the APC list and offer guidance on which ones would be stable and beneficial for the potential peers. A study of more than 3,000 ISP pairs (mostly active in North America) shows that our peer selection methods can attain around 80% accuracy in predicting peering relations. | 10.1109/TNSM.2024.3459796 |
| Ke Gu, Jiaqi Lei, Jingjing Tan, Xiong Li | A Verifiable Federated Learning Scheme With Privacy-Preserving in MCS | 2026 | Vol. 23, Issue | Federated learning Sensors Servers Security Training Protocols Privacy Homomorphic encryption Computational modeling Mobile computing Mobile crowd sensing verifiable federated learning privacy-preserving sampling verification | The popularity of edge smart devices and the explosive growth of generated data have driven the development of mobile crowd sensing (MCS). Also, federated learning (FL), as a new paradigm of privacy-preserving distributed machine learning, integrates with MCS to offer a novel approach for processing large-scale edge device data. However, it also brings about many security risks. In this paper, we propose a verifiable federated learning scheme with privacy-preserving for mobile crowd sensing. In our federated learning scheme, the double-layer random mask partition method combined with homomorphic encryption is constructed to protect the local gradients and enhance system security (strong anti-collusion ability) based on the multi-cluster structure of federated learning. Also, a sampling verification mechanism is proposed to allow the mobile sensing clients to quickly and efficiently verify the correctness of their received gradient aggregation results. Further, a dropout handling mechanism is constructed to improve the robustness of mobile crowd sensing-based federated learning. Related experimental results demonstrate that our verifiable federated learning scheme is effective and efficient in mobile crowd sensing environments. | 10.1109/TNSM.2025.3627581 |