Packet flow capacity autonomous operation based on reinforcement learning
Visualitza/Obre
Cita com:
hdl:2117/359115
Tipus de documentArticle
Data publicació2021-12-12
EditorMultidisciplinary Digital Publishing Institute (MDPI)
Condicions d'accésAccés obert
Llevat que s'hi indiqui el contrari, els
continguts d'aquesta obra estan subjectes a la llicència de Creative Commons
:
Reconeixement 4.0 Internacional
ProjecteB5G-OPEN - Beyond 5G – OPtical nEtwork coNtinuum (EC-H2020-101016663)
AI-POWERED INTENT-BASED PACKET AND OPTICAL TRANSPORT NETWORKS AND EDGE AND CLOUD COMPUTING FOR BEYOND 5G (AEI-PID2020-114135RB-I00)
AI-POWERED INTENT-BASED PACKET AND OPTICAL TRANSPORT NETWORKS AND EDGE AND CLOUD COMPUTING FOR BEYOND 5G (AEI-PID2020-114135RB-I00)
Abstract
As the dynamicity of the traffic increases, the need for self-network operation becomes more evident. One of the solutions that might bring cost savings to network operators is the dynamic capacity management of large packet flows, especially in the context of packet over optical networks. Machine Learning, particularly Reinforcement Learning, seems to be an enabler for autonomicity as a result of its inherent capacity to learn from experience. However, precisely because of that, RL methods might not be able to provide the required performance (e.g., delay, packet loss, and capacity overprovisioning) when managing the capacity of packet flows, until they learn the optimal policy. In view of that, we propose a management lifecycle with three phases: (i) a self-tuned threshold-based approach operating just after the packet flow is set up and until enough data on the traffic characteristics are available; (ii) an RL operation based on models pre-trained with a generic traffic profile; and (iii) an RL operation with models trained for real traffic. Exhaustive simulation results confirm the poor performance of RL algorithms until the optimal policy is learnt and when traffic characteristics change over time, which prevents deploying such methods in operators’ networks. In contrast, the proposed lifecycle outperforms benchmarking approaches, achieving noticeable performance from the beginning of operation while showing robustness against traffic changes.
CitacióBarzegar, S.; Ruíz, M.; Velasco, L. Packet flow capacity autonomous operation based on reinforcement learning. "Sensors (Switzerland)", 12 Desembre 2021, vol. 21, núm. 24, article 8306, p. 1-24.
ISSN1424-8220
Versió de l'editorhttps://www.mdpi.com/1424-8220/21/24/8306
Fitxers | Descripció | Mida | Format | Visualitza |
---|---|---|---|---|
sensors-21-08306-v2.pdf | 15,28Mb | Visualitza/Obre |