Benchmarking Q-learning methods for intelligent network orchestration in the edge
Osuva_Reijonen_Opsenica_Kauppinen_Komu_Kjällman_Mecklin_Hiltunen_Arkko_Simanainen_Elmusrati_2020.pdf - Hyväksytty kirjoittajan käsikirjoitus - 3.14 MB
Pysyvä osoite
Kuvaus
©2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
We benchmark Q-learning methods, with various action selection strategies, in intelligent orchestration of the network edge. Q-learning is a reinforcement learning technique that aims to find optimal action policies by taking advantage of the experiences in the past without utilizing a model that describes the dynamics of the environment. With experiences, we refer to the observed causality between the action and the corresponding impact to the environment. In this paper, the environment for Q-learning is composed of virtualized networking resources along with their dynamics that are monitored with Spindump, an in-network latency measurement tool with support for QUIC and TCP. We optimize the orchestration of these networking resources by introducing Q-learning as part of the machine learning driven, intelligent orchestration that is applicable in the edge. Based on the benchmarking results, we identify which action selection strategies support network orchestration that provides low latency and packet loss by considering network resource allocation in the edge.
Emojulkaisu
2020 2nd 6G Wireless Summit (6G SUMMIT) : 17-20 March 2020, Levi, Finland
ISBN
978-1-7281-6047-4
ISSN
Aihealue
OKM-julkaisutyyppi
A4 Artikkeli konferenssijulkaisussa
