Benchmarking Q-learning methods for intelligent network orchestration in the edge

article
Osuva_Reijonen_Opsenica_Kauppinen_Komu_Kjällman_Mecklin_Hiltunen_Arkko_Simanainen_Elmusrati_2020.pdf - Hyväksytty kirjoittajan käsikirjoitus - 3.14 MB

Kuvaus

©2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
We benchmark Q-learning methods, with various action selection strategies, in intelligent orchestration of the network edge. Q-learning is a reinforcement learning technique that aims to find optimal action policies by taking advantage of the experiences in the past without utilizing a model that describes the dynamics of the environment. With experiences, we refer to the observed causality between the action and the corresponding impact to the environment. In this paper, the environment for Q-learning is composed of virtualized networking resources along with their dynamics that are monitored with Spindump, an in-network latency measurement tool with support for QUIC and TCP. We optimize the orchestration of these networking resources by introducing Q-learning as part of the machine learning driven, intelligent orchestration that is applicable in the edge. Based on the benchmarking results, we identify which action selection strategies support network orchestration that provides low latency and packet loss by considering network resource allocation in the edge.

Emojulkaisu

2020 2nd 6G Wireless Summit (6G SUMMIT) : 17-20 March 2020, Levi, Finland

ISBN

978-1-7281-6047-4

ISSN

Aihealue

OKM-julkaisutyyppi

A4 Artikkeli konferenssijulkaisussa