このエントリーをはてなブックマークに追加


ID 67970
フルテキストURL
著者
Fukushima, Yukinobu Faculty of Environmental, Life, Natural Science and Technology Okayama University
Koujitani, Yuki Graduate School of Natural Science and Technology Okayama University
Nakane, Kazutoshi Graduate School of Information Science Nagoya University
Tarutani, Yuya Graduate School of Engineering Osaka University
Wu, Celimuge Graduate School of Informatics and Engineering The Univ. of Electro-Commun.
Ji, Yusheng Information Systems Architecture Research Division National Institute of Informatics
Yokohira, Tokumi Faculty of Interdisciplinary Science and Engineering in Health Systems Okayama University Kaken ID publons researchmap
Murase, Tutomu Graduate School of Information Science Nagoya University
抄録
This paper tackles a Virtual Machine (VM) migration control problem to maximize the progress (accuracy) of information processing tasks in multi-stage information processing systems. The conventional methods for this problem are effective only for specific situations, such as when the system load is high. In this paper, in order to adaptively achieve high accuracy in various situations, we propose a VM migration method using a Deep Reinforcement Learning (DRL) algorithm. It is difficult to directly apply a DRL algorithm to the VM migration control problem because the size of the solution space of the problem dynamically changes according to the number of VMs staying in the system while the size of the agent’s action space is fixed in DRL algorithms. To cope with this difficulty, the proposed method divides the VM migration control problem into two problems: the problem of determining only the VM distribution (i.e., the proportion of the number of VMs deployed on each edge server) and the problem of determining the locations of all the VMs so that it follows the determined VM distribution. The former problem is solved by a DRL algorithm, and the latter by a heuristic method. This approach makes it possible to apply a DRL algorithm to the VM migration control problem because the VM distribution is expressed by a vector with a fixed number of dimensions and can be directly outputted by the agent. The simulation results confirm that our proposed method can adaptively achieve quasi-optimal accuracy in various situations with different link delays, types of the information processing tasks and the number of VMs.
キーワード
Multi-stage information processing system
VM migration control
Deep reinforcement learning
Deep Deterministic Policy Gradient (DDPG)
備考
https://www.thinkmind.org/library/NetSer/NetSer_v17_n34_2024/netser_v17_n34_2024_7.html
発行日
2024-12-30
出版物タイトル
International Journal On Advances in Networks and Services
17巻
3-4号
出版者
IARIA
開始ページ
116
終了ページ
125
ISSN
1942-2644
資料タイプ
学術雑誌論文
言語
英語
OAI-PMH Set
岡山大学
著作権者
© Copyright by authors
論文のバージョン
publisher
助成機関名
Japan Society for the Promotion of Science
助成番号
JP23K11065
JP24K02937