ID | 67970 |
フルテキストURL | |
著者 |
Fukushima, Yukinobu
Faculty of Environmental, Life, Natural Science and Technology Okayama University
Koujitani, Yuki
Graduate School of Natural Science and Technology Okayama University
Nakane, Kazutoshi
Graduate School of Information Science Nagoya University
Tarutani, Yuya
Graduate School of Engineering Osaka University
Wu, Celimuge
Graduate School of Informatics and Engineering The Univ. of Electro-Commun.
Ji, Yusheng
Information Systems Architecture Research Division National Institute of Informatics
Yokohira, Tokumi
Faculty of Interdisciplinary Science and Engineering in Health Systems Okayama University
Kaken ID
publons
researchmap
Murase, Tutomu
Graduate School of Information Science Nagoya University
|
抄録 | This paper tackles a Virtual Machine (VM) migration control problem to maximize the progress (accuracy) of information processing tasks in multi-stage information processing systems. The conventional methods for this problem are effective only for specific situations, such as when the system load is high. In this paper, in order to adaptively achieve high accuracy in various situations, we propose a VM migration method using a Deep Reinforcement Learning (DRL) algorithm. It is difficult to directly apply a DRL algorithm to the VM migration control problem because the size of the solution space of the problem dynamically changes according to the number of VMs staying in the system while the size of the agent’s action space is fixed in DRL algorithms. To cope with this difficulty, the proposed method divides the VM migration control problem into two problems: the problem of determining only the VM distribution (i.e., the proportion of the number of VMs deployed on each edge server) and the problem of determining the locations of all the VMs so that it follows the determined VM distribution. The former problem is solved by a DRL algorithm, and the latter by a heuristic method. This approach makes it possible to apply a DRL algorithm to the VM migration control problem because the VM distribution is expressed by a vector with a fixed number of dimensions and can be directly outputted by the agent. The simulation results confirm that our proposed method can adaptively achieve quasi-optimal accuracy in various situations with different link delays, types of the information processing tasks and the number of VMs.
|
キーワード | Multi-stage information processing system
VM migration control
Deep reinforcement learning
Deep Deterministic Policy Gradient (DDPG)
|
備考 | https://www.thinkmind.org/library/NetSer/NetSer_v17_n34_2024/netser_v17_n34_2024_7.html
|
発行日 | 2024-12-30
|
出版物タイトル |
International Journal On Advances in Networks and Services
|
巻 | 17巻
|
号 | 3-4号
|
出版者 | IARIA
|
開始ページ | 116
|
終了ページ | 125
|
ISSN | 1942-2644
|
資料タイプ |
学術雑誌論文
|
言語 |
英語
|
OAI-PMH Set |
岡山大学
|
著作権者 | © Copyright by authors
|
論文のバージョン | publisher
|
助成機関名 |
Japan Society for the Promotion of Science
|
助成番号 | JP23K11065
JP24K02937
|