Skip to main content
Top

2022 | OriginalPaper | Chapter

Meta-MADDPG: Achieving Transfer-Enhanced MEC Scheduling via Meta Reinforcement Learning

Authors : Yiming Yao, Tao Ren, Meng Cui, Dong Liu, Jianwei Niu

Published in: Wireless Algorithms, Systems, and Applications

Publisher: Springer Nature Switzerland

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

With the assistance of mobile edge computing (MEC), mobile devices (MDs) can optionally offload local computationally heave tasks to edge servers that are generally deployed at the edge of networks. As thus, the latency of task and energy consumption of MDs can be both reduced, significantly improving mobile users’ quality of experience. Although considerable MEC scheduling algorithms have been designed by researchers, most of them are trained to solve specific tasks, leaving the performance in other MEC environments remaining dubious. To address the issue, this paper first formulates the optimization problem to minimize both task delay and energy consumption, and then transforms it into Markov decision problem that is further solved by using the state-of-the-art multi-agent deep reinforcement learning method, i.e., MADDPG. Furthermore, aiming at improving the overall performance in various MEC environments, we integrate MADDPG with meta-learning and propose Meta-MADDPG which is carefully designed with dedicated reward functions. The evaluation results are given to showcase the more satisfactory performances of Meta-MADDPG over the state-of-the-art algorithms when confronting new environments.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Sabella, D., Vaillant, A., Kuure, P., Rauschenbach, U., Giust, F.: Mobile edge computing architecture: the role of MEC in the internet of things. IEEE Consum. Electr. Mag. 5(4), 84–91 (2016)CrossRef Sabella, D., Vaillant, A., Kuure, P., Rauschenbach, U., Giust, F.: Mobile edge computing architecture: the role of MEC in the internet of things. IEEE Consum. Electr. Mag. 5(4), 84–91 (2016)CrossRef
2.
go back to reference Mach, P., Becvar, Z.: Mobile edge computing: a survey on architecture and computation offloading. IEEE Commun. Surv. Tutorials 19(3), 1628–1656 (2017)CrossRef Mach, P., Becvar, Z.: Mobile edge computing: a survey on architecture and computation offloading. IEEE Commun. Surv. Tutorials 19(3), 1628–1656 (2017)CrossRef
3.
go back to reference Arulkumaran, K., Deisenroth, M.P., Brundage, M., Bharath, A.A.: Deep reinforcement learning: a brief survey. IEEE Signal Process. Mag. 34(6), 26–38 (2017)CrossRef Arulkumaran, K., Deisenroth, M.P., Brundage, M., Bharath, A.A.: Deep reinforcement learning: a brief survey. IEEE Signal Process. Mag. 34(6), 26–38 (2017)CrossRef
4.
go back to reference Li, J., Gao, H., Lv, T., Lu, Y.: Deep reinforcement learning based computation offloading and resource allocation for MEC. In: Proceedings of the 2018 IEEE Wireless Communications and Networking Conference (2018) Li, J., Gao, H., Lv, T., Lu, Y.: Deep reinforcement learning based computation offloading and resource allocation for MEC. In: Proceedings of the 2018 IEEE Wireless Communications and Networking Conference (2018)
5.
go back to reference Wang, Y., Ru, Z., Wang, K., Huang, P.: Joint deployment and task scheduling optimization for large-scale mobile users in multi-UAV-enabled mobile edge computing. IEEE Trans. Cybern. 50(9), 3984–3997 (2020)CrossRef Wang, Y., Ru, Z., Wang, K., Huang, P.: Joint deployment and task scheduling optimization for large-scale mobile users in multi-UAV-enabled mobile edge computing. IEEE Trans. Cybern. 50(9), 3984–3997 (2020)CrossRef
6.
go back to reference Peng, H., Shen, X.: Multi-agent reinforcement learning based resource management in MEC- and UAV-assisted vehicular networks. IEEE J. Sel. Areas Commun. 39(1), 131–141 (2021)MathSciNetCrossRef Peng, H., Shen, X.: Multi-agent reinforcement learning based resource management in MEC- and UAV-assisted vehicular networks. IEEE J. Sel. Areas Commun. 39(1), 131–141 (2021)MathSciNetCrossRef
7.
go back to reference Vilalta, R., Drissi, Y.: A perspective view and survey of meta-learning. Artif. Intell. Rev. 18, 77–95 (2002)CrossRef Vilalta, R., Drissi, Y.: A perspective view and survey of meta-learning. Artif. Intell. Rev. 18, 77–95 (2002)CrossRef
8.
go back to reference R. Lowe, Y. Wu, A. Tamar, J Harb, OpenAI Pieter Abbeel, Igor Mordatch: Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. Advances in Neural Information Processing Systems 30 (NIPS 2017) R. Lowe, Y. Wu, A. Tamar, J Harb, OpenAI Pieter Abbeel, Igor Mordatch: Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. Advances in Neural Information Processing Systems 30 (NIPS 2017)
9.
go back to reference Duan, Y., Schulman, J., Chen, X., et al.: RL ^ 2: Fast reinforcement learning via slow reinforcement learning. arXiv preprint arXiv:1611.02779 (2016) Duan, Y., Schulman, J., Chen, X., et al.: RL ^ 2: Fast reinforcement learning via slow reinforcement learning. arXiv preprint arXiv:​1611.​02779 (2016)
10.
go back to reference Wang, C., Yu, F.R., Liang, C., Chen, Q., Tang, L.: Joint computation offloading and interference management in wireless cellular networks with mobile edge computing. IEEE Trans. Veh. Technol. 66(8), 7432–7445 (2017)CrossRef Wang, C., Yu, F.R., Liang, C., Chen, Q., Tang, L.: Joint computation offloading and interference management in wireless cellular networks with mobile edge computing. IEEE Trans. Veh. Technol. 66(8), 7432–7445 (2017)CrossRef
11.
go back to reference Wen, Y., Zhang, W., Luo, H.: Energy-optimal mobile application execution: taming resource-poor mobile devices with cloud clones. In: Proceedings of the 2012 IEEE INFOCOM (2012) Wen, Y., Zhang, W., Luo, H.: Energy-optimal mobile application execution: taming resource-poor mobile devices with cloud clones. In: Proceedings of the 2012 IEEE INFOCOM (2012)
12.
go back to reference Zhou, P., Finley, B., Li, X., Tarkoma, S., Kangasharju, J., Ammar, M., Hui, P.: 5G MEC computation handoff for mobile augmented reality. Computer Science Networking and Internet Architecture (2021) Zhou, P., Finley, B., Li, X., Tarkoma, S., Kangasharju, J., Ammar, M., Hui, P.: 5G MEC computation handoff for mobile augmented reality. Computer Science Networking and Internet Architecture (2021)
13.
go back to reference Hochba, D.S.: Approximation algorithms for NP-hard problems. ACM SIGACT News 28(2), 40–52 (1997)CrossRef Hochba, D.S.: Approximation algorithms for NP-hard problems. ACM SIGACT News 28(2), 40–52 (1997)CrossRef
14.
go back to reference Kaelbling, L.P., Littman, M.L., Moore, A.W.: Reinforcement learning: a survey. J. Artif. Intell. Res. 4, 237–285 (1996)CrossRef Kaelbling, L.P., Littman, M.L., Moore, A.W.: Reinforcement learning: a survey. J. Artif. Intell. Res. 4, 237–285 (1996)CrossRef
15.
go back to reference Cao, Y., Jiang, T., Wang, C.: Optimal radio resource allocation for mobile task offloading in cellular networks. IEEE Network 28(5), 68–73 (2014)CrossRef Cao, Y., Jiang, T., Wang, C.: Optimal radio resource allocation for mobile task offloading in cellular networks. IEEE Network 28(5), 68–73 (2014)CrossRef
Metadata
Title
Meta-MADDPG: Achieving Transfer-Enhanced MEC Scheduling via Meta Reinforcement Learning
Authors
Yiming Yao
Tao Ren
Meng Cui
Dong Liu
Jianwei Niu
Copyright Year
2022
DOI
https://doi.org/10.1007/978-3-031-19211-1_47