There are no files associated with this item.
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.citation.endPage | 31 | - |
dc.citation.number | 4 | - |
dc.citation.startPage | 24 | - |
dc.citation.title | IEEE WIRELESS COMMUNICATIONS | - |
dc.citation.volume | 29 | - |
dc.contributor.author | Hwang, Sangwon | - |
dc.contributor.author | Lee, Hoon | - |
dc.contributor.author | Park, Juseong | - |
dc.contributor.author | Lee, Inkyu | - |
dc.date.accessioned | 2023-12-21T13:43:31Z | - |
dc.date.available | 2023-12-21T13:43:31Z | - |
dc.date.created | 2023-09-06 | - |
dc.date.issued | 2022-08 | - |
dc.description.abstract | Limited computing resources of internet-of-things (IoT) nodes incur prohibitive latency in processing input data. This triggers new research opportunities toward task offloading systems where edge servers handle intensive computations of IoT devices. Deploying the computing servers at existing base stations may not be sufficient to support IoT nodes operating in a harsh environment. This requests mobile edge servers to be mounted on unmanned aerial vehicles (UAVs) that provide on-demand mobile edge computing (MEC) services. Time-varying offloading demands and mobility of UAVs need a joint design of the optimization variables for all time instances. Therefore, an online decision mechanism is essential for UAV-aided MEC networks. This article presents an overview of recent deep reinforcement learning (DRL) approaches where decisions about UAVs and IoT nodes are taken in an online manner. Specifically, joint optimization over task offloading, resource allocation, and UAV mobility is addressed from the DRL perspective. For the decentralized implementation, a multi-agent DRL method is proposed where multiple intelligent UAVs cooperatively determine their computations and communication policies without central coordination. Numerical results demonstrate that the proposed decentralized learning strategy is superior to existing DRL solutions. The proposed framework sheds light on the viability of the decentralized DRL techniques in designing self-organizing IoT networks. | - |
dc.identifier.bibliographicCitation | IEEE WIRELESS COMMUNICATIONS, v.29, no.4, pp.24 - 31 | - |
dc.identifier.doi | 10.1109/MWC.003.2100690 | - |
dc.identifier.issn | 1536-1284 | - |
dc.identifier.scopusid | 2-s2.0-85140894041 | - |
dc.identifier.uri | https://scholarworks.unist.ac.kr/handle/201301/65443 | - |
dc.identifier.url | https://ieeexplore.ieee.org/document/9920737 | - |
dc.identifier.wosid | 000870727800014 | - |
dc.language | 영어 | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Decentralized Computation Offloading with Cooperative UAVs: Multi-Agent Deep Reinforcement Learning Perspective | - |
dc.type | Article | - |
dc.description.isOpenAccess | FALSE | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Hardware & Architecture; Computer Science, Information Systems; Engineering, Electrical & Electronic; Telecommunications | - |
dc.relation.journalResearchArea | Computer Science; Engineering; Telecommunications | - |
dc.type.docType | Article | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.subject.keywordAuthor | Base stations | - |
dc.subject.keywordAuthor | Multi-access edge computing | - |
dc.subject.keywordAuthor | Reinforcement learning | - |
dc.subject.keywordAuthor | Autonomous aerial vehicles | - |
dc.subject.keywordAuthor | Internet of Things | - |
dc.subject.keywordAuthor | Servers | - |
dc.subject.keywordAuthor | Resource management | - |
dc.subject.keywordAuthor | Multi-agent systems | - |
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Tel : 052-217-1404 / Email : scholarworks@unist.ac.kr
Copyright (c) 2023 by UNIST LIBRARY. All rights reserved.
ScholarWorks@UNIST was established as an OAK Project for the National Library of Korea.