File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

이훈

Lee, Hoon
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 31 -
dc.citation.number 4 -
dc.citation.startPage 24 -
dc.citation.title IEEE WIRELESS COMMUNICATIONS -
dc.citation.volume 29 -
dc.contributor.author Hwang, Sangwon -
dc.contributor.author Lee, Hoon -
dc.contributor.author Park, Juseong -
dc.contributor.author Lee, Inkyu -
dc.date.accessioned 2023-12-21T13:43:31Z -
dc.date.available 2023-12-21T13:43:31Z -
dc.date.created 2023-09-06 -
dc.date.issued 2022-08 -
dc.description.abstract Limited computing resources of internet-of-things (IoT) nodes incur prohibitive latency in processing input data. This triggers new research opportunities toward task offloading systems where edge servers handle intensive computations of IoT devices. Deploying the computing servers at existing base stations may not be sufficient to support IoT nodes operating in a harsh environment. This requests mobile edge servers to be mounted on unmanned aerial vehicles (UAVs) that provide on-demand mobile edge computing (MEC) services. Time-varying offloading demands and mobility of UAVs need a joint design of the optimization variables for all time instances. Therefore, an online decision mechanism is essential for UAV-aided MEC networks. This article presents an overview of recent deep reinforcement learning (DRL) approaches where decisions about UAVs and IoT nodes are taken in an online manner. Specifically, joint optimization over task offloading, resource allocation, and UAV mobility is addressed from the DRL perspective. For the decentralized implementation, a multi-agent DRL method is proposed where multiple intelligent UAVs cooperatively determine their computations and communication policies without central coordination. Numerical results demonstrate that the proposed decentralized learning strategy is superior to existing DRL solutions. The proposed framework sheds light on the viability of the decentralized DRL techniques in designing self-organizing IoT networks. -
dc.identifier.bibliographicCitation IEEE WIRELESS COMMUNICATIONS, v.29, no.4, pp.24 - 31 -
dc.identifier.doi 10.1109/MWC.003.2100690 -
dc.identifier.issn 1536-1284 -
dc.identifier.scopusid 2-s2.0-85140894041 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/65443 -
dc.identifier.url https://ieeexplore.ieee.org/document/9920737 -
dc.identifier.wosid 000870727800014 -
dc.language 영어 -
dc.publisher IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC -
dc.title Decentralized Computation Offloading with Cooperative UAVs: Multi-Agent Deep Reinforcement Learning Perspective -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.relation.journalWebOfScienceCategory Computer Science, Hardware & Architecture; Computer Science, Information Systems; Engineering, Electrical & Electronic; Telecommunications -
dc.relation.journalResearchArea Computer Science; Engineering; Telecommunications -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor Base stations -
dc.subject.keywordAuthor Multi-access edge computing -
dc.subject.keywordAuthor Reinforcement learning -
dc.subject.keywordAuthor Autonomous aerial vehicles -
dc.subject.keywordAuthor Internet of Things -
dc.subject.keywordAuthor Servers -
dc.subject.keywordAuthor Resource management -
dc.subject.keywordAuthor Multi-agent systems -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.