There are no files associated with this item.
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.citation.endPage | 2266 | - |
dc.citation.number | 10 | - |
dc.citation.startPage | 2251 | - |
dc.citation.title | IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS | - |
dc.citation.volume | 37 | - |
dc.contributor.author | Lee, Hoon | - |
dc.contributor.author | Lee, Sang Hyun | - |
dc.contributor.author | Quek, Tony Q. S. | - |
dc.date.accessioned | 2023-12-21T18:36:50Z | - |
dc.date.available | 2023-12-21T18:36:50Z | - |
dc.date.created | 2023-09-06 | - |
dc.date.issued | 2019-10 | - |
dc.description.abstract | This paper studies a deep learning (DL) framework to solve distributed non-convex constrained optimizations in wireless networks where multiple computing nodes, interconnected via backhaul links, desire to determine an efficient assignment of their states based on local observations. Two different configurations are considered: First, an infinite-capacity backhaul enables nodes to communicate in a lossless way, thereby obtaining the solution by centralized computations. Second, a practical finite-capacity backhaul leads to the deployment of distributed solvers equipped along with quantizers for communication through capacity-limited backhaul. The distributed nature and the non-convexity of the optimizations render the identification of the solution unwieldy. To handle them, deep neural networks (DNNs) are introduced to approximate an unknown computation for the solution accurately. In consequence, the original problems are transformed to training tasks of the DNNs subject to non-convex constraints where existing DL libraries fail to extend straightforwardly. A constrained training strategy is developed based on the primal-dual method. For distributed implementation, a novel binarization technique at the output layer is developed for quantization at each node. Our proposed distributed DL framework is examined in various network configurations of wireless resource management. Numerical results verify the effectiveness of our proposed approach over existing optimization techniques. | - |
dc.identifier.bibliographicCitation | IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, v.37, no.10, pp.2251 - 2266 | - |
dc.identifier.doi | 10.1109/JSAC.2019.2933890 | - |
dc.identifier.issn | 0733-8716 | - |
dc.identifier.scopusid | 2-s2.0-85070718736 | - |
dc.identifier.uri | https://scholarworks.unist.ac.kr/handle/201301/65464 | - |
dc.identifier.url | https://ieeexplore.ieee.org/document/8792179 | - |
dc.identifier.wosid | 000487055400007 | - |
dc.language | 영어 | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Deep Learning for Distributed Optimization: Applications to Wireless Resource Management | - |
dc.type | Article | - |
dc.description.isOpenAccess | FALSE | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic; Telecommunications | - |
dc.relation.journalResearchArea | Engineering; Telecommunications | - |
dc.type.docType | Article | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.subject.keywordAuthor | Deep neural network | - |
dc.subject.keywordAuthor | distributed deep learning | - |
dc.subject.keywordAuthor | primal-dual method | - |
dc.subject.keywordAuthor | wireless resource management | - |
dc.subject.keywordPlus | NONCONVEX SPECTRUM OPTIMIZATION | - |
dc.subject.keywordPlus | MULTIPLE-ACCESS | - |
dc.subject.keywordPlus | DUAL METHODS | - |
dc.subject.keywordPlus | INFORMATION | - |
dc.subject.keywordPlus | FRAMEWORK | - |
dc.subject.keywordPlus | NETWORKS | - |
dc.subject.keywordPlus | CHANNELS | - |
dc.subject.keywordPlus | DESIGN | - |
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Tel : 052-217-1404 / Email : scholarworks@unist.ac.kr
Copyright (c) 2023 by UNIST LIBRARY. All rights reserved.
ScholarWorks@UNIST was established as an OAK Project for the National Library of Korea.