File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

이훈

Lee, Hoon
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 11402 -
dc.citation.number 11 -
dc.citation.startPage 11385 -
dc.citation.title IEEE TRANSACTIONS ON COMMUNICATIONS -
dc.citation.volume 73 -
dc.contributor.author Lee, Hoon -
dc.contributor.author Zhou Wentao -
dc.contributor.author Debbah Merouane -
dc.contributor.author Lee Inkyu -
dc.date.accessioned 2025-12-10T09:44:07Z -
dc.date.available 2025-12-10T09:44:07Z -
dc.date.created 2025-12-09 -
dc.date.issued 2025-11 -
dc.description.abstract Future wireless networks are expected to incorporate diverse services that often lack general mathematical models. To address such black-box network management tasks, the large language model (LLM) optimizer framework, which leverages pretrained LLMs as optimization agents, has recently been promoted as a promising solution. This framework utilizes natural language prompts describing the given optimization problems along with past solutions generated by LLMs themselves. As a result, LLMs can obtain efficient solutions autonomously without knowing the mathematical models of the objective functions. Although the viability of the LLM optimizer (LLMO) framework has been studied in various black-box scenarios, it has so far been limited to numerical simulations. For the first time, this paper establishes a theoretical foundation for the LLMO framework. With careful investigations of LLM inference steps, we can interpret the LLMO procedure as a finite-state Markov chain, and prove the convergence of the framework. Our results are extended to a more advanced multiple LLM architecture, where the impact of multiple LLMs is rigorously verified in terms of the convergence rate. Comprehensive numerical simulations validate our theoretical results and provide a deeper understanding of the underlying mechanisms of the LLMO framework. -
dc.identifier.bibliographicCitation IEEE TRANSACTIONS ON COMMUNICATIONS, v.73, no.11, pp.11385 - 11402 -
dc.identifier.doi 10.1109/TCOMM.2025.3592598 -
dc.identifier.issn 0090-6778 -
dc.identifier.scopusid 2-s2.0-105012299486 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/88975 -
dc.identifier.wosid 001618327600014 -
dc.language 영어 -
dc.publisher IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC -
dc.title On the Convergence of Large Language Model Optimizer for Black-Box Network Management -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor OptimizationMathematical modelsConvergenceVectorsLinear programmingClosed boxNumerical simulationLarge language modelsResource managementGenetic algorithmsLarge language models (LLMs)black-box optimization (BBO)finite-state Markov chain -
dc.subject.keywordPlus MULTIUSER MIMO SYSTEMS -
dc.subject.keywordPlus COMPUTATION -
dc.subject.keywordPlus DESIGN -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.