File Download

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

오현동

Oh, Hyondong
Autonomous Systems Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 87792 -
dc.citation.startPage 87777 -
dc.citation.title IEEE ACCESS -
dc.citation.volume 13 -
dc.contributor.author Lee, Junhee -
dc.contributor.author Jang, Hongro -
dc.contributor.author Park, Minkyu -
dc.contributor.author Oh, Hyondong -
dc.date.accessioned 2025-06-17T17:00:02Z -
dc.date.available 2025-06-17T17:00:02Z -
dc.date.created 2025-06-12 -
dc.date.issued 2025-05 -
dc.description.abstract This study investigates the design of reward functions for deep reinforcement learning-based source term estimation (STE). Estimating the properties of unknown hazardous gas leakage using a mobile sensor, known as STE problems, is challenging due to environmental turbulence and sensor noise. To address this issue, the particle filter is employed to estimate the source term under noisy sensor measurements, and the deep Q-network is used to find the optimal source search policy. In deep reinforcement learning, selecting an appropriate reward function is crucial as it directly impacts the learning performance. Specifically, this paper first reviews existing reward functions based on penalty, distance, concentration, and entropy metrics. To overcome the limitations of existing rewards, we combine their strengths and propose new reward functions such as the Gaussian mixture model (GMM) variance-based reward and the GMM information gain-based reward. To validate the robustness of the proposed approach, simulations are conducted in two types of environments: basic and turbulent, by adjusting the parameters of the noise condition. The simulation results demonstrate that the proposed reward functions outperform existing ones and are particularly robust in noisy environments. -
dc.identifier.bibliographicCitation IEEE ACCESS, v.13, pp.87777 - 87792 -
dc.identifier.doi 10.1109/ACCESS.2025.3569827 -
dc.identifier.issn 2169-3536 -
dc.identifier.scopusid 2-s2.0-105005189194 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/87210 -
dc.identifier.wosid 001494102600018 -
dc.language 영어 -
dc.publisher IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC -
dc.title Enhanced Reward Function Design for Source Term Estimation Based on Deep Reinforcement Learning -
dc.type Article -
dc.description.isOpenAccess TRUE -
dc.relation.journalWebOfScienceCategory Computer Science, Information Systems; Engineering, Electrical & Electronic; Telecommunications -
dc.relation.journalResearchArea Computer Science; Engineering; Telecommunications -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor Deep reinforcement learning -
dc.subject.keywordAuthor Position measurement -
dc.subject.keywordAuthor Entropy -
dc.subject.keywordAuthor Robot sensing systems -
dc.subject.keywordAuthor Noise measurement -
dc.subject.keywordAuthor Uncertainty -
dc.subject.keywordAuthor Search problems -
dc.subject.keywordAuthor Source term estimation -
dc.subject.keywordAuthor deep reinforcement learning -
dc.subject.keywordAuthor deep Q-network -
dc.subject.keywordAuthor reward function -
dc.subject.keywordAuthor Bayesian inference -
dc.subject.keywordAuthor particle filter -
dc.subject.keywordAuthor path planning -
dc.subject.keywordAuthor Noise -
dc.subject.keywordAuthor Estimation -
dc.subject.keywordAuthor Mobile agents -
dc.subject.keywordPlus INFOTAXIS -
dc.subject.keywordPlus SOURCE SEARCH -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.