File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 1993 -
dc.citation.number 2 -
dc.citation.startPage 1986 -
dc.citation.title IEEE ROBOTICS AND AUTOMATION LETTERS -
dc.citation.volume 11 -
dc.contributor.author Chong, Haechan -
dc.contributor.author Lee, Jongwon -
dc.contributor.author Ahn, Hyemin -
dc.date.accessioned 2026-01-19T17:56:45Z -
dc.date.available 2026-01-19T17:56:45Z -
dc.date.created 2026-01-19 -
dc.date.issued 2026-02 -
dc.description.abstract Recent robot task planners utilize large language models (LLMs) or vision-language models (VLMs) as a failure detector. These methods perform well by leveraging their semantic reasoning capabilities but often assume full environment understanding, which can lead to unreliable planning in complex scenes lacking explicit structural modeling. To address these limitations, we propose a novel multi-view scene understanding framework that explicitly models object-level relationships, enabling failure detection and effective task replanning. Our approach first captures multi-view images for comprehensive coverage, and generates local 2D scene graphs encoding object identities and relational information. Building on this, we introduce a model based on a graph neural network that merges the local 2D scene graphs into a unified representation. This process results in the unified scene graph, used to detect task success and identify failure causes. For each sub-task, our framework compares the unified scene graph with the expected scene graph predicted by the LLM during the task planning stage, identifying potential failure causes based on their deviations. These causes are then fed back into the LLM to facilitate effective replanning, thereby reducing repetitive failures and enhancing adaptability. We evaluate our framework on five real-world benchmark tasks to demonstrate its applicability. Separately, we compare failure detection and reasoning performance with other methods, showing the benefits of combining multi-view perception with explicit graph-based reasoning. -
dc.identifier.bibliographicCitation IEEE ROBOTICS AND AUTOMATION LETTERS, v.11, no.2, pp.1986 - 1993 -
dc.identifier.doi 10.1109/LRA.2025.3645659 -
dc.identifier.issn 2377-3766 -
dc.identifier.scopusid 2-s2.0-105025426564 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/90317 -
dc.identifier.wosid 001651966100017 -
dc.language 영어 -
dc.publisher IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC -
dc.title Robust Task Planning via Failure Detection Using Scene Graph From Multi-View Images -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.relation.journalWebOfScienceCategory Robotics -
dc.relation.journalResearchArea Robotics -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor Task and motion planning -
dc.subject.keywordAuthor failure detection and recovery -
dc.subject.keywordAuthor AI-enabled robotics -
dc.subject.keywordPlus LANGUAGE -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.