There are no files associated with this item.
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.citation.endPage | 1507 | - |
dc.citation.number | 5 | - |
dc.citation.startPage | 1499 | - |
dc.citation.title | IEEE TRANSACTIONS ON MEDICAL IMAGING | - |
dc.citation.volume | 40 | - |
dc.contributor.author | Tang, Yucheng | - |
dc.contributor.author | Gao, Riqiang | - |
dc.contributor.author | Han, Shizhong | - |
dc.contributor.author | Chen, Yunqiang | - |
dc.contributor.author | Gao, Dashan | - |
dc.contributor.author | Nath, Vishwesh | - |
dc.contributor.author | Bermudez, Camilo | - |
dc.contributor.author | Savona, Michael R. | - |
dc.contributor.author | Bao, Shunxing | - |
dc.contributor.author | Lyu, Ilwoo | - |
dc.contributor.author | Huo, Yuankai | - |
dc.contributor.author | Landman, Bennett A. | - |
dc.date.accessioned | 2023-12-21T15:46:26Z | - |
dc.date.available | 2023-12-21T15:46:26Z | - |
dc.date.created | 2022-04-21 | - |
dc.date.issued | 2021-05 | - |
dc.description.abstract | Body part regression is a promising new technique that enables content navigation through self-supervised learning. Using this technique, the global quantitative spatial location for each axial view slice is obtained from computed tomography (CT). However, it is challenging to define a unified global coordinate system for body CT scans due to the large variabilities in image resolution, contrasts, sequences, and patient anatomy. Therefore, the widely used supervised learning approach cannot be easily deployed. To address these concerns, we propose an annotation-free method named blind-unsupervised-supervision network (BUSN). The contributions of the work are in four folds: (1) 1030 multi-center CT scans are used in developing BUSN without any manual annotation. (2) the proposed BUSN corrects the predictions from unsupervised learning and uses the corrected results as the new supervision; (3) to improve the consistency of predictions, we propose a novel neighbor message passing (NMP) scheme that is integrated with BUSN as a statistical learning based correction; and (4) we introduce a new pre-processing pipeline with inclusion of the BUSN, which is validated on 3D multi-organ segmentation. The proposed method is trained on 1,030 whole body CT scans (230,650 slices) from five datasets, as well as an independent external validation cohort with 100 scans. From the body part regression results, the proposed BUSN achieved significantly higher median R-squared score (=0.9089) than the state-of-the-art unsupervised method (=0.7153). When introducing BUSN as a preprocessing stage in volumetric segmentation, the proposed pre-processing pipeline using BUSN approach increases the total mean Dice score of the 3D abdominal multi-organ segmentation from 0.7991 to 0.8145. | - |
dc.identifier.bibliographicCitation | IEEE TRANSACTIONS ON MEDICAL IMAGING, v.40, no.5, pp.1499 - 1507 | - |
dc.identifier.doi | 10.1109/tmi.2021.3058281 | - |
dc.identifier.issn | 0278-0062 | - |
dc.identifier.scopusid | 2-s2.0-85101447194 | - |
dc.identifier.uri | https://scholarworks.unist.ac.kr/handle/201301/58350 | - |
dc.identifier.wosid | 000645866500017 | - |
dc.language | 영어 | - |
dc.publisher | Institute of Electrical and Electronics Engineers | - |
dc.title | Body Part Regression With Self-Supervision | - |
dc.type | Article | - |
dc.description.isOpenAccess | FALSE | - |
dc.type.docType | Article | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.subject.keywordAuthor | Body part regression | - |
dc.subject.keywordAuthor | multi-organ segmentation | - |
dc.subject.keywordAuthor | organ navigation | - |
dc.subject.keywordAuthor | robust regression | - |
dc.subject.keywordAuthor | self-supervised learning | - |
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Tel : 052-217-1404 / Email : scholarworks@unist.ac.kr
Copyright (c) 2023 by UNIST LIBRARY. All rights reserved.
ScholarWorks@UNIST was established as an OAK Project for the National Library of Korea.