File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

류일우

Lyu, Ilwoo
3D Shape Analysis Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

High-resolution 3D abdominal segmentation with random patch network fusion

Author(s)
Tang, Y.Gao, R.Lee, H.H.Han, S.Chen, Y.Gao, D.Nath, V.Bermudez, C.Savona, M.R.Abramson, R.G.Bao, S.Lyu, IlwooHuo, Y.Landman, B.A.
Issued Date
2021-04
DOI
10.1016/j.media.2020.101894
URI
https://scholarworks.unist.ac.kr/handle/201301/50086
Citation
MEDICAL IMAGE ANALYSIS, v.69
Abstract
Deep learning for three dimensional (3D) abdominal organ segmentation on high-resolution computed tomography (CT) is a challenging topic, in part due to the limited memory provide by graphics processing units (GPU) and large number of parameters and in 3D fully convolutional networks (FCN). Two prevalent strategies, lower resolution with wider field of view and higher resolution with limited field of view, have been explored but have been presented with varying degrees of success. In this paper, we propose a novel patch-based network with random spatial initialization and statistical fusion on overlapping regions of interest (ROIs). We evaluate the proposed approach using three datasets consisting of 260 subjects with varying numbers of manual labels. Compared with the canonical “coarse-to-fine” baseline methods, the proposed method increases the performance on multi-organ segmentation from 0.799 to 0.856 in terms of mean DSC score (p-value < 0.01 with paired t-test). The effect of different numbers of patches is evaluated by increasing the depth of coverage (expected number of patches evaluated per voxel). In addition, our method outperforms other state-of-the-art methods in abdominal organ segmentation. In conclusion, the approach provides a memory-conservative framework to enable 3D segmentation on high-resolution CT. The approach is compatible with many base network structures, without substantially increasing the complexity during inference. Given a CT scan with at high resolution, a low-res section (left panel) is trained with multi-channel segmentation. The low-res part contains down-sampling and normalization in order to preserve the complete spatial information. Interpolation and random patch sampling (mid panel) is employed to collect patches. The high-dimensional probability maps are acquired (right panel) from integration of all patches on field of views. © 2020
Publisher
Elsevier BV
ISSN
1361-8415
Keyword (Author)
3D CTAbdominal organ segmentationCoarse to fineHigh resolutionNetwork fusion
Keyword
malememoryComputer graphicsConvolutional neural networksDeep learningGraphics processing unitProgram processorsConvolutional networksHigh-resolution computed tomographyMulti-organ segmentationsNetwork structuresOverlapping regionsSpatial informationsState-of-the-art methodsThreedimensional (3-d)Computerized tomographyabdominal visceraadultarticlecontrolled studyprobabilityfemalehigh resolution computer tomographyhumanhuman experimentmajor clinical study

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.