File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

류일우

Lyu, Ilwoo
3D Shape Analysis Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace US -
dc.citation.title Medical Imaging 2018: Image Processing -
dc.contributor.author Bobo, M.F. -
dc.contributor.author Bao, S. -
dc.contributor.author Huo, Y. -
dc.contributor.author Yao, Y. -
dc.contributor.author Virostko, J. -
dc.contributor.author Plassard, A.J. -
dc.contributor.author Lyu, Ilwoo -
dc.contributor.author Assad, A. -
dc.contributor.author Abramson, R.G. -
dc.contributor.author Hilmes, M.A. -
dc.contributor.author Landman, B.A. -
dc.date.accessioned 2023-12-19T17:37:14Z -
dc.date.available 2023-12-19T17:37:14Z -
dc.date.created 2021-03-09 -
dc.date.issued 2018-02-11 -
dc.description.abstract Abdominal image segmentation is a challenging, yet important clinical problem. Variations in body size, position, and relative organ positions greatly complicate the segmentation process. Historically, multi-atlas methods have achieved leading results across imaging modalities and anatomical targets. However, deep learning is rapidly overtaking classical approaches for image segmentation. Recently, Zhou et al. showed that fully convolutional networks produce excellent results in abdominal organ segmentation of computed tomography (CT) scans. Yet, deep learning approaches have not been applied to whole abdomen magnetic resonance imaging (MRI) segmentation. Herein, we evaluate the applicability of an existing fully convolutional neural network (FCNN) designed for CT imaging to segment abdominal organs on T2 weighted (T2w) MRI's with two examples. In the primary example, we compare a classical multi-atlas approach with FCNN on forty-five T2w MRI's acquired from splenomegaly patients with five organs labeled (liver, spleen, left kidney, right kidney, and stomach). Thirty-six images were used for training while nine were used for testing. The FCNN resulted in a Dice similarity coefficient (DSC) of 0.930 in spleens, 0.730 in left kidneys, 0.780 in right kidneys, 0.913 in livers, and 0.556 in stomachs. The performance measures for livers, spleens, right kidneys, and stomachs were significantly better than multi-atlas (p < 0.05, Wilcoxon rank-sum test). In a secondary example, we compare the multi-atlas approach with FCNN on 138 distinct T2w MRI's with manually labeled pancreases (one label). On the pancreas dataset, the FCNN resulted in a median DSC of 0.691 in pancreases versus 0.287 for multi-atlas. The results are highly promising given relatively limited training data and without specific training of the FCNN model and illustrate the potential of deep learning approaches to transcend imaging modalities. 1 © COPYRIGHT SPIE. Downloading of the abstract is permitted for personal use only. -
dc.identifier.bibliographicCitation Medical Imaging 2018: Image Processing -
dc.identifier.doi 10.1117/12.2293751 -
dc.identifier.issn 1605-7422 -
dc.identifier.scopusid 2-s2.0-85047332849 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/50144 -
dc.language 영어 -
dc.publisher SPIE -
dc.title Fully convolutional neural networks improve abdominal organ segmentation -
dc.type Conference Paper -
dc.date.conferenceDate 2018-02-11 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.