File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

김성필

Kim, Sung-Phil
Brain-Computer Interface Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Neural decoding of Aristotle tactile illusion using deep learning-based fMRI classification

Author(s)
Lee, EunjiKim, Ji-HyunPark, JaeseokKim, Sung-PhilShin, Taehoon
Issued Date
2025-06
DOI
10.3389/fnins.2025.1606801
URI
https://scholarworks.unist.ac.kr/handle/201301/87462
Citation
FRONTIERS IN NEUROSCIENCE, v.19, pp.1606801
Abstract
Introduction Aristotle illusion is a well-known tactile illusion which causes the perception of one object as two. EEG analysis was employed to investigate the neural correlates of Aristotle illusion, yet was limited due to low spatial resolution of EEG. This study aimed to identify brain regions involved in the Aristotle illusion using functional magnetic resonance imaging (fMRI) and deep learning-based analysis of fMRI data.Methods While three types of tactile stimuli (Aristotle, Reverse, Asynchronous) were applied to thirty participants' fingers, we collected fMRI data, and recorded the number of stimuli each participant perceived. Four convolutional neural network (CNN) models were trained for perception-based classification tasks (the occurrence of Aristotle illusion vs. Reverse illusion, the occurrence vs. absence of Reverse illusion), and stimulus-based classification tasks (Aristotle vs. Reverse, Reverse vs. Asynchronous, and Aristotle vs. Asynchronous).Results Simple fully convolution network (SFCN) achieved the highest classification accuracy of 68.4% for the occurrence of Aristotle illusion vs. Reverse illusion, and 80.1% for the occurrence vs. absence of Reverse illusion. For stimulus-based classification tasks, all CNN models yielded accuracies around 50% failing to distinguish among the three types of applied stimuli. Gradient-weighted class activation mapping (Grad-CAM) analysis revealed salient brain regions-of-interest (ROIs) for the perception-based classification tasks, including the somatosensory cortex and parietal regions.Discussion Our findings demonstrate that perception-driven neural responses are classifiable using fMRI-based CNN models. Saliency analysis of the trained CNNs reveals the involvement of the somatosensory cortex and parietal regions in making classification decisions, consistent with previous research. Other salient ROIs include orbitofrontal cortex, middle temporal pole, supplementary motor area, and middle cingulate cortex.
Publisher
FRONTIERS MEDIA SA
ISSN
1662-4548
Keyword (Author)
fMRIdeep learningbrain mappingsomatosensorytactile illusion
Keyword
INFORMATION

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.