File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

임성훈

Lim, Sunghoon
Industrial Intelligence Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 157 -
dc.citation.startPage 133 -
dc.citation.title JOURNAL OF MANUFACTURING SYSTEMS -
dc.citation.volume 76 -
dc.contributor.author Kim, Gyeongho -
dc.contributor.author Yang, Sang Min -
dc.contributor.author Kim, Dong Min -
dc.contributor.author Choi, Jae Gyeong -
dc.contributor.author Lim, Sunghoon -
dc.contributor.author Park, Hyung Wook -
dc.date.accessioned 2024-08-08T11:35:05Z -
dc.date.available 2024-08-08T11:35:05Z -
dc.date.created 2024-08-07 -
dc.date.issued 2024-10 -
dc.description.abstract Accurately predicting tool wear is crucial for intelligent machining process monitoring, control, and quality improvement. Recent studies on tool wear prediction predominantly apply deep learning-based data-driven approaches that use multivariate time-series signals from high-precision sensors. However, the reliance on these sensors incurs high installation and operation costs, posing practical challenges for small and mediumsized enterprises. This work proposes a novel deep learning-based approach that employs smartphone sensors to predict tool wear, which addresses the problems associated with smartphone sensor data, including higher noise levels and increased data and model uncertainties. To this end, this work develops various data-driven techniques for effective tool wear prediction and uncertainty quantification. First, a Kalman filter-based noise suppression method is applied to reduce undesired noise effects. Second, a novel uncertainty modeling method consisting of a Bayesian deep learning approach and a density output structure is proposed to capture both aleatoric and epistemic uncertainties during tool wear prediction. The proposed method not only takes into account high noise levels and induced uncertainty, but also continuously quantifies and dissects predictive uncertainty. The proposed method's effectiveness is validated with real-world datasets from Ti-6Al-4V turning experiments under three different machining conditions. The comprehensive experimental results indicate the superior prediction performance of the proposed method compared to existing data-driven methods, probabilistic deep learning-based methods, and state-of-the-art methods. For each of the three distinct datasets, the proposed method provides the lowest mean absolute error (MAE) values of 2.5815, 1.2414, and 1.2269, with the highest R2 2 values of 0.9951, 0.9971, and 0.9982, respectively. -
dc.identifier.bibliographicCitation JOURNAL OF MANUFACTURING SYSTEMS, v.76, pp.133 - 157 -
dc.identifier.doi 10.1016/j.jmsy.2024.07.010 -
dc.identifier.issn 0278-6125 -
dc.identifier.scopusid 2-s2.0-85199874999 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/83419 -
dc.identifier.wosid 001285392900001 -
dc.language 영어 -
dc.publisher Elsevier BV -
dc.title Developing a deep learning-based uncertainty-aware tool wear prediction method using smartphone sensors for the turning process of Ti-6Al-4V -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.relation.journalWebOfScienceCategory Engineering, Industrial;Engineering, Manufacturing -
dc.relation.journalResearchArea Engineering;Operations Research & Management Science -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor Deep learning -
dc.subject.keywordAuthor Aleatoric uncertainty -
dc.subject.keywordAuthor Epistemic uncertainty -
dc.subject.keywordPlus PARTICLE FILTER -
dc.subject.keywordPlus MARKOV MODEL -
dc.subject.keywordPlus SYSTEM -
dc.subject.keywordPlus OPTIMIZATION -
dc.subject.keywordPlus DEMAND -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.