File Download

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

이용재

Lee, Yongjae
Financial Engineering Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 52608 -
dc.citation.startPage 52588 -
dc.citation.title IEEE ACCESS -
dc.citation.volume 8 -
dc.contributor.author Cho, Hyunghun -
dc.contributor.author Kim, Yongjin -
dc.contributor.author Lee, Eunjung -
dc.contributor.author Choi, Daeyoung -
dc.contributor.author Lee, Yongjae -
dc.contributor.author Rhee, Wonjong -
dc.date.accessioned 2023-12-21T17:48:51Z -
dc.date.available 2023-12-21T17:48:51Z -
dc.date.created 2020-04-01 -
dc.date.issued 2020-03 -
dc.description.abstract Compared to the traditional machine learning models, deep neural networks (DNN) are known to be highly sensitive to the choice of hyperparameters. While the required time and effort for manual tuning has been rapidly decreasing for the well developed and commonly used DNN architectures, undoubtedly DNN hyperparameter optimization will continue to be a major burden whenever a new DNN architecture needs to be designed, a new task needs to be solved, a new dataset needs to be addressed, or an existing DNN needs to be improved further. For hyperparameter optimization of general machine learning problems, numerous automated solutions have been developed where some of the most popular solutions are based on Bayesian Optimization (BO). In this work, we analyze four fundamental strategies for enhancing BO when it is used for DNN hyperparameter optimization. Specifically, diversification, early termination, parallelization, and cost function transformation are investigated. Based on the analysis, we provide a simple yet robust algorithm for DNN hyperparameter optimization - DEEP-BO (Diversified, Early-termination-Enabled, and Parallel Bayesian Optimization). When evaluated over six DNN benchmarks, DEEP-BO mostly outperformed well-known solutions including GP-Hedge, BOHB, and the speed-up variants that use Median Stopping Rule or Learning Curve Extrapolation. In fact, DEEP-BO consistently provided the top, or at least close to the top, performance over all the benchmark types that we have tested. This indicates that DEEP-BO is a robust solution compared to the existing solutions. The DEEP-BO code is publicly available at https://github.com/snu-adsl/DEEP-BO. -
dc.identifier.bibliographicCitation IEEE ACCESS, v.8, pp.52588 - 52608 -
dc.identifier.doi 10.1109/access.2020.2981072 -
dc.identifier.issn 2169-3536 -
dc.identifier.scopusid 2-s2.0-85082536913 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/31854 -
dc.identifier.url https://ieeexplore.ieee.org/abstract/document/9037259 -
dc.identifier.wosid 000524748500118 -
dc.language 영어 -
dc.publisher IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC -
dc.title Basic Enhancement Strategies When Using Bayesian Optimization for Hyperparameter Tuning of Deep Neural Networks -
dc.type Article -
dc.description.isOpenAccess TRUE -
dc.relation.journalWebOfScienceCategory Computer Science, Information Systems; Engineering, Electrical & Electronic; Telecommunications -
dc.relation.journalResearchArea Computer Science; Engineering; Telecommunications -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor Deep neural networks -
dc.subject.keywordAuthor hyperparameter optimization -
dc.subject.keywordAuthor Bayesian optimization -
dc.subject.keywordAuthor diversification -
dc.subject.keywordAuthor early termination -
dc.subject.keywordAuthor parallelization -
dc.subject.keywordAuthor cost function transformation -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.