File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace KO -
dc.citation.endPage 777 -
dc.citation.startPage 775 -
dc.citation.title 8th International Conference on Information and Communication Technology Convergence, ICTC 2017 -
dc.contributor.author Park, Jung Kyu -
dc.contributor.author Kim, Jaeho -
dc.date.accessioned 2023-12-19T18:08:34Z -
dc.date.available 2023-12-19T18:08:34Z -
dc.date.created 2018-08-10 -
dc.date.issued 2017-10-18 -
dc.description.abstract In this paper, we attempt to manage GC overhead at the operating system level. In our approach, first, we use a machine learning technique to devise a GC detecting mechanism at the operating system level, and second, we show that by making use of this mechanism performance variance normally observed on SSDs can be reduced. We develop a GC-detector that detects garbage collection of SSDs and request TRIM operations to the SSD when GC is detected. Experimental results running the GC-detector show increase average bandwidth and low performance variance compared to when not using GC-detector. -
dc.identifier.bibliographicCitation 8th International Conference on Information and Communication Technology Convergence, ICTC 2017, pp.775 - 777 -
dc.identifier.doi 10.1109/ICTC.2017.8190778 -
dc.identifier.scopusid 2-s2.0-85046885779 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/37686 -
dc.identifier.url https://ieeexplore.ieee.org/document/8190778/ -
dc.language 영어 -
dc.publisher ICTC -
dc.title A method for reducing garbage collection overhead of SSD using machine learning algorithms -
dc.type Conference Paper -
dc.date.conferenceDate 2017-10-18 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.