File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

공태식

Gong, Taesik
Ubiquitous AI Lab
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace SI -
dc.citation.endPage 41512 -
dc.citation.startPage 41487 -
dc.citation.title International Conference on Learning Representations -
dc.contributor.author Baek, Eunsu -
dc.contributor.author Han, Sung-Hwan -
dc.contributor.author Gong, Taesik -
dc.contributor.author Kim, Hyung-Sin -
dc.date.accessioned 2026-01-05T14:31:34Z -
dc.date.available 2026-01-05T14:31:34Z -
dc.date.created 2025-12-26 -
dc.date.issued 2025-04-24 -
dc.description.abstract Domain shift remains a persistent challenge in deep-learning-based computer vision, often requiring extensive model modifications or large labeled datasets to address. Inspired by human visual perception, which adjusts input quality through corrective lenses rather than over-training the brain, we propose Lens, a novel camera sensor control method that enhances model performance by capturing high-quality images from the model's perspective, rather than relying on traditional human-centric sensor control. Lens is lightweight and adapts sensor parameters to specific models and scenes in real-time (i.e., test-time input adaptation). At its core, Lens utilizes VisiT, a training-free, model-specific quality indicator that evaluates individual unlabeled samples at test time using confidence scores, without additional adaptation costs. To validate Lens, we introduce ImageNet-ES Diverse, a new benchmark dataset capturing natural perturbations from varying sensor and lighting conditions. Extensive experiments on both ImageNet-ES and our new ImageNet-ES Diverse show that Lens significantly improves model accuracy across various baseline schemes for sensor control and model modification, while maintaining low latency in image captures. Lens effectively compensates for large model size differences and integrates synergistically with model improvement techniques. Our code and dataset are available at github.com/Edw2n/Lens.git. -
dc.identifier.bibliographicCitation International Conference on Learning Representations, pp.41487 - 41512 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/89771 -
dc.language 영어 -
dc.publisher International Conference on Learning Representations, ICLR -
dc.title ADAPTIVE CAMERA SENSOR FOR VISION MODELS -
dc.type Conference Paper -
dc.date.conferenceDate 2025-04-24 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.