There are no files associated with this item.
Cited time in
Full metadata record
| DC Field | Value | Language |
|---|---|---|
| dc.citation.startPage | 103035 | - |
| dc.citation.title | INTERNATIONAL JOURNAL OF HUMAN-COMPUTER STUDIES | - |
| dc.citation.volume | 176 | - |
| dc.contributor.author | Azim, Md Aashikur Rahman | - |
| dc.contributor.author | Rahman, Adil | - |
| dc.contributor.author | Heo, Seongkook | - |
| dc.date.accessioned | 2026-03-31T14:31:17Z | - |
| dc.date.available | 2026-03-31T14:31:17Z | - |
| dc.date.created | 2026-03-31 | - |
| dc.date.issued | 2023-08 | - |
| dc.description.abstract | Foot-based gestures enable people to interact with mobile and wearable devices when their hands are unavailable for interaction. For the foot gestures to be truly usable, the gestures should be recognizable by the system without being confused by daily activities and still be easy to perform. However, designing such gestures often requires multiple iterations of gesture design, model training, and evaluation. In this paper, we present SequenceSense, a tool developed to help designers efficiently design a usable gesture set using inertial sensors, which eliminates the need for multiple data collection studies to evaluate the gestures' usability through gesture modification by sequencing atomic actions and instant false positive analysis, and instead requires only the initial gesture sample collection. Unlike gesture recognizers using complete gestures to train a model, SequenceSense segments gesture into a sequence of atomic actions. For example, a foot tap to the right may have (1) lift the foot, (2) move the foot to the right, and (3) land the foot. SequenceSense also compares the gesture sequence with the sequence database created from the daily activities to identify possible conflicts. This allows gesture designers to build easily usable foot-based gestures without the need for recollecting and evaluating gestures. We validated SequenceSense's efficacy in designing usable gestures with low false positives through a user study with nine gesture designers. | - |
| dc.identifier.bibliographicCitation | INTERNATIONAL JOURNAL OF HUMAN-COMPUTER STUDIES, v.176, pp.103035 | - |
| dc.identifier.doi | 10.1016/j.ijhcs.2023.103035 | - |
| dc.identifier.issn | 1071-5819 | - |
| dc.identifier.scopusid | 2-s2.0-85151807559 | - |
| dc.identifier.uri | https://scholarworks.unist.ac.kr/handle/201301/91183 | - |
| dc.identifier.url | https://www.sciencedirect.com/science/article/pii/S1071581923000447?pes=vor&utm_source=clarivate&getft_integrator=clarivate | - |
| dc.identifier.wosid | 001054128200001 | - |
| dc.language | 영어 | - |
| dc.publisher | ACADEMIC PRESS LTD- ELSEVIER SCIENCE LTD | - |
| dc.title | SequenceSense: A Tool for Designing Usable Foot-Based Gestures Using a Sequence-Based Gesture Recognizer | - |
| dc.type | Article | - |
| dc.description.isOpenAccess | FALSE | - |
| dc.relation.journalWebOfScienceCategory | Computer Science, Cybernetics; Ergonomics; Psychology, Multidisciplinary | - |
| dc.relation.journalResearchArea | Computer Science; Engineering; Psychology | - |
| dc.type.docType | Article | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | ssci | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.subject.keywordAuthor | Gesture recognizer | - |
| dc.subject.keywordAuthor | Usable gestures | - |
| dc.subject.keywordAuthor | Gesture designer | - |
| dc.subject.keywordAuthor | Foot-based gestures | - |
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Tel : 052-217-1403 / Email : scholarworks@unist.ac.kr
Copyright (c) 2023 by UNIST LIBRARY. All rights reserved.
ScholarWorks@UNIST was established as an OAK Project for the National Library of Korea.