Projects per year
Abstract
Gesture recognition on smartwatches is challenging not only due to resource constraints but also due to the dynamically changing conditions of users. It is currently an open problem how to engineer gesture recognisers that are robust and yet deployable on smartwatches. Recent research has found that common everyday events, such as a user removing and wearing their smartwatch again, can deteriorate recognition accuracy significantly. In this paper, we suggest that prior understanding of causes behind everyday variability and false positives should be exploited in the development of recognisers. To this end, first, we present a data collection method that aims at diversifying gesture data in a representative way, in which users are taken through experimental conditions that resemble known causes of variability (e.g., walking while gesturing) and are asked to produce deliberately varied, but realistic gestures. Secondly, we review known approaches in machine learning for recogniser design on constrained hardware. We propose convolution-based network variations for classifying raw sensor data, achieving greater than 98% accuracy reliably under both individual and situational variations where previous approaches have reported significant performance deterioration. This performance is achieved with a model that is two orders of magnitude less complex than previous state-of-the-art models. Our work suggests that deployable and robust recognition is feasible but requires systematic efforts in data collection and network design to address known causes of gesture variability.
Original language | English |
---|---|
Title of host publication | 27th International Conference on Intelligent User Interfaces, IUI 2022 |
Publisher | ACM |
Pages | 277-291 |
Number of pages | 15 |
ISBN (Electronic) | 978-1-4503-9144-3 |
DOIs | |
Publication status | Published - 22 Mar 2022 |
MoE publication type | A4 Conference publication |
Event | International Conference on Intelligent User Interfaces - Virtual, Online, Finland Duration: 22 Mar 2022 → 25 Mar 2022 Conference number: 27 |
Publication series
Name | International Conference on Intelligent User Interfaces, Proceedings IUI |
---|
Conference
Conference | International Conference on Intelligent User Interfaces |
---|---|
Abbreviated title | IUI |
Country/Territory | Finland |
City | Virtual, Online |
Period | 22/03/2022 → 25/03/2022 |
Keywords
- Deep Learning
- Gestures
- Mobile Devices
- Sensing
- Wearables
Fingerprint
Dive into the research topics of 'Robust and Deployable Gesture Recognition for Smartwatches'. Together they form a unique fingerprint.Projects
- 3 Finished
-
Human Automata: Simulator-based Methods for Collaborative AI
Oulasvirta, A. (Principal investigator), Shiripour, M. (Project Member), Putkonen, A.-M. (Project Member), Rastogi, A. (Project Member), Hegemann, L. (Project Member), Iyer, A. (Project Member), Santala, S. (Project Member), Dayama, N. (Project Member), Laine, M. (Project Member), Halasinamara Chandramouli, S. (Project Member), Li, C. (Project Member), Zhu, Y. (Project Member), Liao, Y.-C. (Project Member), Kylmälä, J. (Project Member), Nioche, A. (Project Member) & Kompatscher, J. (Project Member)
01/01/2020 → 31/12/2023
Project: Academy of Finland: Other research funding
-
-: Finnish Center for Artificial Intelligence
Kaski, S. (Principal investigator)
01/01/2019 → 31/12/2022
Project: Academy of Finland: Other research funding
-
-: Bayesian Artefact Design
Oulasvirta, A. (Principal investigator), Shin, J. (Project Member), Hegemann, L. (Project Member), Todi, K. (Project Member), Putkonen, A.-M. (Project Member), Halasinamara Chandramouli, S. (Project Member), Hassinen, H. (Project Member), Dayama, N. (Project Member), Leiva, L. (Project Member), Laine, M. (Project Member), Zhu, Y. (Project Member), Liao, Y.-C. (Project Member), Peng, Z. (Project Member) & Nioche, A. (Project Member)
01/09/2018 → 31/08/2023
Project: Academy of Finland: Other research funding