Hyppää sisältöön
    • FI
    • ENG
  • FI
  • /
  • EN
OuluREPO – Oulun yliopiston julkaisuarkisto / University of Oulu repository
Näytä viite 
  •   OuluREPO etusivu
  • Oulun yliopisto
  • Avoin saatavuus
  • Näytä viite
  •   OuluREPO etusivu
  • Oulun yliopisto
  • Avoin saatavuus
  • Näytä viite
JavaScript is disabled for your browser. Some features of this site may not work without it.

MyoKey : surface electromyography and inertial motion sensing-based text entry in AR

Kwon, Young D.; Shatilov, Kirill A.; Lee, Lik-Hang; Kumyol, Serkan; Lam, Kit-Yung; Yau, Yui-Pan; Hui, Pan (2020-08-04)

 
Avaa tiedosto
nbnfi-fe2020081148292.pdf (311.1Kt)
nbnfi-fe2020081148292_meta.xml (42.02Kt)
nbnfi-fe2020081148292_solr.xml (35.15Kt)
Lataukset: 

URL:
https://doi.org/10.1109/PerComWorkshops48775.2020.9156084

Kwon, Young D.
Shatilov, Kirill A.
Lee, Lik-Hang
Kumyol, Serkan
Lam, Kit-Yung
Yau, Yui-Pan
Hui, Pan
Institute of Electrical and Electronics Engineers
04.08.2020

Y. D. Kwon et al., "MyoKey: Surface Electromyography and Inertial Motion Sensing-based Text Entry in AR," 2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Austin, TX, USA, 2020, pp. 1-4, doi: 10.1109/PerComWorkshops48775.2020.9156084

https://rightsstatements.org/vocab/InC/1.0/
© 2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
https://rightsstatements.org/vocab/InC/1.0/
doi:https://doi.org/10.1109/PerComWorkshops48775.2020.9156084
Näytä kaikki kuvailutiedot
Julkaisun pysyvä osoite on
https://urn.fi/URN:NBN:fi-fe2020081148292
Tiivistelmä

Abstract

The seamless textual input in Augmented Reality (AR) is very challenging and essential for enabling user-friendly AR applications. Existing approaches such as speech input and vision-based gesture recognition suffer from environmental obstacles and the large default keyboard size, sacrificing the majority of the screen’s real estate in AR. In this paper, we propose MyoKey, a system that enables users to effectively and unobtrusively input text in a constrained environment of AR by jointly leveraging surface Electromyography (sEMG) and Inertial Motion Unit (IMU) signals transmitted by wearable sensors on a user’s forearm. MyoKey adopts a deep learning-based classifier to infer hand gestures using sEMG. In order to show the feasibility of our approach, we implement a mobile AR application using the Unity application building framework. We present novel interaction and system designs to incorporate information of hand gestures from sEMG and arm motions from IMU to provide seamless text entry solution. We demonstrate the applicability of MyoKey by conducting a series of experiments achieving the accuracy of 0.91 on identifying five gestures in real-time (Inference time: 97.43 ms).

Kokoelmat
  • Avoin saatavuus [37957]
oulurepo@oulu.fiOulun yliopiston kirjastoOuluCRISLaturiMuuntaja
SaavutettavuusselosteTietosuojailmoitusYlläpidon kirjautuminen
 

Selaa kokoelmaa

NimekkeetTekijätJulkaisuajatAsiasanatUusimmatSivukartta

Omat tiedot

Kirjaudu sisäänRekisteröidy
oulurepo@oulu.fiOulun yliopiston kirjastoOuluCRISLaturiMuuntaja
SaavutettavuusselosteTietosuojailmoitusYlläpidon kirjautuminen