Accurate and fast classification of foot gestures for virtual locomotion

  • Xinyu Shi
  • , Junjun Pan
  • , Zeyong Hu
  • , Juncong Lin
  • , Shihui Guo*
  • , Minghong Liao
  • , Ye Pan
  • , Ligang Liu
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

This work explores the use of foot gestures for locomotion in virtual environments. Foot gestures are represented as the distribution of plantar pressure and detected by three sparsely-located sensors on each insole. The Long Short-Term Memory model is chosen as the classifier to recognize the performer's foot gesture based on the captured signals of pressure information. The trained classifier directly takes the noisy and sparse input of sensor data, and handles seven categories of foot gestures (stand, walk forward/backward, run, jump, slide left and right) without manual definition of signal features for classifying these gestures. This classifier is capable of recognizing the foot gestures, even with the existence of large sensor-specific, inter-person and intra-person variations. Results show that an accuracy of ~80% can be achieved across different users with different shoe sizes and ~85% for users with the same shoe size. A novel method, Dual-Check Till Consensus, is proposed to reduce the latency of gesture recognition from 2 seconds to 0.5 seconds and increase the accuracy to over 97%. This method offers a promising solution to achieve lower latency and higher accuracy at a minor cost of computation workload. The characteristics of high accuracy and fast classification of our method could lead to wider applications of using foot patterns for human-computer interaction.

Original languageEnglish
Title of host publicationProceedings - 2019 IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2019
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages178-189
Number of pages12
ISBN (Electronic)9781728109879
DOIs
StatePublished - Oct 2019
Event18th IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2019 - Beijing, China
Duration: 14 Oct 201918 Oct 2019

Publication series

NameProceedings - 2019 IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2019

Conference

Conference18th IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2019
Country/TerritoryChina
CityBeijing
Period14/10/1918/10/19

Keywords

  • Gestural input
  • Human centered computing
  • Human computer interaction (HCI)
  • Interaction techniques
  • Interactive systems and tools
  • User interface programming

Fingerprint

Dive into the research topics of 'Accurate and fast classification of foot gestures for virtual locomotion'. Together they form a unique fingerprint.

Cite this