Fusion from Multimodal Gait Spatiotemporal Data for Human Gait Speed Classifications

Research output: Contribution to conferencePaperpeer-review

Abstract

Human gait pattens remain largely undefined when relying on a single sensing modality. We report a pilot implementation of sensor fusion to classify gait spatiotemporal signals, from a publicly available dataset of 50 participants, harvested from four different type of sensors. For fusion we propose a hybrid Convolutional Neural Network and Long Short- Term Memory (hybrid CNN+LSTM) and Multi-stream CNN. The classification results are compared to single modality data using Single-stream CNN, a state-of-the-art Vision Transformer, and statistical classifiers algorithms. The fusion models outperformed the single modality methods and classified gait speed of previously unseen 10 random subjects with 97% F1-score prediction accuracy of the four gait speed classes.
Original languageEnglish
Pages1
Number of pages4
DOIs
Publication statusPublished - 31 Oct 2021
EventIEEE SENSORS 2021 - Virtual conference, Sydney, Australia
Duration: 31 Oct 20214 Nov 2021
https://2021.ieee-sensorsconference.org/

Conference

ConferenceIEEE SENSORS 2021
Country/TerritoryAustralia
CitySydney
Period31/10/214/11/21
Internet address

Keywords

  • Deep Convolutional Neural Networks (CNN)
  • Gait Speed
  • Long Short-Term Memory (LSTM)
  • Multimodal Data
  • Transformers

Fingerprint

Dive into the research topics of 'Fusion from Multimodal Gait Spatiotemporal Data for Human Gait Speed Classifications'. Together they form a unique fingerprint.

Cite this