A new wearable stretch sensor array for 3D spine model visualization during therapeutic exercise

J. Caviedes, B. Li, P. Swan, J. Chen

Research output: Contribution to journalArticlepeer-review

Abstract

Mobile visualization of the spine during therapeutic exercise may unlock the potential benefits of biofeedback for home-based therapy and fitness programs. In this paper we present the design and validation of a new approach to wearable sensors to close the digital-physical gap between sensor data and 3D spine posture by supporting animation of a customizable 3D thoracolumbar spine model that may eventually enable a mobile, virtual reality (VR) visual biofeedback. First, we propose a linear model of the dependency between four stretch signals from the dorsal surface and the angular positions of the thoracolumbar spine on the sagittal, coronal, and transverse planes. We then describe validation experiments and demonstrate the concept by animating the spine model with monoaxial, biaxial and tri-axial motions. The linear model was validated using a 3-way comparison of exercise video, sensor signals, and spine model animation. The computed angular positions were consistent with the video recording, and the animation of the model was visually accurate with a mean absolute error of 3.62 ° for single axis motions and 8.74 ° for dual axis motions. At the end, we provide a discussion on improvements to the linear model performance for cross-axial interactions, plus an outlook of future work.

Original languageEnglish (US)
Article number103732
JournalMedical Engineering and Physics
Volume99
DOIs
StatePublished - Jan 2022

Keywords

  • Spine modeling
  • Spine therapy
  • Spine visualization
  • Stretch sensors
  • Visual biofeedback
  • Wearable sensors

ASJC Scopus subject areas

  • Biophysics
  • Biomedical Engineering

Fingerprint

Dive into the research topics of 'A new wearable stretch sensor array for 3D spine model visualization during therapeutic exercise'. Together they form a unique fingerprint.

Cite this