Pressure Eye: In-bed Contact Pressure Estimation via Contact-less Imaging
Computer vision has achieved great success in interpreting semantic meanings from images, yet estimating underlying (non-visual) physical properties of an object is often limited to their bulk values rather than reconstructing a dense map. In this work, we present our pressure eye (PEye) approach to...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Computer vision has achieved great success in interpreting semantic meanings
from images, yet estimating underlying (non-visual) physical properties of an
object is often limited to their bulk values rather than reconstructing a dense
map. In this work, we present our pressure eye (PEye) approach to estimate
contact pressure between a human body and the surface she is lying on with high
resolution from vision signals directly. PEye approach could ultimately enable
the prediction and early detection of pressure ulcers in bed-bound patients,
that currently depends on the use of expensive pressure mats. Our PEye network
is configured in a dual encoding shared decoding form to fuse visual cues and
some relevant physical parameters in order to reconstruct high resolution
pressure maps (PMs). We also present a pixel-wise resampling approach based on
Naive Bayes assumption to further enhance the PM regression performance. A
percentage of correct sensing (PCS) tailored for sensing estimation accuracy
evaluation is also proposed which provides another perspective for performance
evaluation under varying error tolerances. We tested our approach via a series
of extensive experiments using multimodal sensing technologies to collect data
from 102 subjects while lying on a bed. The individual's high resolution
contact pressure data could be estimated from their RGB or long wavelength
infrared (LWIR) images with 91.8% and 91.2% estimation accuracies in
$PCS_{efs0.1}$ criteria, superior to state-of-the-art methods in the related
image regression/translation tasks. |
---|---|
DOI: | 10.48550/arxiv.2201.11828 |