Human Pose Estimation in Monocular Omnidirectional Top-View Images
Human pose estimation (HPE) with convolutional neural networks (CNNs) for indoor monitoring is one of the major challenges in computer vision. In contrast to HPE in perspective views, an indoor monitoring system can consist of an omnidirectional camera with a field of view of 180{\deg} to detect the...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Human pose estimation (HPE) with convolutional neural networks (CNNs) for
indoor monitoring is one of the major challenges in computer vision. In
contrast to HPE in perspective views, an indoor monitoring system can consist
of an omnidirectional camera with a field of view of 180{\deg} to detect the
pose of a person with only one sensor per room. To recognize human pose, the
detection of keypoints is an essential upstream step. In our work we propose a
new dataset for training and evaluation of CNNs for the task of keypoint
detection in omnidirectional images. The training dataset, THEODORE+, consists
of 50,000 images and is created by a 3D rendering engine, where humans are
randomly walking through an indoor environment. In a dynamically created 3D
scene, persons move randomly with simultaneously moving omnidirectional camera
to generate synthetic RGB images and 2D and 3D ground truth. For evaluation
purposes, the real-world PoseFES dataset with two scenarios and 701 frames with
up to eight persons per scene was captured and annotated. We propose four
training paradigms to finetune or re-train two top-down models in MMPose and
two bottom-up models in CenterNet on THEODORE+. Beside a qualitative evaluation
we report quantitative results. Compared to a COCO pretrained baseline, we
achieve significant improvements especially for top-view scenes on the PoseFES
dataset. Our datasets can be found at
https://www.tu-chemnitz.de/etit/dst/forschung/comp_vision/datasets/index.php.en. |
---|---|
DOI: | 10.48550/arxiv.2304.08186 |