论文标题
同时收集的多模式说谎姿势数据集:在不利视力条件下进行床内的姿势监测
Simultaneously-Collected Multimodal Lying Pose Dataset: Towards In-Bed Human Pose Monitoring under Adverse Vision Conditions
论文作者
论文摘要
计算机视觉(CV)在解释图像的语义含义方面取得了巨大的成功,但是对于具有不良视力条件的任务以及患有数据/标签对的限制的任务,CV算法可能是脆弱的。其中之一是内床的姿势估计,在许多医疗保健应用中具有重要的价值。自然设置中的床内姿势监测可能涉及完整的黑暗或完全遮挡。此外,缺乏公开可用的内BED姿势数据集阻碍了许多成功的姿势估计算法来解决此任务。在本文中,我们介绍了同时收集的多模式说谎姿势(SLP)数据集,其中包括使用多种成像方式捕获的109名参与者的床内姿势图像,包括RGB,长波红外,深度,压力图和压力图。我们还为地面真相姿势标签在极端条件下(例如灯光熄灭并被床单/毯子完全覆盖)提供了一个物理超级参数调整策略。 SLP设计与主流人类姿势数据集兼容,因此,最先进的2D姿势估计模型可以通过SLP数据有效地训练,其表现可有望高达95%,[email protected]在单个方式上。通过通过协作包括其他方式,可以进一步提高姿势估计绩效。
Computer vision (CV) has achieved great success in interpreting semantic meanings from images, yet CV algorithms can be brittle for tasks with adverse vision conditions and the ones suffering from data/label pair limitation. One of this tasks is in-bed human pose estimation, which has significant values in many healthcare applications. In-bed pose monitoring in natural settings could involve complete darkness or full occlusion. Furthermore, the lack of publicly available in-bed pose datasets hinders the use of many successful pose estimation algorithms for this task. In this paper, we introduce our Simultaneously-collected multimodal Lying Pose (SLP) dataset, which includes in-bed pose images from 109 participants captured using multiple imaging modalities including RGB, long wave infrared, depth, and pressure map. We also present a physical hyper parameter tuning strategy for ground truth pose label generation under extreme conditions such as lights off and being fully covered by a sheet/blanket. SLP design is compatible with the mainstream human pose datasets, therefore, the state-of-the-art 2D pose estimation models can be trained effectively with SLP data with promising performance as high as 95% at [email protected] on a single modality. The pose estimation performance can be further improved by including additional modalities through collaboration.