论文标题
AI游乐场:用于深度学习的基于不真实的引擎数据消融工具
AI Playground: Unreal Engine-based Data Ablation Tool for Deep Learning
论文作者
论文摘要
机器学习需要数据,但是获取和标记现实世界的数据具有挑战性,昂贵且耗时。更重要的是,几乎不可能在收购后更改实际数据(例如,更改房间的照明),这使得很难衡量数据的特定属性如何影响性能。在本文中,我们介绍了AI Playground(AIP),这是一种开源,虚幻的基于发动机的工具,用于生成和标记虚拟图像数据。使用AIP,在不同条件下(例如,忠诚度,照明等)捕获相同图像并具有不同的地面真理(例如,深度或表面正常值)是微不足道的。 AIP很容易扩展,可以在有或没有代码的情况下使用。为了验证我们提出的工具,我们生成了八个原本相同但不同的照明和忠诚度条件的数据集。然后,我们训练了深度神经网络,以预测(1)深度值,(2)表面正态或(3)对象标签,并评估了每个网络的内部和交叉数据表性能。除其他见解外,我们验证了对不同设置的敏感性与问题有关。我们证实了其他研究的发现,分割模型对忠诚度非常敏感,但我们还发现它们对照明同样敏感。相反,深度和正常估计模型似乎对忠诚度或照明敏感,对图像结构更敏感。最后,我们在两个现实世界数据集上测试了训练有素的深度估计网络,并获得了与仅实际数据培训相当的结果,证实了我们的虚拟环境足以实现现实世界任务。
Machine learning requires data, but acquiring and labeling real-world data is challenging, expensive, and time-consuming. More importantly, it is nearly impossible to alter real data post-acquisition (e.g., change the illumination of a room), making it very difficult to measure how specific properties of the data affect performance. In this paper, we present AI Playground (AIP), an open-source, Unreal Engine-based tool for generating and labeling virtual image data. With AIP, it is trivial to capture the same image under different conditions (e.g., fidelity, lighting, etc.) and with different ground truths (e.g., depth or surface normal values). AIP is easily extendable and can be used with or without code. To validate our proposed tool, we generated eight datasets of otherwise identical but varying lighting and fidelity conditions. We then trained deep neural networks to predict (1) depth values, (2) surface normals, or (3) object labels and assessed each network's intra- and cross-dataset performance. Among other insights, we verified that sensitivity to different settings is problem-dependent. We confirmed the findings of other studies that segmentation models are very sensitive to fidelity, but we also found that they are just as sensitive to lighting. In contrast, depth and normal estimation models seem to be less sensitive to fidelity or lighting and more sensitive to the structure of the image. Finally, we tested our trained depth-estimation networks on two real-world datasets and obtained results comparable to training on real data alone, confirming that our virtual environments are realistic enough for real-world tasks.