论文标题
生成用于训练深度学习的合成摄影数据,基于3D点云分割模型
Generating synthetic photogrammetric data for training deep learning based 3D point cloud segmentation models
论文作者
论文摘要
在I/ITSEC 2019上,作者提出了一个完全自动化的工作流程,以段3D摄影测量点云/网格和提取对象信息,包括各个树的位置和地面材料(Chen等,2019)。最终目标是创建现实的虚拟环境,并为仿真提供必要的信息。我们使用在美国陆军的One World Terrain(OWT)项目下创建的数据库测试了先前提出的框架的概括性,并具有各种景观(即各种建筑风格,植被类型和城市密度)以及不同的数据质量(即图像之间的飞行高度和重叠)。尽管数据库比现有数据库大得多,但由于目前缺乏相当大的培训和验证数据集,深入学习算法是否确实在准确性方面确实实现了全部潜力。获得大型注释的3D点云数据库是耗时且劳动密集型的,不仅是从数据注释的角度来看,必须从训练有素的人员手动标记数据,而且还从原始数据收集和处理的角度来看。此外,分割模型通常很难区分对象,例如建筑物和树木质量,这些类型的方案并不总是存在于收集的数据集中。因此,这项研究的目的是使用合成摄影测量数据研究在训练深度学习算法中替代现实世界数据。我们已经研究了生成基于合成无人机的摄影数据数据的方法,以提供足够尺寸的数据库,用于训练深度学习算法,并能够扩大深度学习模型难度的方案的数据大小。
At I/ITSEC 2019, the authors presented a fully-automated workflow to segment 3D photogrammetric point-clouds/meshes and extract object information, including individual tree locations and ground materials (Chen et al., 2019). The ultimate goal is to create realistic virtual environments and provide the necessary information for simulation. We tested the generalizability of the previously proposed framework using a database created under the U.S. Army's One World Terrain (OWT) project with a variety of landscapes (i.e., various buildings styles, types of vegetation, and urban density) and different data qualities (i.e., flight altitudes and overlap between images). Although the database is considerably larger than existing databases, it remains unknown whether deep-learning algorithms have truly achieved their full potential in terms of accuracy, as sizable data sets for training and validation are currently lacking. Obtaining large annotated 3D point-cloud databases is time-consuming and labor-intensive, not only from a data annotation perspective in which the data must be manually labeled by well-trained personnel, but also from a raw data collection and processing perspective. Furthermore, it is generally difficult for segmentation models to differentiate objects, such as buildings and tree masses, and these types of scenarios do not always exist in the collected data set. Thus, the objective of this study is to investigate using synthetic photogrammetric data to substitute real-world data in training deep-learning algorithms. We have investigated methods for generating synthetic UAV-based photogrammetric data to provide a sufficiently sized database for training a deep-learning algorithm with the ability to enlarge the data size for scenarios in which deep-learning models have difficulties.