论文标题
杂草识别增强的深度数据增强:一种扩散概率模型和基于转移学习的方法
Deep Data Augmentation for Weed Recognition Enhancement: A Diffusion Probabilistic Model and Transfer Learning Based Approach
论文作者
论文摘要
杂草管理在许多现代农业应用中都起着重要作用。常规的杂草控制方法主要依赖化学除草剂或手工除草,这些除草剂通常是成本不适,环境不友好,甚至对食品安全和人类健康构成威胁。最近,使用机器视觉系统的自动化/机器人除草,其研究的注意力增加了,其潜力和个性化的杂草处理可能性。但是,需要专用,大规模且标记的杂草图像数据集以开发出强大而有效的杂草识别系统,但是它们通常很难且昂贵。为了解决这个问题,已经探索了数据增强方法,例如生成对抗网络(GAN),以生成农业应用程序高度现实的图像。然而,尽管有一些进展,但这些方法通常很复杂,无法在图像中保留细节的良好细节。在本文中,我们介绍了采用扩散概率模型(也称为扩散模型)的第一批工作,以基于转移学习来生成高质量的合成杂草图像。全面的实验结果表明,开发的方法始终优于几种最先进的gan模型,这代表了样本保真度和多样性之间的最佳权衡以及在常见的杂草数据集(CottonWeedId15)上的最高差异。此外,使用合成杂草图像的扩展数据集显然可以在杂草分类任务的四个深度学习(DL)模型上提高模型性能。此外,在CottonWeedID15数据集上接受培训的DL模型只有10%的真实图像和90%的合成杂草图像的测试精度超过94%,显示出生成的杂草样品的高质量。这项研究的代码可在https://github.com/dongchen06/dmweeds上公开提供。
Weed management plays an important role in many modern agricultural applications. Conventional weed control methods mainly rely on chemical herbicides or hand weeding, which are often cost-ineffective, environmentally unfriendly, or even posing a threat to food safety and human health. Recently, automated/robotic weeding using machine vision systems has seen increased research attention with its potential for precise and individualized weed treatment. However, dedicated, large-scale, and labeled weed image datasets are required to develop robust and effective weed identification systems but they are often difficult and expensive to obtain. To address this issue, data augmentation approaches, such as generative adversarial networks (GANs), have been explored to generate highly realistic images for agricultural applications. Yet, despite some progress, those approaches are often complicated to train or have difficulties preserving fine details in images. In this paper, we present the first work of applying diffusion probabilistic models (also known as diffusion models) to generate high-quality synthetic weed images based on transfer learning. Comprehensive experimental results show that the developed approach consistently outperforms several state-of-the-art GAN models, representing the best trade-off between sample fidelity and diversity and highest FID score on a common weed dataset, CottonWeedID15. In addition, the expanding dataset with synthetic weed images can apparently boost model performance on four deep learning (DL) models for the weed classification tasks. Furthermore, the DL models trained on CottonWeedID15 dataset with only 10% of real images and 90% of synthetic weed images achieve a testing accuracy of over 94%, showing high-quality of the generated weed samples. The codes of this study are made publicly available at https://github.com/DongChen06/DMWeeds.