论文标题
SCG-NET:用于语义分割的自构建图神经网络
SCG-Net: Self-Constructing Graph Neural Networks for Semantic Segmentation
论文作者
论文摘要
通过利用远程像素像素依赖性来捕获全局上下文表示,已显示出可以改善语义分割性能。但是,如何有效地做到这一点是一个开放的问题,因为当前使用注意方案或非常深的模型来增加模型的视野,从而导致复杂的模型具有较大的内存消耗。受图形神经网络的最新工作的启发,我们提出了直接从图像中学习长距离依赖图的自我构造图(SCG)模块,并使用它来有效地传播上下文信息以改善语义细分。该模块是通过新型的自适应对角线增强方法和由定制的图形重建项和Kullback-Leibler Divergence正则化项组成的各种下限进行了优化的。当合并到神经网络(SCG-NET)中时,在公开可用的ISPRS Potsdam和Vaihingen数据集上以端到端的方式和竞争性能(分别为92.0%和89.8%的平均F1分数(分别为92.0%和89.8%)执行语义分段(分别为92.0%和89.8%)。
Capturing global contextual representations by exploiting long-range pixel-pixel dependencies has shown to improve semantic segmentation performance. However, how to do this efficiently is an open question as current approaches of utilising attention schemes or very deep models to increase the models field of view, result in complex models with large memory consumption. Inspired by recent work on graph neural networks, we propose the Self-Constructing Graph (SCG) module that learns a long-range dependency graph directly from the image and uses it to propagate contextual information efficiently to improve semantic segmentation. The module is optimised via a novel adaptive diagonal enhancement method and a variational lower bound that consists of a customized graph reconstruction term and a Kullback-Leibler divergence regularization term. When incorporated into a neural network (SCG-Net), semantic segmentation is performed in an end-to-end manner and competitive performance (mean F1-scores of 92.0% and 89.8% respectively) on the publicly available ISPRS Potsdam and Vaihingen datasets is achieved, with much fewer parameters, and at a lower computational cost compared to related pure convolutional neural network (CNN) based models.