论文标题
通过3D感知的全球对应学习进行硬置虚拟的尝试
Towards Hard-pose Virtual Try-on via 3D-aware Global Correspondence Learning
论文作者
论文摘要
在本文中,我们针对基于图像的人对人的虚拟试验,在存在各种姿势和较大的观点变化的情况下。现有方法在这种情况下受到限制,因为它们估计了服装扭曲流量主要基于2D姿势和外观,这省略了3D人体形状的几何图。此外,当前的服装翘曲方法仅限于局部区域,这使得它们无效地捕获长期依赖性,并导致伪像的下部流动。为了解决这些问题,我们提出了3D感知的全球对应关系,它们是可靠的流量,共同编码3D人体的全球语义相关性,局部变形和几何学先验。特别是,给定描绘源和目标人的图像对,(a)我们首先通过两个编码器获得其姿势感知和高级表示形式,并引入具有多个细化模块的粗到精细解码器,以预测像素的全局对应关系。 (b)从图像推断出的3D参数人模型被合并为正规化对应性改进过程的先验,以便我们的流量可以是3D感知的,并且可以更好地处理姿势和观点的变化。 (c)最后,一个对抗发电机将3D感知流程的服装以及目标人的图像作为输入,以综合光合现实的尝试效果。对公共基准和我们的硬孔测试集进行了广泛的实验,证明了我们方法与SOTA Try-On方法的优越性。
In this paper, we target image-based person-to-person virtual try-on in the presence of diverse poses and large viewpoint variations. Existing methods are restricted in this setting as they estimate garment warping flows mainly based on 2D poses and appearance, which omits the geometric prior of the 3D human body shape. Moreover, current garment warping methods are confined to localized regions, which makes them ineffective in capturing long-range dependencies and results in inferior flows with artifacts. To tackle these issues, we present 3D-aware global correspondences, which are reliable flows that jointly encode global semantic correlations, local deformations, and geometric priors of 3D human bodies. Particularly, given an image pair depicting the source and target person, (a) we first obtain their pose-aware and high-level representations via two encoders, and introduce a coarse-to-fine decoder with multiple refinement modules to predict the pixel-wise global correspondence. (b) 3D parametric human models inferred from images are incorporated as priors to regularize the correspondence refinement process so that our flows can be 3D-aware and better handle variations of pose and viewpoint. (c) Finally, an adversarial generator takes the garment warped by the 3D-aware flow, and the image of the target person as inputs, to synthesize the photo-realistic try-on result. Extensive experiments on public benchmarks and our HardPose test set demonstrate the superiority of our method against the SOTA try-on approaches.