论文标题
Langevin冷却域翻译
Langevin Cooling for Domain Translation
论文作者
论文摘要
域翻译是在两个域之间找到对应关系的任务。在无监督的设置下,几种深神网络(DNN)模型,例如Cyclean和跨语性语言模型在此任务上取得了巨大的成功 - - 从两个域中的两个独立的培训数据(没有配对的样品)中学到了域之间的映射。但是,这些方法通常在很大一部分的测试样品上表现不佳。在本文中,我们假设许多这样的失败样本位于边缘 - 相对较低的数据分布区域 - 数据分布,DNN的训练不太良好,并建议执行Langevin动力学以将此类边缘样品带入高密度区域。我们在定性和定量上证明我们的策略称为Langevin冷却(L-Cool),增强了图像翻译和语言翻译任务中最新的方法。
Domain translation is the task of finding correspondence between two domains. Several Deep Neural Network (DNN) models, e.g., CycleGAN and cross-lingual language models, have shown remarkable successes on this task under the unsupervised setting---the mappings between the domains are learned from two independent sets of training data in both domains (without paired samples). However, those methods typically do not perform well on a significant proportion of test samples. In this paper, we hypothesize that many of such unsuccessful samples lie at the fringe---relatively low-density areas---of data distribution, where the DNN was not trained very well, and propose to perform Langevin dynamics to bring such fringe samples towards high density areas. We demonstrate qualitatively and quantitatively that our strategy, called Langevin Cooling (L-Cool), enhances state-of-the-art methods in image translation and language translation tasks.