论文标题
差异化私人语言模型用于安全数据共享
Differentially Private Language Models for Secure Data Sharing
论文作者
论文摘要
为了保护共享数据的个人的隐私,开发方法非常重要,使研究人员和公司可以发布文本数据,同时为其发起人提供正式的隐私保证。在NLP领域,已经大力努力遵循当地差异隐私的框架建立机制,从而在发布之前将单个文本样本匿名化。在实践中,由于当地差异性隐私所需的强烈噪音,这些方法通常对其输出语言的质量不满意。在本文中,我们使用全球差异隐私解决了目前的问题,尤其是通过以差异方式培训生成语言模型,从而从中训练生成语言模型。使用自然语言提示和新的及时不匹配的损失,我们能够创建高度准确,流利的文本数据集,以采用特定所需属性,例如情感或主题,并类似于培训数据的统计属性。我们进行彻底的实验,表明我们的合成数据集不会从我们的原始数据中泄漏信息,并且具有高语言质量,并且非常适合培训模型,以进一步分析现实世界数据。值得注意的是,我们还证明,私人合成数据的培训分类器与DP-SGD上的真实数据直接训练分类器。
To protect the privacy of individuals whose data is being shared, it is of high importance to develop methods allowing researchers and companies to release textual data while providing formal privacy guarantees to its originators. In the field of NLP, substantial efforts have been directed at building mechanisms following the framework of local differential privacy, thereby anonymizing individual text samples before releasing them. In practice, these approaches are often dissatisfying in terms of the quality of their output language due to the strong noise required for local differential privacy. In this paper, we approach the problem at hand using global differential privacy, particularly by training a generative language model in a differentially private manner and consequently sampling data from it. Using natural language prompts and a new prompt-mismatch loss, we are able to create highly accurate and fluent textual datasets taking on specific desired attributes such as sentiment or topic and resembling statistical properties of the training data. We perform thorough experiments indicating that our synthetic datasets do not leak information from our original data and are of high language quality and highly suitable for training models for further analysis on real-world data. Notably, we also demonstrate that training classifiers on private synthetic data outperforms directly training classifiers on real data with DP-SGD.