论文标题

几乎没有意见的学习摘要

Few-Shot Learning for Opinion Summarization

论文作者

Bražinskas, Arthur, Lapata, Mirella, Titov, Ivan

论文摘要

意见摘要是反映多个文档中表达的主观信息的文本的自动创建,例如对产品的用户评论。该任务实际上很重要,并且引起了很多关注。但是,由于摘要生产的高成本,缺乏足够大的数据集,足以培训监督模型。取而代之的是,该任务传统上是通过提取方法来完成的,这些方法学会以无监督或弱监督的方式选择文本片段。最近,已经表明,抽象性摘要可能会更加流利,并且可以更好地反映相互矛盾的信息,也可以以无监督的方式产生。但是,这些模型没有暴露于实际摘要,无法捕获其基本属性。在这项工作中,我们表明,即使少数摘要也足以引导具有所有预期属性的摘要文本,例如写作风格,信息性,流利性和情感保存。我们首先训练有条件的变压器语言模型,以生成新产品评论,鉴于该产品的其他可用评论。该模型还基于与摘要直接相关的审核属性;这些属性是从没有手动努力的评论中得出的。在第二阶段,我们微调一个插件模块,该模块学会了在少数摘要中预测属性值。这使我们可以将生成器切换到汇总模式。我们在亚马逊和Yelp数据集上显示,我们的方法在自动和人类评估中大大优于以前的提取和抽象方法。

Opinion summarization is the automatic creation of text reflecting subjective information expressed in multiple documents, such as user reviews of a product. The task is practically important and has attracted a lot of attention. However, due to the high cost of summary production, datasets large enough for training supervised models are lacking. Instead, the task has been traditionally approached with extractive methods that learn to select text fragments in an unsupervised or weakly-supervised way. Recently, it has been shown that abstractive summaries, potentially more fluent and better at reflecting conflicting information, can also be produced in an unsupervised fashion. However, these models, not being exposed to actual summaries, fail to capture their essential properties. In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text with all expected properties, such as writing style, informativeness, fluency, and sentiment preservation. We start by training a conditional Transformer language model to generate a new product review given other available reviews of the product. The model is also conditioned on review properties that are directly related to summaries; the properties are derived from reviews with no manual effort. In the second stage, we fine-tune a plug-in module that learns to predict property values on a handful of summaries. This lets us switch the generator to the summarization mode. We show on Amazon and Yelp datasets that our approach substantially outperforms previous extractive and abstractive methods in automatic and human evaluation.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源