论文标题

几个嵌套的命名实体识别

Few-Shot Nested Named Entity Recognition

论文作者

Ming, Hong, Yang, Jiaoyun, Jiang, Lili, Pan, Yan, An, Ning

论文摘要

虽然指定的实体识别(NER)是一项经过广泛研究的任务,但对只有少数标记数据的实体推断很具有挑战性,尤其是对于具有嵌套结构的实体而言。与扁平实体不同,实体及其嵌套实体更有可能具有类似的语义特征表示形式,从而在几个射击设置中对不同实体类别进行分类的困难大大增加了困难。尽管据我们所知,但先前的工作已经在几乎没有学习的背景下简要讨论了嵌套的结构,但本文是专门研究几个嵌套的NER任务的第一个论文。利用上下文依赖性以区分嵌套实体,我们提出了一个基于Biaffine的对比学习(BCL)框架。我们首先设计一个Biaffine跨度表示模块,用于学习每个实体跨度的上下文跨度依赖性表示形式,而不仅仅是学习其语义表示。然后,我们通过残差连接合并这两个表示,以区分嵌套实体。最后,我们建立一个对比度学习框架,以调整更大边界和更广泛的域转移学习能力的表示分布。我们对三个英语,德语和俄罗斯嵌套的NER数据集进行了实验研究。结果表明,在F1分数方面,BCL在1-Shot和5-Shot任务上的表现优于三个基线模型。

While Named Entity Recognition (NER) is a widely studied task, making inferences of entities with only a few labeled data has been challenging, especially for entities with nested structures. Unlike flat entities, entities and their nested entities are more likely to have similar semantic feature representations, drastically increasing difficulties in classifying different entity categories in the few-shot setting. Although prior work has briefly discussed nested structures in the context of few-shot learning, to our best knowledge, this paper is the first one specifically dedicated to studying the few-shot nested NER task. Leveraging contextual dependency to distinguish nested entities, we propose a Biaffine-based Contrastive Learning (BCL) framework. We first design a Biaffine span representation module for learning the contextual span dependency representation for each entity span rather than only learning its semantic representation. We then merge these two representations by the residual connection to distinguish nested entities. Finally, we build a contrastive learning framework to adjust the representation distribution for larger margin boundaries and more generalized domain transfer learning ability. We conducted experimental studies on three English, German, and Russian nested NER datasets. The results show that the BCL outperformed three baseline models on the 1-shot and 5-shot tasks in terms of F1 score.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源