文章目录

  • 往期文章链接目录
  • Information Extraction v.s. Relation Extraction
  • Existing Works of RE
    • Pattern-based Methods
    • Statistical Relation Extraction Models
    • Neural Relation Extraction Methods
  • Future Directions
    • Utilizing More Data
      • Methods to Denoise DS Data
      • Open Problem for Utilizing More Data
    • Performing More Efficient Learning
      • Few-shot Learning
    • Handling More Complicated Context
    • Orienting More Open Domains
  • 往期文章链接目录

往期文章链接目录

Information Extraction v.s. Relation Extraction

Information Extraction: Information extraction is the task of automatically extracting structured information from unstructured and/or semi-structured machine-readable documents and other electronically represented sources.

Relation extraction (RE) is an important task in IE. It focuses on extracting relations between entities. A complete relation RE system consists of

  1. a named entity recognizer to identify named entities from text.
  2. an entity linker to link entities to existing knowledge graphs.
  3. a relational classifier to determine relations between entities by given context (most difficult and important).

Existing Works of RE

RE methods follows the typical supervised setting, from early pattern-based methods, statistical approaches, to recent neural models.

Pattern-based Methods

The early methods use sentence analysis tools to identify syntactic elements in text, then automatically construct pattern rules from these elements. Later work involves larger corpora, more formats of patterns and more efficient ways of extraction.

Statistical Relation Extraction Models

Statistical Methods requires less human efforts so that statistical relation extraction (SRE) has been extensively studied. Approaches includes:

  • feature-based methods which design lexical, syntactic and semantic features for entity pairs and their corresponding context (hard to design features).
  • kernel-based methods measures the similarities between relation representations and textual instances (hard to design kernel functions).
  • Graphical methods abstract the dependencies between entities, text and relations in the form of directed acyclic graphs, and then use inference models to identify the correct relations (still limited model capacities).
  • embedding models. Encode text into low-dimensional semantic spaces and extract relations from textual embeddings. E.g. Knowledge Graph (KG) embeddings.

Neural Relation Extraction Methods

The performance of SOTA RE models. The adoption of neural models began in 2013.

Neural relation extraction (NRE) models can effectively capture textual information and generalize to wider range of data. NRE mainly utilizes both word embeddings and positional embeddings and focus on designing and utilizing various network architectures to capture the relational semantics within text. Methods includes:

  • Recursive Neural Networks. Learn compositional representations for sentences recursively.
  • Convolutional neural networks (CNNs). Model local textual patterns.
  • Recurrent neural networks (RNNs). Handle long sequential data.
  • Graph neural networks (GNNs). Build word/entity graphs for reasoning.
  • Attention-based neural networks. Aggregate global relational information.

Currently, Transformers and Pre-trained LM models achieves SOTA on intra-sentence RE.

Future Directions

Despite the success of existing RE methods, most of them still work in a simplified setting. These methods mainly focus on training models with large amounts of human annotations to classify two given entities within one sentence into pre-defined relations. However, the real world is much more complicated than this simple setting:

  • collecting high-quality human annotations is expensive and time-consuming.
  • many long-tail relations cannot provide large amounts of training examples.
  • most facts are expressed by long context consisting of multiple sentences.
  • using a pre-defined set to cover those relations with open-ended growth is difficult.

Utilizing More Data

The researchers have formed a consensus that utilizing more data is a potential way towards more powerful RE models.

Distant supervision (DS) assumption has been used to automatically label data by aligning existing KGs with plain text. For any entity pair in KGs, sentences mentioning both the entities will be labeled with their corresponding relations in KGs.

Here is an illustration of DS relation extraction. With the fact (Apple Inc., product, iPhone), DS finds all sentences mentioning the two entities and annotates them with the relation product, which inevitably brings noise labels.

Methods to Denoise DS Data

  • Adopt multi-instance learning by combining sentences with same entity pairs and then selecting informative instances from them.
  • Incorporating extra context information such as KGs.
  • Utilize sophisticated mechanisms and training strategies to enhance distantly supervised NRE models.

Open Problem for Utilizing More Data

  • Existing DS methods focus on denoising auto-labeled instances. Explore better DS schemes is valuable.
  • Perform unsupervised or semi-supervised learning for utilizing large-scale unlabeled data as well as using knowledge from KGs and introducing human experts in the loop.

Performing More Efficient Learning

Real-world relation distributions are long-tail and most relations have very limited relational facts and corresponding sentences. We can see the long tail distributions from two DS datasets in the following figure:

Few-shot Learning

Few-shot learning is a good fit for learning long-tail relations efficiently. It trains good representations of instances or learns ways of fast adaptation from existing large-scale data, and then transfer to new tasks.

A typical few-shot learning setting is the N-way K-shot setting, where models are given N random-sampled new relations, along with K training examples for each relation. Here is an example. Give a few instances for new relation types, few-shot RE models classify query sentences into one of the given relations.

Here are a few challenges for few-shot learning:

  • Few-shot domain adaptation.
  • Few-shot none-of-the-above detection. Conventional few-shot models have the difficulty to form a good representation for the none-of-the-above (NOTA) relation in the N-way K-shot setting. Therefore, it is crucial to study how to identify NOTA instances .
  • Few-shot RE may result in a easy classification task if total amount of relations is small. As the following figure shows, as the number of relations increase, the performance drops.
  • Current models cannot truly understand relations similar in semantics.

Handling More Complicated Context

Most existing methods focus on intra-sentence RE and thus are inadequate for identifying relational facts expressed in a long document. Extracting relations from complicated context is a challenging task requiring reading, memorizing and reasoning for discovering relational facts across multiple sentences.

Here is a post (Connecting the Dots: Document-level Neural Relation Extraction with Edge-oriented Graphs) I wrote for explaining using GNN on document-level relation extraction task.

Orienting More Open Domains

Our world undergoes growth of relations and it is not possible to pre-specify all relations by human experts. Thus, we need RE systems that do not rely on pre-defined relation schemas and can work in open scenarios. Here are current explorations in handling open relations:

  • Open information extraction (Open IE), which extracts relation phrases and arguments (entities) from text.
  • Relation discovery, which discover unseen relation types from unsupervised data. Here is an example of casting relation discovery as a clustering task.

Note: There are many redundant extracted relations. Normalizing these phrases is crucial for downstream tasks. For example, relations on (Barack Obama, was born in , Honolulu) and (Obama, place of birth, Honolulu) are actually identical. So they should be normalized.


Reference:

  • More Data, More Relations, More Context and More Openness: A Review and Outlook for Relation Extraction. https://arxiv.org/pdf/2004.03186.pdf.
  • Connecting the Dots: Document-level Neural Relation Extraction with Edge-oriented Graphs. https://arxiv.org/pdf/1909.00228.pdf.
  • Distant supervision for relation extraction without labeled data. https://www.aclweb.org/anthology/P09-1113.pdf.

往期文章链接目录

Relation Extraction 关系抽取综述相关推荐

  1. Exploiting the Syntax-Model Consistency for Neural Relation Extraction(关系抽取,语法模型,跨领域关系抽取

    文章目录 1.Introductiony 1.1 神经网络的关系抽取 1.2使用依赖树的关系抽取 1.2.2 依赖树的一般树 1.3 本文做法 2. 相关工作 2.1核方法 2.2 深度学习方法 2. ...

  2. 中文实体、关系抽取工具

    文章目录 1.Information Extraction 信息提取工具 这个是model 2. CoreNLP 3.AllenNLP 1.Information Extraction 信息提取工具 ...

  3. 关系抽取调研-工业界

    关系抽取调研--工业界 目录 1. 任务 1.1. 任务定义 1.2. 数据集 1.3. 评测标准 2. 方法总结 2.1. 基于模板的方法 2.1.1. 基于触发词/字符串 2.1.2. 基于依存句 ...

  4. 文献阅读15-OntoILPER:A logic-based relational learning approach关系抽取,NER+RE

    文章目录 abstract 1.Introduction 2.2. Supervised Machine Learning Approaches to Relation Extraction 2.1. ...

  5. 构建关系抽取的动词源

    2016 ACM | Constructing Linguistic Verb Source for Relation 构建关系抽取的动词源 摘要 在生物文献挖掘中,关系抽取是一个重要的课题.在正则中 ...

  6. 关系抽取论文总结(relation extraction)不断更新

    2000 1.Miller, Scott, et al. "A novel use of statistical parsing to extract information from te ...

  7. {ACL2020}In Layman’s Terms: Semi-Open Relation Extraction from Scientific Texts(关系抽取,半开放关系抽取)

    半开放关系抽取 用FOBIE训练一个narrow IE 训练一个OpenIE 用narrowIE矫正OpenIE 1.Introduction 标注困难. 跨领域标注更难 TRADE-OFF rela ...

  8. 文献阅读课10-Neural Relation Extraction for Knowledge Base Enrichment(提取+嵌入+消歧+规范化联合模型,实体已知,仅关系抽取,多词实体)

    文章目录 Abstract 1.Introduction 2. 相关工作 2.2 Entity-aware Relation Extraction 3.提出的模型 3.1 Solution Frame ...

  9. 论文阅读课4-Long-tail Relation Extraction via Knowledge Graph Embeddings(GCN,关系抽取,2019,远程监督,少样本不平衡,2注意

    文章目录 abstract 1.introduction 2.相关工作 2.1 关系提取 2.2 KG embedding 2.3 GCNN 3. 方法 3.1符号 3.2框架 3.2.1 Insta ...

最新文章

  1. 独家 | 浅谈Python/Pandas中管道的用法
  2. linux下压缩工具总结与使用(参考私房菜)
  3. CASE软件Enterprise Architect简介和使用入门图解
  4. 【STM32】SYSTEM文件夹介绍,delay,sys,usart
  5. 2022-2028年中国石油焦行业运行现状与发展态势展望报告
  6. linux ftp 登录慢,linux中vsftpd登录,上传下载文件慢解决办法linux操作系统 -电脑资料...
  7. html ie8上传图片,图片上传本地预览兼容ie8
  8. python递归排序_Python归并排序(递归实现)
  9. linux内核 address_space 结构
  10. javascript实现的中国象棋单机双人
  11. 搞懂Java代理和反射一篇就够了
  12. 计算机用word做海报,如何使用word文档制作精美海报
  13. Using the URLconf defined in djangolx.urls, Django tried these URL patterns, in this order: 报错
  14. 工作中遇到的小技巧 一(暂停更新)
  15. 公众号H5页面接入微信登录流程
  16. C#大恒相机采集图片时图片上下对称折叠了
  17. Android获取汉字首字母工具类
  18. php pdo 支持mysql表类型_全新的PDO数据库操作类php版(仅适用Mysql)
  19. LRE板卡交换机在轨交CBCT系统中的应用
  20. Eclipse替换和查找快捷键

热门文章

  1. 学计算机的会重装系统吗,人人都能学会重装系统,你还在抱着电脑去实体店重装系统吗?...
  2. 世界上50大开心的事
  3. linux编译taglib,在Windows上将TagLib编译到Qt C项目中
  4. CPC (Contrastive Predictive Coding)
  5. JFlow手机端开发说明与案例
  6. 《李天命的思考艺术》笔记
  7. seo模拟点击软件_哪些SEO排名工具是有效的呢?
  8. DISC 免费性格测试题
  9. 火狐浏览器添加脚本(可代替手动做一些操作)
  10. jsp田径运动会管理系统计算机毕业设计