想看原文戳这里

1. introduction

(1) the definition of KG

A KG is a multi-relational graph composed of entities (nodes) and relations (different types of edges).

(2) examples

Freebase, DBpedia, YAGO, NELL

(3) key idea of KGE

The key idea is to embed components of a KG including entities and relations into continuous vector spaces, so as to simplify the manipulation while preserving the inherent structure of the KG.

(4) traditional procedure of KGE

Given a KG, such a technique first represents entities and relations in a continuous vector space, and defines a scoring function on each fact to measure its plausibility. Entity and relation embeddings can then be obtained by maximizing the total plausibility of observed facts.

2. notations

3. KG Embedding with Facts Alone

(1) facts

Facts observed in the KG are stored as a collection of triples D+={(h,r,t)}. Each triple is composed of a head entity h∈E, a tail entity t∈E, and a relation r∈R between them.

(2) steps

① representing entities and relations

Entities are usually represented as vectors, i.e., deterministic points in the vector space. Relations are typically taken as operations in the vector space.

② defining a scoring function

A scoring function fr(h,t) is defined on each fact (h,r,t) to measure its plausibility. Facts observed in the KG tend to have higher scores than those that have not been observed.

③ learning entity and relation representations

The third step solves an optimization problem that maximizes the total plausibility of observed facts.

(3) embedding techniques

① Translational Distance Models

exploit distance-based scoring functions; measure the plausibility of a fact as the distance between the two entities, usually after a translation carried out by the relation

(i) TransE

-represents both entities and relations as vectors in the same space, say Rd

-Given a fact (h,r,t), the relation is interpreted as a translation vector r so that the embedded entities h and t can be connected by r with low error, i.e., h+r≈t when (h,r,t) holds.

-

-TransE has flaws in dealing with 1-to-N, N-to-1, and N-to-N relations : given a 1-to-N relation, e.g., DirectorOf, TransE might learn very similar vector representations for Psycho, Rebecca, and RearWindow which are all films directed by AlfredHitchcock, even though they are totally different entities

(ii) TransE’s Extensions

Relation-Specific Entity Embeddings: allow an entity to have distinct representations when involved in different relations

- TransH

* relation-specific hyperplanes

* models entities as vectors

* models each relation r as a vector r on a hyperplane with wr as the normal vector

*         

* By introducing the mechanism of projecting to relation-specific hyperplanes, TransH enables different roles of an entity in different relations.

- TransR

* relation-specific spaces

* entities are represented as vectors in an entity space Rd, and each relation is associated with a specific space Rk and modeled as a translation vector in that space

* Mr∈Rk×d is a projection matrix from the entity space to the relation space of r

*              

- TransD

* simplifies TransR

*       

- TranSparse (share)

* simplifies TransR

*

- TranSparse (separate)

* simplifies TransR

*

Relaxing Translational Requirement: relaxing the overstrict requirement of h+r≈t

- TransM

*

- ManifoldE

*

- TransF

* only requires t to lie in the same direction with h+r, and meanwhile h in the same direction with t-r

*

- TransA

*

(iii) Gaussian Embeddings

take into account entities’ uncertainties, and model them as random variables

- KG2E

*

*

*

- TransG

*

*

*

(iv) Other Distance Models

- UM (unstructured model)

* a naive version of TransE by setting all r=0

- SE (structured embedding)

*

② Semantic Matching Models

exploit similarity-based scoring functions; measure plausibility of facts by matching latent semantics of entities and relations

(i) RESCAL and Its Extensions

- RESCAL

*

- TATEC

* models not only the three-way interaction h⊤Mrt but also two-way interactions, e.g., those between an entity and a relation

*

- DistMult

* this over-simplified model can only deal with symmetric relations

*

- HolE

* combines the expressive power of RESCAL with the efficiency and simplicity of DistMult

*

*

- ComplEx

* extends DistMult by introducing complex-valued embeddings so as to better model asymmetric relations

* entity and relation embeddings h,r,t no longer lie in a real space but a complex space

* every ComplEx has an equivalent HolE

*

- ANALOGY

* extends RESCAL so as to further model the analogical properties of entities and relations

*

(ii) Matching with Neural Networks

- Semantic Matching Energy (SME)

- Neural Tensor Network (NTN)

- Multi-Layer Perceptron (MLP)

- Neural Association Model (NAM)

③ Model Training

(i) Training under Open World Assumption

- KGs contain only true facts and non-observed facts can be either false or just missing. D+ stores only positive examples.

- Given the positive set D+ and a negative set D- constructed accordingly, we can learn entity and relation representations Θ by minimizing the logistic loss/ pairwise ranking loss.

- logistic loss :      

- pairwise ranking loss :

- Minimizing the pairwise ranking loss has an additional advantage: it does not assume that negative examples are necessarily false, just that they are more invalid than those positive ones.

- the logistic loss generally yields better results for semantic matching models such as DistMult and ComplEx, while the pairwise ranking loss may be more suitable for translational distance models like TransE

- steps

* Initializing Entity and Relation Embeddings

~ usually initialized randomly from uniform distributions or Gaussian distributions

~ also initialized with results of simple models such as TransE

~ also represent an entity as the average word vector of its name or description, and initialize word vectors with those pre-trained on a text corpus

* Generating Negative Training Examples

~

~

~ To reduce false-negative examples, (i) set different probabilities for replacing the head and the tail, i.e., to give more chance to replacing the head if the relation is 1-to-N and the tail if the relation is N-to-1 or (ii) given a positive fact, it corrupts a position (i.e., head or tail) using only entities that have appeared in that position with the same relation

~ 50 negatives per positive example is a good trade-off between accuracy and training time

(ii) Training under Closed World Assumption

- all facts that are not contained in D+ are false

- squared loss :

- other possible loss functions include the logistic loss and absolute loss

④ Model Comparison

⑤ Other Approaches

learn representations for head-tail entity pairs rather than individual entities

4. Incorporating Additional Information

(1) Entity Types

① a straight forward method : take IsA as an ordinary relation and the corresponding triples as ordinary training examples

② semantically smooth embedding (SSE)

③ type-embodied knowledge representation learning (TKRL)

(i) can handle hierarchical entity categories and multiple category labels

(ii) a translational distance model with type-specific entity projections

(2) Relation Paths

A relation path is typically defined as a sequence of relations r1→⋯→rℓ through which two entities can be connected on the graph. A key challenge then is how to represent such paths in the same vector space along with entities and relations.

① a straight forward method : represent a path as a composition of the representations of its constituent relations; typical composition operations include addition, multiplication, and recurrent neural network (RNN)

② path-based TransE (PTransE)

(3) Textual Descriptions

① align the given KG with an auxiliary text corpus, and then jointly conduct KG embedding and word embedding

② description-embodied knowledge representation learning (DKRL)

③ a text-enhanced KG embedding model, referred to as TEKE

(4) Logical Rules

systems such as WARMR, ALEPH, and AMIE can extract logical rules automatically from KGs

① utilize rules to refine embedding models during KG completion

② a joint model which embeds KG facts and logical rules simultaneously; a key ingredient of this approach, called KALE, is to represent and model facts and rules in a unified framework

③ vector embeddings are introduced for entity pairs rather than individual entities, making it particularly useful for relation extraction

(5) Other Information

Entity Attributes; Temporal Information; Graph Structures; Evidence from Other Relational Learning Methods

5. Applications in Downstream Tasks

…待补充

6. Concluding Remarks

…待补充

Knowledge Graph Embedding: A Survey of Approaches and Applications 摘要相关推荐

  1. Knowledge Graph Embedding: A Survey of Approaches and Applications (2)

    本文是对知识图谱相关文章阅读翻译,<Knowledge Graph Embedding: A Survey of Approaches and Applications>只针对文字部分,图 ...

  2. Knowledge Graph Embedding: A Survey of Approaches and Applications【翻译】

    原文地址:https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=8047276 论文名称:Knowledge Graph Embed ...

  3. 论文阅读《Knowledge Graph Refinement: A Survey of Approaches and Evaluation Methods》

    论文链接 一.谷歌的知识图谱 谷歌的知识图谱是在2012年向公众推出的,也是在这个时候,"知识图谱"这个术语被创造出来. 谷歌本身对于知识图谱的构建是相当保密的:只有少数外部来源讨 ...

  4. 《knowledge graph embedding:a survey of approaches and applications》论文阅读

    发表于TKDE 2017. knowledge graph embedding:a survey of approaches and applications abstract 1. introduc ...

  5. 【实体对齐·BootEA】Bootstrapping Entity Alignment with Knowledge Graph Embedding

    文章目录 0.总结 1.动机 2. 贡献+方法 3.应用场景 4.其他模型 5.数据集 6.效果 以下的是组内比较 BootEA: "Bootstrapping Entity Alignme ...

  6. 学习Knowledge Graph Embedding Based Question Answering代码笔记

    前言 最近被导师安排学习一下[Knowledge Graph Embedding Based Question Answering] 这篇paper,这篇paper的重点在于运用了Knowledge ...

  7. AutoSF- Searching Scoring Functions for Knowledge Graph Embedding

    AutoSF: Searching Scoring Functions for Knowledge Graph Embedding Abstract 衡量知识图(KG)中三元组合理性的得分函数(SFs ...

  8. 知识图谱-第三方工具:LibKGE(用于Knowledge Graph Embedding)【包含多种模型:TransE、DistMult、ComplEx、ConvE、Transformer等】

    用于Knowledge Graph Embedding的向量表示库有很多,比如: LibKGE GraphVite AmpliGraph OpenKE pykeen pykg2vec LibKGE的主 ...

  9. 知识图谱论文阅读(十三)【2020 arXiv】Attentive Knowledge Graph Embedding for Personalized Recommendation

    题目: Attentive Knowledge Graph Embedding for Personalized Recommendation 论文链接: 代码链接: 知识图谱特征学习在推荐系统中的应 ...

最新文章

  1. 【bzoj2081】[Poi2010]Beads Hash
  2. 一条语句获得一年所有月份
  3. 量子叠加态系数_1.2 量子比特
  4. 网络校时(非NTP)
  5. Monent.js:强大的日期处理类库
  6. stm32双串口通信_一文读懂,基于 STM32 和 CAN 总线的温度监控系统的设计方法
  7. Q:How to read attribute from a tag
  8. Haswell 事务内存相关的一些资源介绍
  9. arping 本机出现 timeout 问题
  10. matlab 求最大值函数,利用matlab, 二元函数求最大值
  11. 魔力岛服务器稳定吗,魔力岛《飘渺仙剑》寻找GM你愉我乐线上活动
  12. 润乾报表简单报表制作流程(胎盘级教程)
  13. 自定义线程池拒绝策略缓解高并发下线程池压力
  14. 【Silvaco example】Temperature Ramping - Effect on Leakage
  15. 图像分割之常用损失函数-Dice Loss
  16. 小白用Python抓取豆瓣高评分喜剧电影
  17. python多线程批量过滤文件关键字
  18. Windows Vista正式版何时提供下载
  19. 怎么给自己的html网页加个密码,密码正确才能显示网页
  20. python自带的框架是什么_Python Django框架是什么?Python学习入门!

热门文章

  1. 汽车以太网- MAC地址
  2. 服务器重启之后,V2无法使用,查看日志出现drained connection > proxy/vmess/encoding: invalid user:
  3. 创智播客微服务_周立-Spring Cloud微服务实战视频课程
  4. 苹果公司:2022年,或将推出更便宜的、大屏的iPhone14
  5. cas集成oauth,用新浪微博账号登录示例
  6. python unicode 标点范围_不同语言Unicode的编码范围
  7. 左连接,外连接,内连接的区别是什么,以及默认连接是什么
  8. 手机怎么从网上下载简历模板?如何在手机上做简历?
  9. nyoj737 石子合并
  10. 官宣 | iPayLinks艾贝盈 跨税云达成合作,为跨境企业健康出海保驾护航