Xmind思维导图:

deep knowledge-aware network(DKN)

  • properties:

    • incorporates knowledge graph representation into news recommendation.
    • a content-based model for CTR(click-through rate) prediction.
  • key components:
    • a multi-channel and word-entity-aligned knowledge-aware convolution neural network(KCNN) that fuses semantic-level and knowledge-level representations of news.
    • an attention module to dynamically aggregate a user’s history with respect to current candidate news.

Motivation

  • news language is highly condensed, full of knowledges and entities and common sense.However, existing methods are unaware of such external knowledge and cannot fully discover latent knowledge-level connections among news.
  • to extract deep logical connections among news, it’s necessary to introduce additional knowledge graph information into news recommendation.

Challenges in news recommendation

  • news article are highly time-sensitive and their relevance expires quickly within a short period. (Section 5.1)

    • Out-of-date news are substituted by newer ones frequently,which make traditional ID-based methods such as collaborative filtering(CF) less effective.
  • people are topic-sensitive in news reading as they are usually interested in multiple specific news categories.(Section 5.5)
    • How to dynamically measure a user’s interest based on his diversified reading history for current candidate news is key.
  • news language is usually highly condensed and comprised of a large amount of knowledge entities and common sense.
    • However, traditional semantic models or topic models can only find their relatedness based on co-occurrence or clustering structure of words, but are hardly able to discover their latent knowledge-level connection.
    • As a result, a user’s reading pattern will be narrowed down to a limited circle and cannot be reasonably extended based on existing recommendation methods.

DKN Framework

  • Knowledge distillation

    • step-1: distinguish knowledge entities in news contest utilizing the technique of entity linking
      notes: to disambiguate mentions in texts by associating them with predefined entities in a knowledge graph.
    • step-2: construct a sub-graph and extract all relational links among these identified entities from the original knowledge graph.
      notes: Note that the relations among identified entities only may be sparse and lack diversity. Therefore,expand the knowledge sub-graph to all entities within one hop of identified ones
    • step-3: knowledge graph embedding methods for entity representation learning
      notes: 1) Given the extracted knowledge graph, a great many knowledge graph embedding methods, such as TransE , TransH, TransR , and TransD, can be utilized for entity representation learning. 2)the goal of knowledge graph embedding is to learn a low-dimensional representation vector for each entity and relation that preserves the structural information of the original knowledge graph. 3) It should be noted that though state-of-the-art knowledge graph embedding methods could generally preserve the structural information in the original graph, we find that the information of learned embedding for a single entity is still limited when used in subsequent recommendations. To help identify the position of entities in the knowledge graph, we propose extracting additional contextual information for each entity.
    • step-4:learned entity embedding
  • Knowledge-aware convolutional neural networks module (KCNN)
    notes: KCNN is used to process each piece of title and generate an embedding vector.
    KCNN treats words and entities as multiple channels, and explicitly keeps their alignment relationship during convolution.

    • propose a multi-channel and word-entity-aligned KCNN for combining word semantics and knowledge information. for each news title, use its word embeddings transformed entity embeddings and transformed context embeddings as source input.
      notes: First, try to combine words and associated entities is to treat the entities as “pseudo words” and concatenate them to the word sequence W.e.g.W = [w1 w2 … wn et1 et2 …]However, we argue that this simple concatenating strategy has the following limitations: 1) The concatenating strategy breaks up the connection between words and associated entities and is unaware of their alignment. 2) Word embeddings and entity embeddings are learned by different methods, meaning it is not suitable to convolute them together in a single vector space.3) The concatenating strategy implicitly forces word embeddings and entity embeddings to have the same dimension, which may not be optimal in practical settings since the optimal dimensions for word and entity embeddings may differ from each other.Since the transformation function is continuous, it can map the entity embeddings and context embeddings from the entity space to the word space while preserving their original spatial relationship.
    • feed the obtained new sentence W into CNN(similar to Kim CNN for sentence representation learning applying multiple filters with varying window sizes to extract specific local patterns in the news title)
  • Attention-based user interest extraction
    notes: Motivation:To get final embedding of the user with respect to the current candidate news, we use an attention-based method to automatically match the candidate news to each piece of his clicked news, and aggregate the user’s historical interests with different weights.

    • concatenate candidate news and use clicked news
      apply a DNN as the attention network and the softmax function to calculate the normalized impact weight.
      the embedding of user i with respect to the candidate news can be calculated as the weight sum of his clicked news title embedding.
  • Calculate the predicted probability with DNN
    • Calculate the predicted probability with DNN

Experiment

  • Dataset: Bing News
  • Estimation: outperform baselines by 2.8% to 17% on F1 and 2.6% to 16.1% on AUC.
  • Conclusion: the usage of knowledge and an attention model can bring additional 3.5% and 1.4% in improvement.

2018_WWW_DKN- Deep Knowledge-Aware Network for News Recommendation阅读笔记相关推荐

  1. DKN: Deep Knowledge-Aware Network for News Recommendation阅读笔记

    这篇论文发表在2018年的WWW上.引入知识来进行新闻推荐. 关键词:News recommendation; knowledge graph representation; deep neural ...

  2. Alleviating the Inconsistency Problem of Applying Graph Neural Network to Fraud Detection阅读笔记

    Alleviating the Inconsistency Problem of Applying Graph Neural Network to Fraud Detection阅读笔记 文章标题:A ...

  3. 25. Deep Weighted MaxSAT for Aspect-based Opinion Extraction 阅读笔记

    25. Deep Weighted MaxSAT for Aspect-based Opinion Extraction 阅读笔记 Author Information::Meixi Wu* , We ...

  4. 《Deep Learning for Computer Vision withPython》阅读笔记-PractitionerBundle(第9 - 11章)

    9.使用HDF5和大数据集 到目前为止,在本书中,我们只使用了能够装入机器主存储器的数据集.对于小数据集来说,这是一个合理的假设--我们只需加载每一个单独的图像,对其进行预处理,并允许其通过我们的网络 ...

  5. 论文解读《Evaluating the visualization of what a Deep Neural Network has learned》–阅读笔记

    本文属于原创,转载请注明出处 *本论文解读的初衷: 1.由于某些原因,最近有关注到神经网络可解释性与可视化方向的发展. 2.本人习惯阅读优秀的博文后直接点赞收藏,而这篇却没有搜到相关解读,不知道是不是 ...

  6. 《Deep Learning for Computer Vision withPython》阅读笔记-StarterBundle(第18 - 23章)

    18.检查点模型 截止到P265页 //2022.1.18日22:14开始学习 在第13章中,我们讨论了如何在培训完成后将模型保存和序列化到磁盘上.在上一章中,我们学习了如何在发生欠拟合和过拟合时发现 ...

  7. EnlightenGAN: Deep Light Enhancement without Paired Supervision论文阅读笔记

    EnlightenGAN: Deep Light Enhancement without Paired Supervision论文解读 Motivation and introduction 最近在x ...

  8. 《Deep Learning for Computer Vision withPython》阅读笔记-StarterBundle(第6 - 7章)

    6.配置您的开发环境 当涉及到学习新技术(尤其是深度学习)时,配置开发环境往往是成功的一半.在不同的操作系统.不同的依赖版本以及实际的库本身之间,配置您自己的深度学习开发环境可能是相当令人头痛的事情. ...

  9. Deep High-Resolution Representation Learning for Visual Recognition阅读笔记

    用于视觉识别的深度高分辨率表示学习 论文链接 摘要: 高分辨率表示对于人体姿态估计.语义分割和目标检测这类位置敏感的视觉问题至关重要.现有的 sota 框架首先通过串联 high-to-low 分辨率 ...

最新文章

  1. SQL里各种联合查询总结
  2. Leetcode-最长回文子串(5)
  3. 生活本该如此——莫言
  4. C# 选择按钮 数字排序 找最大值最小值
  5. OpenJudge计算概论-字符串排序
  6. ecs 云服务器 管理控制台_【弹性计算】教您快速学会云服务器ECS 创建命令!
  7. 聚焦效率与目标差距,数据才是远程办公的内核!
  8. vs qt 在linux运行,QT安装以及使用(QT支持linux和windows,也支持C/C++代码的编译运行,比vs简洁多)...
  9. CentOS 6.5 安装mysql5.5
  10. 如何还原已经推送到远程分支的合并提交?
  11. Ubuntu安装jdk10
  12. Monad B2 For Dotnet Framework 2.0 RC/RTM Released!
  13. 数据人必会的Excel|还在给老板看苍白无力的数据?有了这份Excel炫酷的报表,还愁不升职加薪?
  14. EFCore-脚手架Scaffold发生Build Failed问题的终极解决
  15. 百度热搜词词云图-python
  16. 毛玻璃 / 玻璃拟态 处理 backdrop-filter 兼容性问题
  17. 2023最新最全vscode插件精选
  18. 离线语音蓝牙设计应用案例
  19. 【HTML】HTML网页设计---海贼王网页前端设计
  20. 计算机网络 网络层功能

热门文章

  1. Allegro PCB Design GXL (legacy) - 导出 PCB 中元件的坐标文件
  2. 为什么刀剑总显示服务器繁忙,刀剑乱舞萌新必看攻略 常见问题篇
  3. Linux系统常见目录分类
  4. Django中多对多查询方式
  5. Sql 从一个表往另一个表里插数据
  6. 大数据运维存档(5)HBaseKafka调优与排障
  7. 【PMAC】Chapter4:PMAC的C#开发
  8. 人民网:山西省 - 地方领导留言板
  9. nginx安装、配置文件详解、测试
  10. 64位Ubuntu系统安装Chrome浏览器