!猫在家里看论文,写论文的日子真爽

!我常常自嘲自己的英文写的很像老太太的裹脚布,又臭又长

!主要是将一些GCN的英文表达方式记录下来,收藏起来慢慢学习

!会给出论文题目,还有一些小小的note

-------------------------------------------------------一条开始认真脸的分界线---------------------------------------------------------

Aspect-based Sentiment Classification with Aspect-specific Graph Convolutional Networks 

1. To tackle this problem, we propose to build a Graph Convolutional Network (GCN) over the dependency tree of a sentence to exploit syntactical information and word dependencies.

注意over 和 exploit 的使用

2. GCN has a multi-layer architecture, with each layer encoding and updating the representation of nodes in the graph using features of immediate neighbors.

注意multi-layer的使用,

以及用with 的使用

这句话常常需要用来表示多层的GCN

3. Furthermore, following the idea of self-looping in Kipf and Welling (2017), each word is manually set adjacent to itself, i.e. the diagonal values of A are all ones.

Following the idea of …

the diagonal values of A are all ones. 对角线为1的矩阵A

set adjacent to itself 设置自链接

4. Experimental results have indicated that GCN brings benefit to the overall performance by leveraging both syntactical infor- mation and long-range word dependencies.

Bing benefit to

Leverage 可以翻译为利用的意思

5. While attention-based models are promising, they are insufficient to capture syntactical dependencies between context words and the aspect within a sentence.

这里描述了attention-based的缺陷,不能充分地捕捉句子的句法依赖,其实还是由于word与word之间距离远,而 导致的,其实也不能完全这么说吧,self attention 会考虑句内所有word的attention,可能能解决一些远距离的信息丢失问题吧。

While 是尽管

SEMI-SUPERVISED CLASSIFICATION WITH GRAPH CONVOLUTIONAL NETWORKS 

1. Our contributions are two-fold. Firstly, we introduce a simple and well-behaved layer-wise prop- agation rule for neural network models which operate directly on graphs and show how it can be motivated from a first-order approximation of spectral graph convolutions (Hammond et al., 2011). Secondly, we demonstrate how this form of a graph-based neural network model can be used for fast and scalable semi-supervised classification of nodes in a graph. Experiments on a number of datasets demonstrate that our model compares favorably both in classification accuracy and effi- ciency (measured in wall-clock time) against state-of-the-art methods for semi-supervised learning.

经典GCN是这样来描述

从本质上讲,GCN 是谱图卷积(spectral graph convolution) 的局部一阶近似(localized first-order approximation)。GCN的另一个特点在于其模型规模会随图中边的数量的增长而线性增长。总的来说,GCN 可以用于对局部图结构与节点特征进行编码。

2. Semantic role labeling (SRL)can be informally described as the task of discovering who did what to whom.

之前在任务定义,形式化时常常会用 is formalized as ……或者是 is define as ……problem

其实也可以使用 is described as the task of    …..被描述为这样….的任务

GRAPH ATTENTION NETWORKS 

1. In its most general formulation, the model allows every node to attend on every other node, dropping all structural information. We inject the graph structure into the mechanism by performing masked attention—we only compute eij for nodes j ∈ Ni, where Ni is some neighborhood of nodei in the graph.

这里介绍了GAT的两种机制,一种是每个节点考虑图中所有节点的影响,这是极端情况,忽略了结构信息。

另外一种则是只考虑节点i领域范围内的节点。

注意表达方式

every node to attend on every other node 来表达节点相互attend的感觉

Drop all structural information. 尤其是drop的使用,这里有比较多normal的词,比如ignore,lose

injectsth into sth by sth 将某种机制,某种结构通过某种方式注入到….

mask attention 这种说法

Neighborhood of node i in the graph

2. By stacking layers in which nodes are able to attend over their neighborhoods’ features, we enable (implicitly) specifying different weights to different nodes in a neighborhood, without requiring any kind of costly matrix operation (such as inversion).

这里的which 是指在stack layers.

nodes are able to attend over their neihborhoods’ features.

specifying different weights to different nodes

Without 的使用

3. However, many interesting tasks involve data that can not be represented in a grid-like structure and that instead lies in an irregular domain.

4. This is the case of 3D meshes, social networks, telecommunication networks, biological networks or brain connectomes. Such data can usually be represented in the form of graphs.

注意表达方式

这段话的常用来的表达的是两种结构

一种是grid-like structure这样的网格结构是可以通过CNN,

还有一种是irregular domain 非规则的,比如社交网络,电信网络等

5. The idea is to compute the hidden representations of each node in the graph, by attending over its neighbors, following a self-attention strategy.

注意表达方式

By attending over its neighbors

Following a self-attention strategy

Attention Guided Graph Convolutional Networks for Relation Extraction 

1. However, how to effectively make use of relevant information while ignoring irrelevant information from the dependency trees remains a challenge research question.

注意表达方式

以how to do sth 作为主语

While 的使用,这里的while 表示同时

然而,如何在有效利用相关信息的同时忽略依赖树中的无关信息,仍然是一个具有挑战性的研究问题

remains a challengng research question , 这里的remain用的好,比 is 表达出了这不仅仅是个问题,还是个遗留问题

2. Intuitively, we develop a “soft pruning” strategy that transforms the original dependency tree into a fully connected edge- weighted graph.

注意表达方式

Intuitively

develop a  strategy that

3. With the help of dense connections, we are able to train the AGGCN model with a large depth, allowing rich local and non-local de- pendency information to be captured.

这一段描述的是dense connections 对网络的作用,虽然都是表达DC能够训练更深的网络,降低过拟合的风险,但是这个with the help of 用的好啊

With the help of

train model with a large depth  这个就比deeper network要高大上的多

local and no-local dependency information

allow的主语是model, 也更客观

model allow sth to be done.

allowing rich local and non-local dependency information to be captured. 其实这里可以借此衍生出很多改写

GCNs are neural networks that operate directly on graph structures

operate的使用

大体上描述GCNs

4. Opposite direction of a dependency arc is also included, which means Aij = 1 and Aji = 1 if there is an edge going from node i to node j,

otherwise Aij = 0 and Aji = 0.

5. Our model can be understood as a soft-pruning approach that automatically learns how to selectively attend to the relevant sub-structures useful for the relation extraction task

Sth can be understood as a ….approach that

how to selectively attend to the relevant sub-structures useful 这里的attend

6. Instead of using rule-based prun- ing, we develop a “soft pruning” strategy in the attention guided layer, which assigns weights to all edges. These weights can be learned by the model in an end-to-end fashion.

纯碎觉的写的又简单又清晰,我有的时候觉得我写的文像老奶奶的裹脚布,又臭又长。

Syntax-Aware Aspect Level Sentiment Classification with Graph Attention Networks 

1. When an aspect term is separated away from its sentiment phrase, it is hard to find the associated sentiment words in a sequence.

描述了为什么我们要将syntax引入到很多nlp任务中,这里指的是apsect-level sentiment classification.

因为通常我们的模型是建立在序列输入上,在序列关系上,有些文本距离一些关键信息距离很远,但是如果将其转换为句法树其实上两者存在直接的关系,这就是为什么要引入syntactic dependencies,因为能从一定程度上降低由于长距离依赖而导致的信息丢失问题。

2. Unlike these previous methods, our approach represents a sentence as a dependency graph instead of a word sequence.

注意表达

其实就是将文本从词序列的结构转换为依赖图的形式

our approach represents A as a B instead of C

我们将A用B来表示而不是用C

3. We employ a multi-layer graph attention network to propagate sentiment features from important syntax neighbourhood words to the aspect target.

注意表达方式

我很喜欢这个propagate的使用方式

employsb to do sth

Propagate the ….from the important syntax neighbourhood words to the aspect target

这句话就很形象地表达了图结构的信息传播方式,沿着图的边将邻居结点的信息聚合起来

Graph Convolution over Pruned Dependency Trees Improves Relation Extraction 

1. To resolve these issues, we further apply a Contextualized GCN (C-GCN) model, where the input word vectors are first fed into a bi-directional long short-term memory (LSTM) network to generate contextualized representations, which are then used as h(0) in the original model.

这里解释了C-GCN, 其实C-GCN很好理解,其实就是在word embedding 和 GCN layer之间插一个Bi-LSTM层(有时也被称为contextualized layer), 现将word embedding 过一遍bi-lstm再输入到gin 中对contextualized features 做propagate.

We note that this relation extraction model is conceptually similar to graph kernel-based models (Zelenko et al., 2003), in that it aims to utilize local dependency tree patterns to inform relation classification.

In that 的使用

is conceptually similar to 在概念上与…..相似

Intuitively, during each graph convolution, each node gathers and summarizes information from its neighboring nodes in the graph.

描述GCN更新结点表示的方式 gathers and summarizes information from its neighboring nodes in the graph

However, naively applying the graph convolution operation in Equation (1) could lead to node representations with drastically different magnitudes, since the degree of a token varies a lot. This could bias our sentence representation towards favoring high-degree nodes regardless of the information carried in the node (see details in Section 2.2).

这段话用来解释在GCN中对邻接矩阵进行归一化的原因。我们用邻接矩阵来表示图结构,但是由于图中结点的差异比较大,直接使用原始的邻接矩阵A会导致结点更偏爱度大的结点。

Degree of a token varies a lot

be towards to doing sth 有利于做…..

  • towards doing 这个用法我不太懂

regardless of the information carried in the node 不管

Densely Connected Convolutional Networks 

DenseNets have several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters. 

这里描述了DC的优点:缓解梯度消息的问题,增强特征传播

大大地减少了参数的数量

Compelling 引入注目的

Vanishing-gradient problem 梯度下降的问题

feature propagation 特征传播

Encourage feature reuse 特征重用,这里是不是可以理解为每层的输入都可能被选择,从而保留下来

substantially, 大大地,基本上;大体上;总的来说

As CNNs become increasingly deep, a new research problem emerges: as information about the input or gradient passes through many layers, it can vanish and “wash out” by the time it reaches the end (or beginning) of the network.

信息会随着网络层数变多,在传播过程中出现消失或者是wash out的现象

  • they create short paths from early layers to later layers 这句话直戳本质,解决信息vanish and “wash out” 的问题。

Although these different approaches vary in network topology and training procedure, they all share a key characteristic: they create short paths from early layers to later layers.

这里指的是为了解决由于网络变深而导致得一系列问题,包括DC,ResNets和Highway Networks都是为了缓解这些问题的

network topology 网络拓扑

share a key characteristic

这句话描述了Dense Connected layer的工作原理

To ensure maximum information flow between layers in the network, we connect all layers (with matching feature-map sizes) directly with each other. To preserve the feed-forward nature, each layer obtains additional inputs from all preceding layers and passes on its own feature-maps to all subsequent layers.

注意表达方式

each layer obtains additional inputs from all preceding layers and passes on its own feature-maps to all subsequent layers.

我觉得这句话写的特别好,点明了DC

每一层从所有之前的层获得额外的输入,并结合子层的要素映射传递到所有后续图层

  • 这句话描述了DenseNets 和 ResNets的不同

Concatenating feature-maps learned by different layers increases variation in the input of subsequent layers and improves efficiency. This constitutes a major difference between DenseNets and ResNets

constitutes a major difference between  A and B 构成了A和B的主要不同(差异)

increases variation 增加变化

There are other notable network architecture innovations which have yielded competitive results.

yielded competitive results 取得了具有竞争力的结果

  • 穿插一个ReNet

We empirically demonstrate DenseNet’s effectiveness on several benchmark datasets and compare with state-of-the- art architectures, especially with ResNet and its variants.

注意表达方式

empirically 经验地

Demonstrate  A’s effectiveness on ****(数据集)and compare with *****(models or methods), especially with ****(models or methods)

DialogueGCN: A Graph Convolutional Neural Network for Emotion Recognition in Conversation 

Emotion recognition in conversation (ERC) has received much attention, lately, from researchers due to its potential widespread applications in diverse areas, such as health-care, education, and human resources.

其实我觉得这个表达感觉很奇怪啊

The current state-of-the-art model in emotion recognition in conversation is (Majumder et al., 2019), where authors introduced a party state and global state based recurrent model for modelling the emotional dynamics.

留意这个句式,如何将目前领域最优的模型进行介绍

Thus, there is always a major interplay between the inter-speaker dependency and self- dependency with respect to the emotional dynamics in the conversation

With respect to 关于,至于

there is always a major interplay between A and B with respect to C

就C而言,在A和B之间存在相互作用

We also represent ui ∈ RDm as the feature representation of the utterance, obtained using the feature extraction process described below.

obtained的使用

using的使用

In theory, RNNs like long short-term memory (LSTM) and GRU should propagate long-term contextual information. However, in practice it is not always the case (Bradbury et al., 2017).

指出了LSTM在理论上可以传播长期上下文信息,但是事实上并未如此。

In theory 理论上

theoretically

in practice it is not always the case 并非总是这样

Encoding Sentences with Graph Convolutional Networks for Semantic Role Labeling 

For example, one can observe that many arcs in the syntactic dependency graph (shown in black below the sentence in Figure 1) are mirrored in the semantic dependency graph.

注意表达方式,这里其实想说语义表示与句法表示密切相关,句法依存树上的很多边,也就是依赖关系可以在语义依赖图上体现出来(语义依赖图是什么图???),这里其实通过这种说法,说明可以通过句法信息来挖掘语义上的联系

A is mirrored in B, A在B中有所体现

We believe that one of the reasons for this radical choice is the lack of simple and effective methods for incorporating syntactic information into sequential neural networks (namely, at the level of words).

Incorporate syntactic information into sequential neural networks.

指出了现在的方法缺少一种将句法信息融合到序列化的神经网络中

at the level of words

Lack of

sequential neural networks 序列化的神经网络中

One layer GCN encodes only information about immediate neighbors and K layers are needed to encode K-order neighborhoods (i.e., information about nodes at most K hops aways).

这里表达的单个GCN只能编码一阶紧邻,如果要编码K阶邻居结点,需要K个GCN

注意表达方式

immediate neighbors

encode K-order neighborhoods

  • This contrasts with recurrent and recursive neural networks which, at least in theory, can capture statistical dependencies across unbounded paths in a trees or in a sequence.
  • 这个句式没看懂

Interestingly, again unlike recursive neural networks, GCNs do not constrain the graph to be a tree.

因为之前很多方法为了获得句法、词法上的信息,通常会使用递归神经网络,但是GCNs 操作的图结果并不强制为树结构。

We believe that there are many applications in NLP, where GCN-based encoders of sentences or even documents can be used to incorporate knowledge about linguistic structures (e.g., representations of syntax, semantics or discourse).

因为这篇论文首次将GCN应用与NLP任务中,在此之前没有任何NLP任务使用GCN,因而作者在这里畅想以GCN为基础作为编码句子或者是文档的编码器,用来融合语言结构()

As in standard convolutional networks (LeCun et al., 2001), by stacking GCN layers one can incorporate higher degree neighborhoods

注意表达方式

incorporate higher degree neighborhoods

Our simplification captures the intuition that information should be propagated differently along edges depending whether this is a head-to-dependent or dependent-to-head edge (i.e., along or opposite the corresponding syntactic arc) and whether it is a self-loop.

我就是我见过表达信息沿图结构中的边传播最好的表达,其实是说图结构中的边是不同的,这里将其归为3类,一类是head->dep, 也就是从句法依存中parse的依赖关系,一类是dep->head,  是head->dep的反方向边,最后一个是self-loop, 其实就是在传播过程自我信息要加以保护的一种做法。

这篇文章认为信息沿边的传播要视边的类型而定,在聚合邻居结点是对不同类型的边学习不同的权重,因而是label-specific parameters

注意表达方式

captures the intuition that 从句 出于****的出发点

be propagated differently along edges 沿边缘以不同方式传播

depends Whether this is A or B and whether it is C,取决于这是否是A或者B,还是C

The inability of GCNs to capture dependencies between nodes far away from each other in the graph may seem like a serious problem, especially

in the context of SRL: paths between predicates and arguments often include many dependency arcs。

这里指出了GCN对捕获距离远的两个结点的inability,并指出了这个在SRL中更为严重,因为谓语和arg之间通常会包含很多依赖边。

However, when graph convolution is performed on top of LSTM states (i.e., LSTM states serve as input x = h(1) vv

to GCN) rather than static word embeddings, GCN may not need to capture more than a couple of hops.

这里针对上面GCN存在的问题,表示如果将GCN建在LSTM之上,也就是说用LSTM的状态作为GCN的输入,而不是静态的word embedding 那么GCN或许

不需要跳很多就能捕获

注意表达方式

The inability of ***(model)to do  sth, 模型在do sth的不足

nodes far away from each other in the graph

Paths between A and B often include many dependency arcs.

a couple of hops.

The classifier predicts semantic roles of words given the predicate while relying on word representations provided by GCN

这里的while 如何理解呢

该分类器根据GCN提供的词表示方法预测给定谓词的词的语义角色

This suggests that extra GCN layers are effective but largely redundant with respect to what LSTMs already capture.

这篇文章通过实验说明了额外的GCN层是有效的,但与LSTMs已经捕获的内容相比,它是大量冗余的,这句话要细细品

GCN 论文英语表达总结相关推荐

  1. 【论文解读】GCN论文总结

    本次要总结和分享的是ICLR2017的关于GCN方面的代表作之一论文:SEMI-SUPERVISED CLASSIFICATION WITH GRAPH CONVOLUTIONAL NETWORKS, ...

  2. 英语表达的收集类游戏

    背景: 大脑图像到具体的英语表达的对应是英语学习的砖瓦.但是训练建立这类反射并不容易,因为我们不能轻易的制造某个大脑图像.代替的办法是用母语文字,图片等的表达来激发对某个大脑图像的产生.这也是不容易的 ...

  3. 计算机专业领域和英语,专业、领域和职业的英语表达法

    句子:我的专业是计算机科学. 误译:My profession is computer science. 正译:1.(我是学计算机科学的.) 1)I major in computer science ...

  4. 外贸人必学的四十个地道英语表达

    "yes-man"."big potato"."a man of a woman"......都是什么意思? 很多从事外贸行业的人都会用Sk ...

  5. rust tech trash是什么_最新热点|那些“垃圾分类”你该知道的相关英语表达

    原标题:最新热点|那些"垃圾分类"你该知道的相关英语表达 朗阁亲爱的小伙伴们 我是小朗 7月1日 也就是前天 这意味着上海进入生活垃圾分类强制时代 史上最严垃圾分类开始了 而上海人 ...

  6. 很傻计算机的英语,用英语表达笨蛋,除了Stupid,还有其他的表达

    原标题:用英语表达笨蛋,除了Stupid,还有其他的表达 Are you stupid? 还单单只会用 fool 或 stupid 表示傻蛋 ?其实表示傻瓜的词很多,一些简单的词组也可以表示,比如 b ...

  7. 天气预报的英语表达方式

    天气预报的英语表达方式: sunny晴天 cloudy多云 overcast阴天 light rain 小雨  moderate rain 中雨heavy rain 大雨 moderate snow中 ...

  8. 上厕所’的英语表达方式

    '上厕所'的英语表达方式 一名中国留学生初到美国,在机场找厕所,问老外:「Where is W.C.?」老外听不懂.一名中国太太到医院生产,洋护士问她:「Did you have a bowel mo ...

  9. 分享一些舒服的英语表达,更有利于成交!

    在交流中,包括与国内客户的交谈和老外的交流之中,我们一直都在学习,学习那种能够让彼此听起来都更舒服的表达,简而言之就是会说话!如果能做一个设身处地站在对方立场上思考的人,这将更有利于你和客户谈生意,也 ...

最新文章

  1. linux 多个定时器,timer: 一个Linux下的超级精简的多重定时器:可实现成千上万个定时任务,定时精度可达纳秒级别,且同一时间点可添加不同的任务!适用于云后台服务和嵌入式等各种环境。...
  2. freemarker -自定义指令
  3. vs添加icon总是变为bitmap_欧赛尔vs勒阿弗尔
  4. CMD执行命令出现NOMALY: meaningless REX prefix used以及IDEA提示Cannot run git问题解决
  5. 四川南充计算机经贸学校,四川南充工贸学校计算机应用专业秋季计划
  6. java逆向框架,Android逆向之逆向框架层
  7. 比CopyMemory还要快的函数SuperCopyMemory
  8. DHCP服务器控制企业成本
  9. Win7(64Bit) 安装 PL/SQL Developer图解
  10. ubuntu的文件路径
  11. WEBMAX官方帮助手册 惊艳登场
  12. c语言简单快速排序原理,快速排序原理、复杂度分析及C语言实现
  13. 23种设计模式11---享元模式
  14. GB28181协议视频流媒体平台中国标设备编号和通道国标编号标记唯一的摄像头|视频|镜头通道
  15. stm32 负数_stm32中编码器模式读出“负数”的问题
  16. [学习笔记]什么是存储过程?
  17. uniapp微信小程序打电话
  18. iOS 如何适应 iPhone 5s/6/6 Plus 三种屏幕的尺寸?
  19. java中this和super是否可以同时使用
  20. java库存信息管理系统_Java商户管理系统 客户管理 库存管理 销售报表 SSM项目源码...

热门文章

  1. C++常见内存错误及解决方案
  2. 如何查看网站的真实服务器ip,牧牛生态科技: ipfs项目到底怎么样?
  3. 【C语言】一些概念的基本解释
  4. 什么是“网络空间安全”?
  5. Theme主题的使用及windowIsTranslucent的理解
  6. Nvidia Xavier Nx平台多路imx274相机视频通路调试记录
  7. css animation 图片65%角旋转,沿着z轴旋转
  8. 自定义 jstl 标签 fns
  9. 微型计算机原理指令改错题,微型计算机原理考试题.docx
  10. LaTeX 中 align 公式环境中插入中文的两种方法