AMiner平台由清华大学计算机系研发,拥有我国完全自主知识产权。平台包含了超过2.3亿学术论文/专利和1.36亿学者的科技图谱,提供学者评价、专家发现、智能指派、学术地图等科技情报专业化服务。系统2006年上线,吸引了全球220个国家/地区1000多万独立IP访问,数据下载量230万次,年度访问量超过1100万,成为学术搜索和社会网络挖掘研究的重要数据和实验平台。

AMiner平台:https://www.aminer.cn

导语:EMNLP,自然语言处理经验方法会议(Conference on Empirical Methods in Natural Language Processing),是由国际语言学会(ACL)下属的SIGDAT小组主办的自然语言处理领域的顶级国际会议,也是自然语言算法的A类会议。EMNLP2020共审阅论文3359篇,接收754篇,接收率为22.4%。

Transformer由论文《Attention is All You Need》提出,现在是谷歌云TPU推荐的参考模型。论文相关的Tensorflow的代码可以从GitHub获取,其作为Tensor2Tensor包的一部分。哈佛的NLP团队也实现了一个基于PyTorch的版本,并注释该论文。

根据AMiner-EMNLP2020词云图和论文可以看出,Transformer在本次会议中也有许多不凡的工作,下面我们一起看看Transformer主题的相关论文。

1.论文名称:Understanding the Difficulty of Training Transformers
论文链接:https://www.aminer.cn/pub/5e9d72b391e0117173ad2c33?conf=emnlp2020

作者:Liu Liyuan, Liu Xiaodong, Gao Jianfeng, Chen Weizhu, Han Jiawei

简介:

Transformers (Vaswani et al, 2017) have led to a series of breakthroughs in various deep learning tasks (Devlin et al, 2019; Velickovic et al, 2018).
They do not contain recurrent connections and can parallelize all computations in the same layer, improving effectiveness, efficiency, and scalability.
The authors conduct a comprehensive analysis in theoretical and empirical manners to answer the question: what complicates Transformer training

2.论文名称:A Bilingual Generative Transformer for Semantic Sentence Embedding
论文链接:https://www.aminer.cn/pub/5dca89783a55ac77dcb01f30?conf=emnlp2020

作者:Wieting John, Neubig Graham, Berg-Kirkpatrick Taylor

简介:

Learning useful representations of language has been a source of recent success in natural language processing (NLP).
The authors focus on learning semantic sentence embeddings in this paper, which play an important role in many downstream applications.
Since they do not require any labelled data for fine-tuning, sentence embeddings are useful for a variety of problems right out of the box.
Semantic similarity measures have downstream uses such as fine-tuning machine translation systems (Wieting et al, 2019a)

3.论文名称:Calibration of Pre-trained Transformers
论文链接:https://www.aminer.cn/pub/5e7345fd91e011a051ebf819?conf=emnlp2020

作者:Desai Shrey, Durrett Greg

简介:

Neural networks have seen wide adoption but are frequently criticized for being black boxes, offering little insight as to why predictions are made (Benitez et al, 1997; Dayhoff and DeLeo, 2001; Castelvecchi, 2016) and making it difficult to diagnose errors at test-time.
The authors evaluate the calibration of two pre-trained models, BERT (Devlin et al, 2019) and RoBERTa (Liu et al, 2019), on three tasks: natural language inference (Bowman et al, 2015), paraphrase detection (Iyer et al, 2017), and commonsense reasoning (Zellers et al, 2018)
These tasks represent standard evaluation settings for pretrained models, and critically, challenging out-ofdomain test datasets are available for each.
Such test data allows them to measure calibration in more realistic settings where samples stem from a dissimilar input distribution, which is exactly the scenario where the authors hope a well-calibrated model would avoid making confident yet incorrect predictions

4.论文名称:Attention is Not Only a Weight: Analyzing Transformers with Vector Norms.
论文链接:https://www.aminer.cn/pub/5f7fe6d80205f07f68973153?conf=emnlp2020

作者:Goro Kobayashi, Tatsuki Kuribayashi, Sho Yokoi, Kentaro Inui

简介:

Transformers (Vaswani et al, 2017; Devlin et al, 2019; Yang et al, 2019; Liu et al, 2019; Lan et al, 2020) have improved the state-of-the-art in a wide range of natural language processing tasks.
The attention mechanism computes an output vector by accumulating relevant information from a sequence of input vectors.
It assigns attention weights to each input, and sums up input vectors based on their weights.
Attention computes each output vector yi ∈ Rd from the corresponding pre-update vector yi ∈ Rd and a sequence of input vectors X = {x1, .

5.论文名称:X-LXMERT: Paint, Caption and Answer Questions with Multi-Modal Transformers
论文链接:https://www.aminer.cn/pub/5f6c762f91e0119671e8597f?conf=emnlp2020

作者:Jaemin Cho, Jiasen Lu, Dustin Schwenk, Hannaneh Hajishirzi, Aniruddha Kembhavi

简介:

The past year has seen a spate of BERT-style (Devlin et al, 2019) transformer-based architectures (Lu et al, 2019; Chen et al, 2019; Li et al, 2019) proposed for vision-and-language tasks
These models are typically pre-trained on large image captioning corpora, extending ideas from masked language modeling to mask both the image and text modalities and produce state of the art results on a variety of vision and language tasks including visual question answering, visual grounding and image retrieval.
LXMERT consists of two types of encoders: single-modality encoders for each modality and a cross-modality encoder using bi-directional cross attention to exchange information and align entities across the modalities

6.论文名称:Towards Reasonably-Sized Character-Level Transformer NMT by Finetuning Subword Systems.
论文链接:https://www.aminer.cn/pub/5f7fe6d80205f07f689733fe?conf=emnlp2020

作者:Jindřich Libovický, Alexander Fraser

简介:

State-of-the-art neural machine translation (NMT) models operate almost end-to-end except for input and output text segmentation.
Training character-level Transformer S2S models (Vaswani et al, 2017) is more complicated because the selfattention size is quadratic in the sequence length.
The authors observe that training a character-level model directly from random initialization suffers from instabilities, often preventing it from converging.
The authors’ character-level models show slightly worse translation quality, but have better robustness towards input noise and better capture morphological phenomena.
The authors’ approach is important because previous approaches have relied on very large transformers, which are out of reach for much of the research community

更详细了解EMNLP2020论文,可以关注公众号或者链接直达EMNLP2020专题,最前沿的研究方向和最全面的论文数据等你来~

EMNLP2020 | 近期必读Transformer精选论文相关推荐

  1. EMNLP2020 | 近期必读Question Answering精选论文

    AMiner平台由清华大学计算机系研发,拥有我国完全自主知识产权.平台包含了超过2.3亿学术论文/专利和1.36亿学者的科技图谱,提供学者评价.专家发现.智能指派.学术地图等科技情报专业化服务.系统2 ...

  2. EMNLP2020 | 近期必读Natural Language Inference精选论文

    **AMiner平台**由清华大学计算机系研发,拥有我国完全自主知识产权.平台包含了超过2.3亿学术论文/专利和1.36亿学者的科技图谱,提供学者评价.专家发现.智能指派.学术地图等科技情报专业化服务 ...

  3. 近期AI领域8篇精选论文(附论文、代码)

    来源:PaperWeekly 本文共3050字,建议阅读7分钟. 本文带你发掘近期8篇AI领域精选论文的亮点和痛点,时刻紧跟 AI 前沿成果. 01 Fast and Accurate Reading ...

  4. 近期必读的6篇 NeurIPS 2019 的零样本学习(Zero-Shot Learning)论文

    近期必读的6篇 NeurIPS 2019 的零样本学习(Zero-Shot Learning)论文 PS:转发自"专知"公众号 [导读]NeurIPS 是全球最受瞩目的AI.机器学 ...

  5. 近期必读的12篇「推荐系统」相关论文

    在碎片化阅读充斥眼球的时代,越来越少的人会去关注每篇论文背后的探索和思考. 在这个栏目里,你会快速 get 每篇精选论文的亮点和痛点,时刻紧跟 AI 前沿成果. 点击本文底部的「阅读原文」即刻加入社区 ...

  6. 计算机视觉(CV)领域Transformer最新论文及资源整理分享

    Transformer由论文<Attention is All You Need>提出,现在是谷歌云TPU推荐的参考模型.Transformer模型最早是用于机器翻译任务,当时达到了SOT ...

  7. 解读 | 2019年10篇计算机视觉精选论文(中)

    导读:2019 年转眼已经接近尾声,我们看到,这一年计算机视觉(CV)领域又诞生了大量出色的论文,提出了许多新颖的架构和方法,进一步提高了视觉系统的感知和生成能力.因此,我们精选了 2019 年十大 ...

  8. 计算机维修知识综述论文,机器学习领域各领域必读经典综述论文整理分享

    原标题:机器学习领域各领域必读经典综述论文整理分享 机器学习是一门多领域交叉学科,涉及概率论.统计学.逼近论.凸分析.算法复杂度理论等多门学科.专门研究计算机怎样模拟或实现人类的学习行为,以获取新的知 ...

  9. ICLR 2020丨微软亚洲研究院精选论文解读

    编者按:在全球疫情影响之下,原计划首次在非洲举行的国际 AI 学术会议 ICLR 2020 将成为第一届完全通过网络远程举行的 ICLR 会议.本文为大家介绍的4篇微软亚洲研究院精选论文分别研究了 B ...

最新文章

  1. 扫个地用得到5TOPS算力?自动驾驶芯片公司这样说
  2. 环形队列PHP,环形队列 - Go语言中文网 - Golang中文社区
  3. 稳健+成长股池(转载)
  4. 图像的亮度和对比度区别
  5. Linux 下的 AddressSanitizer
  6. python做abaqus后处理_python进行abaqus后处理的二次开发,我想提取odb文件中的nodeSet,结果总提示Keyerror,请各位大神指教。...
  7. 云主机挂载硬盘 - 开机自动挂载 fdisk and parted
  8. python(7)– 类的反射
  9. “Git 不是昙花一现!”
  10. 2020 年百度之星·程序设计大赛 - 测试赛1001 度度熊保护村庄
  11. Luogu1306 斐波那契公约数
  12. 热门手机刷锤子ROM什么样?
  13. Github中文汉化插件—亲测可用
  14. 目标检测论文综述(一)深度卷积神经网络
  15. ZEMAX 中三种设计优化方法
  16. android webview下载附件几种方法
  17. 扛住100亿次请求——如何做一个“有把握”的春晚红包系统”
  18. lilo是什么意思_Lilo_英文名Lilo是什么意思
  19. 电脑IE打开慢的主要原因
  20. Unity3d C#解决transform.LookAt朝向前后翻转问题(含源码)

热门文章

  1. docker pdf 中文版 linux,Docker入门实战手册PDF
  2. Ionic3安装和项目创建
  3. 【unity3d】如何学习unity3d
  4. 图形界面介绍Violation Browser
  5. 机器学习基础:概率和熵
  6. 云服务器安装并配置nginx
  7. this与super
  8. 学会Zynq(2)Zynq-7000处理器的配置详解
  9. 从 sketch 中获取颜色、字体组件内容
  10. Unity Shader 表面着色器边缘光(Rim Lighting)二