关于 #今日arXiv精选 

这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者。

Attentive fine-tuning of Transformers for Translation of low-resourced languages @LoResMT 2021

Comment: 10 pages

Link: http://arxiv.org/abs/2108.08556

Abstract

This paper reports the Machine Translation (MT) systems submitted by theIIITT team for the English->Marathi and English->Irish language pairs LoResMT2021 shared task. The task focuses on getting exceptional translations forrather low-resourced languages like Irish and Marathi. We fine-tune IndicTrans,a pretrained multilingual NMT model for English->Marathi, using externalparallel corpus as input for additional training. We have used a pretrainedHelsinki-NLP Opus MT English->Irish model for the latter language pair. Ourapproaches yield relatively promising results on the BLEU metrics. Under theteam name IIITT, our systems ranked 1, 1, and 2 in English->Marathi,Irish->English, and English->Irish, respectively.

Augmenting Slot Values and Contexts for Spoken Language Understanding with Pretrained Models

Comment: Accepted by Interspeech2021

Link: http://arxiv.org/abs/2108.08451

Abstract

Spoken Language Understanding (SLU) is one essential step in building adialogue system. Due to the expensive cost of obtaining the labeled data, SLUsuffers from the data scarcity problem. Therefore, in this paper, we focus ondata augmentation for slot filling task in SLU. To achieve that, we aim atgenerating more diverse data based on existing data. Specifically, we try toexploit the latent language knowledge from pretrained language models byfinetuning them. We propose two strategies for finetuning process: value-basedand context-based augmentation. Experimental results on two public SLU datasetshave shown that compared with existing data augmentation methods, our proposedmethod can generate more diverse sentences and significantly improve theperformance on SLU.

PoinTr: Diverse Point Cloud Completion with Geometry-Aware Transformers

Comment: Accepted to ICCV 2021 (Oral Presentation)

Link: http://arxiv.org/abs/2108.08839

Abstract

Point clouds captured in real-world applications are often incomplete due tothe limited sensor resolution, single viewpoint, and occlusion. Therefore,recovering the complete point clouds from partial ones becomes an indispensabletask in many practical applications. In this paper, we present a new methodthat reformulates point cloud completion as a set-to-set translation problemand design a new model, called PoinTr that adopts a transformer encoder-decoderarchitecture for point cloud completion. By representing the point cloud as aset of unordered groups of points with position embeddings, we convert thepoint cloud to a sequence of point proxies and employ the transformers forpoint cloud generation. To facilitate transformers to better leverage theinductive bias about 3D geometric structures of point clouds, we further devisea geometry-aware block that models the local geometric relationshipsexplicitly. The migration of transformers enables our model to better learnstructural knowledge and preserve detailed information for point cloudcompletion. Furthermore, we propose two more challenging benchmarks with morediverse incomplete point clouds that can better reflect the real-worldscenarios to promote future research. Experimental results show that our methodoutperforms state-of-the-art methods by a large margin on both the newbenchmarks and the existing ones. Code is available athttps://github.com/yuxumin/PoinTr

Do Vision Transformers See Like Convolutional Neural Networks?

Comment: None

Link: http://arxiv.org/abs/2108.08810

Abstract

Convolutional neural networks (CNNs) have so far been the de-facto model forvisual data. Recent work has shown that (Vision) Transformer models (ViT) canachieve comparable or even superior performance on image classification tasks.This raises a central question: how are Vision Transformers solving thesetasks? Are they acting like convolutional networks, or learning entirelydifferent visual representations? Analyzing the internal representationstructure of ViTs and CNNs on image classification benchmarks, we find strikingdifferences between the two architectures, such as ViT having more uniformrepresentations across all layers. We explore how these differences arise,finding crucial roles played by self-attention, which enables early aggregationof global information, and ViT residual connections, which strongly propagatefeatures from lower to higher layers. We study the ramifications for spatiallocalization, demonstrating ViTs successfully preserve input spatialinformation, with noticeable effects from different classification methods.Finally, we study the effect of (pretraining) dataset scale on intermediatefeatures and transfer learning, and conclude with a discussion on connectionsto new architectures such as the MLP-Mixer.

Video Relation Detection via Tracklet based Visual Transformer

Comment: 1st of Video Relation Understanding (VRU) Grand Challenge in ACM  Multimedia 2021

Link: http://arxiv.org/abs/2108.08669

Abstract

Video Visual Relation Detection (VidVRD), has received significant attentionof our community over recent years. In this paper, we apply thestate-of-the-art video object tracklet detection pipeline MEGA and deepSORT togenerate tracklet proposals. Then we perform VidVRD in a tracklet-based mannerwithout any pre-cutting operations. Specifically, we design a tracklet-basedvisual Transformer. It contains a temporal-aware decoder which performs featureinteractions between the tracklets and learnable predicate query embeddings,and finally predicts the relations. Experimental results strongly demonstratethe superiority of our method, which outperforms other methods by a largemargin on the Video Relation Understanding (VRU) Grand Challenge in ACMMultimedia 2021. Codes are released athttps://github.com/Dawn-LX/VidVRD-tracklets.

A Multi-input Multi-output Transformer-based Hybrid Neural Network for Multi-class Privacy Disclosure Detection

Comment: 20 pages

Link: http://arxiv.org/abs/2108.08483

Abstract

The concern regarding users' data privacy has risen to its highest level dueto the massive increase in communication platforms, social networking sites,and greater users' participation in online public discourse. An increasingnumber of people exchange private information via emails, text messages, andsocial media without being aware of the risks and implications. Researchers inthe field of Natural Language Processing (NLP) have concentrated on creatingtools and strategies to identify, categorize, and sanitize private informationin text data since a substantial amount of data is exchanged in textual form.However, most of the detection methods solely rely on the existence ofpre-identified keywords in the text and disregard the inference of theunderlying meaning of the utterance in a specific context. Hence, in somesituations, these tools and algorithms fail to detect disclosure, or theproduced results are miss-classified. In this paper, we propose a multi-input,multi-output hybrid neural network which utilizes transfer-learning,linguistics, and metadata to learn the hidden patterns. Our goal is to betterclassify disclosure/non-disclosure content in terms of the context ofsituation. We trained and evaluated our model on a human-annotated ground truthdataset, containing a total of 5,400 tweets. The results show that the proposedmodel was able to identify privacy disclosure through tweets with an accuracyof 77.4% while classifying the information type of those tweets with animpressive accuracy of 99%, by jointly learning for two separate tasks.

·

今日arXiv精选:Transformer专题论文推荐相关推荐

  1. 今日arXiv精选 | 18篇近期值得关注的Transformer工作

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Differentiable Prompt Makes Pre-traine ...

  2. 今日arXiv精选 | 21篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Efficient Domain Adaptation of Languag ...

  3. 今日arXiv精选 | 46篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Neural Machine Translation Quality and ...

  4. 今日arXiv精选 | 31篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从 arXiv 中精选高质量论文,推送给读者. Analysis of Language Change in Colla ...

  5. 今日arXiv精选 | 12篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. You should evaluate your language mode ...

  6. 今日arXiv精选 | 14篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Effective Sequence-to-Sequence Dialogu ...

  7. 今日arXiv精选 | 14 篇 ICCV 2021 最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. LocTex: Learning Data-Efficient Visual ...

  8. 今日arXiv精选 | 近期必读的5篇Transformers相关论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Fastformer: Additive Attention is All ...

  9. 今日arXiv精选 | 23篇顶会论文:ICASSP / ICCV / CIKM / ICME / AAAI

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. VerbCL: A Dataset of Verbatim Quotes f ...

最新文章

  1. 缺陷检测 | PCB AOI质量检测之自动定位核选取算法
  2. 德州市公共资源交易平台大数据
  3. Windows 下noinstall方式安装 mysql-5.7.5-m15-winx64
  4. python语言格式化输出_Python字符串格式化输出
  5. 简单工厂之简单模型(uml)
  6. spring学习(47):bean的作用域
  7. java 获取 t 的类型_如何获取类型为T的字段的类?
  8. 固高运动控制卡的使用 1 启动程序以及环境(C#版)
  9. 思科、华为、H3C交换机命名规则全收录
  10. 印度十大最好电影排行
  11. MyExcel 2.2.0 版本发布,支持公式导出
  12. python如何设置矩阵_Python Numpy如何操作数组和矩阵
  13. activerecords php,ActiveRecord_帮助文档_Thinkphp手册 | 码农网
  14. 2.1 压力泊松方程(OpenFOAM理论笔记系列)
  15. C++学习方法(经典)
  16. 计算机基础操作测试题,计算机基础操作练习题.pdf
  17. Handle初解,看完你就懂了handle
  18. hive表信息查询:查看表结构、表操作等(转)
  19. 《复杂》作者梅拉妮·米歇尔发文直指AI四大谬论,探究AI几度兴衰背后的根源...
  20. 起点导航系统源码最新V2.6开源可运营版

热门文章

  1. mysql中的逻辑类型如何定义_MYSQL存储过程即常用逻辑知识点总结
  2. 服务器上如何安装两个php网站,服务器安装两个php版本吗
  3. discuz php如何开发,Discuzx2开发标准流程
  4. python-windows环境安装
  5. 省选专练(学习)AC自动机
  6. centos7扩展根分区
  7. SaltStack安装Redis-第十篇
  8. Java之关于面向对象
  9. bootstrap源码分析之Carousel
  10. 贪吃蛇游戏(附源码)