关于 #今日arXiv精选 

这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者。

Differentiable Prompt Makes Pre-trained Language Models Better Few-shot  Learners

Category: NLP

Link: https://arxiv.org/abs/2108.13161

Abstract

Large-scale pre-trained language models have contributed significantly to natural language processing. However, their effectiveness depends mainly on scaling the modelparameters and prompt design. This study proposes a novel pluggable, extensible, and efficientapproach named DifferentiAble pRompT.

Shatter: An Efficient Transformer Encoder with Single-Headed  Self-Attention and Relative Sequence Partitioning

Category: NLP

Link: https://arxiv.org/abs/2108.13032

Abstract

Shatter is an alternative self-attention architecture to the popular Transformer architecture. Shatter can be pretrained on GPUs in 7 days, and match the performance of BERT_Base.

ReGen: Reinforcement Learning for Text and Knowledge Base Generation  using Pretrained Language Models

Category: NLP

Link: https://arxiv.org/abs/2108.12472

Abstract

We present ReGen, a bidirectionalgeneration of text and graph leveraging Reinforcement Learning (RL) to improve performance. Our system provides state-of-the-art results on WebNLG+ 2020 and TekGen datasets.

Exploring and Improving Mobile Level Vision Transformers

Category: Computer Vision

Link: https://arxiv.org/abs/2108.13015

Abstract

We study the vision transformer structure in the mobile level in this paper. We find a dramatic performance drop. We propose a novel irregular patch embedding module and adaptive patch merging module to improve the performance.

Multi-Channel Transformer Transducer for Speech Recognition

Category: Machine Learning

Link: https://arxiv.org/abs/2108.12953

Abstract

Multi-channel inputs offer several advantages over single-channel, to improve the robustness of on-device speech recognition systems. Recent work on a multi-channel transformer has proposed a way to incorporate such inputs into end-to-end ASR for improved accuracy.

Making Transformers Solve Compositional Tasks

Category: Artificial Intelligence

Link: https://arxiv.org/abs/2108.04378

Abstract

Several studies have reported the inability of Transformer models to generalize compositionally. This is a key type of generalization in many NLP tasks such as semantic parsing. We show that the inductive biases given to the model by several design decisions significantly impact compositional generalization.

C5T5: Controllable Generation of Organic Molecules with Transformers

Category: Machine Learning

Link: https://arxiv.org/abs/2108.10307

Abstract

C5T5 is a novel self-supervised pretraining method that that that allows transformers to make zero-shot select-and-replace edits, altering substances towards desired property values. It operates on IUPAC names, a standardized molecular representation that intuitively encodes rich structural information

ViTGAN: Training GANs with Vision Transformers

Category: Computer Vision

Link: https://arxiv.org/abs/2107.04589

Abstract

Vision Transformers (ViTs) have shown competitive performance on image recognition while requiring less vision-specific inductive biases. We introduce novel regularization Techniques for training GANs with ViTs. Empirically, our approach achieves comparable performance to state-of-the-art CNN-based Style

Extracting Qualitative Causal Structure with Transformer-Based NLP

Category: Machine Learning

Link: https://arxiv.org/abs/2108.13304

Abstract

In everyday or academic language, we may express interactions between quantities, events or entities. Qualitative causal relationships compactly express the direction, dependency,temporal constraints, and monotonicity constraints of discrete or continuous interactions. This paper presents a transformer-based NLP architecture that jointly identifies and extracts

Leveraging Pre-trained Language Model for Speech Sentiment Analysis

Category: NLP

Link: https://arxiv.org/abs/2106.06598

Abstract

We explore the use of pre-trained language models to learnipientsentiment information of written texts for speech sentiment analysis. Although spoken and written texts have different linguistic characteristics, they can complement each other in understanding sentiment. In these experiments, we demonstrate the proposed approaches can consistentlyimprove F1 scores.

Swin Transformer: Hierarchical Vision Transformer using Shifted Windows

Category: Computer Vision

Link: https://arxiv.org/abs/2103.1403

Abstract

This paper presents a new vision Transformer that can serve as a general-purpose backbone for computer vision. It is compatible with a broad range of vision tasks, including image classification and dense prediction. Its performance surpasses the previous state-of-the-art by a large margin. The code and models will be made publicly available.

GroupFormer: Group Activity Recognition with Clustered Spatial-Temporal  Transformer

Category: Computer Vision

Link: https://arxiv.org/abs/2108.1263

Abstract

Group activity recognition is a crucial yet challenging problem. Previous methods either model spatial and temporal information separately, or directly aggregateindividual features to form group features. We propose a novel group activity recognition network termed GroupFormer.

Vision Transformers with Patch Diversification

Category: Computer Vision

Link: https://arxiv.org/abs/2104.12753

Abstract

Vision transformer has demonstrated promising performance on challenging computer vision tasks. But directly training the vision transformers may give sub-optimal results. We introduce novel loss functions in vision transformer training to encourage diversity across patchrepresentations for more discriminative feature extraction.

The Devil is in the Detail: Simple Tricks Improve Systematic  Generalization of Transformers

Category: Machine Learning

Link: https://arxiv.org/abs/2108.12284

Abstract

Transformers, typically trained with default hyper-parameters from standard tasks, fail dramatically. By revisiting model configurations, we can drasticallyimprove the performance of Transformers on systematic generalization. We report improvements on five popular datasets: SCAN, CFQ, PCFG, COGS, and Mathematics

Geometry-Free View Synthesis: Transformers and no 3D Priors

Category: Computer Vision

Link: https://arxiv.org/abs/2104.07652

Abstract

A transformer-based model can synthesize entirely novel views without any hand-engineered 3Dbiases. This is achieved by (i) a global attention mechanism for implicitlylearning long-range 3D correspondences between source and target views, and (ii) a probabilistic formulation.

Fastformer: Additive Attention is All You Need

Category: NLP

Link: https://arxiv.org/abs/2108.09084

Abstract

Transformer is a powerful model for text understanding. It is inefficient due to its quadratic complexity to input sequence length. In Fastformer, instead of modeling the pair-wise interactionsbetween tokens, we first use additive attention mechanism to model global contexts.

Space-time Mixing Attention for Video Transformer

Category: Computer Vision

Link: https://arxiv.org/abs/2106.05968

Abstract

This paper is on video recognition using Transformers. The complexity of the model scales linearly with the number of frames in the video sequence and induces no overhead compared to an image-based Transformer model.

Span Fine-tuning for Pre-trained Language Models

Category: NLP

Link: https://arxiv.org/abs/2108.12848

Abstract

Pre-trained language models (PrLM) have to carefully manage input units when training on a large text with a vocabulary consisting of millions of words. Previous works have shown that incorporating span-level information over successive words in pre-training could further improve the performance of PrLMs.

·

今日arXiv精选 | 18篇近期值得关注的Transformer工作相关推荐

  1. 今日arXiv精选 | 21篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Efficient Domain Adaptation of Languag ...

  2. 今日arXiv精选 | 28篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Broaden the Vision: Geo-Diverse Visual ...

  3. 今日arXiv精选 | 14 篇 ICCV 2021 最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. LocTex: Learning Data-Efficient Visual ...

  4. 今日arXiv精选 | 29篇顶会论文:ACM MM/ ICCV/ CIKM/ AAAI/ IJCAI

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Group-based Distinctive Image Captioni ...

  5. 今日arXiv精选 | 11篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Does Vision-and-Language Pretraining I ...

  6. 今日arXiv精选 | 13篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Classification-based Quality Estimatio ...

  7. 今日arXiv精选 | 11篇ICCV 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Explain Me the Painting: Multi-Topic K ...

  8. 今日arXiv精选 | 46篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从arXiv中精选高质量论文,推送给读者. Neural Machine Translation Quality and ...

  9. 今日arXiv精选 | 31篇EMNLP 2021最新论文

     关于 #今日arXiv精选  这是「AI 学术前沿」旗下的一档栏目,编辑将每日从 arXiv 中精选高质量论文,推送给读者. Analysis of Language Change in Colla ...

最新文章

  1. RDKit | 基于RDKit获取分子3D距离矩阵
  2. 批处理for中如何实现break
  3. 专访《Javascript设计模式与开发实践》作者曾探:爱编程 爱生活
  4. 如何解决复杂条件下的程序流的控制问题?
  5. Jenkins构建自动化任务
  6. SAP S4HANA里关于生产订单的一些重要数据库表
  7. 旧项目适配iphone6和iphone plus
  8. asp.net 6中的mini api和mvc api性能对比
  9. oracle 12c alert,Oracle 12c DG备库Alert报错ORA-01110
  10. python判断输入的数字是完全平方还是三次方
  11. 银河麒麟4安装MySQL8_银河麒麟4.0.2安装教程-电脑系统安装手册
  12. python怎么执行代码漏洞_如何使用python查找网站漏洞
  13. CvvImage.h和CvvImage.cpp
  14. 7大Python IDE工具推荐
  15. A Magic Lamp
  16. JS(JavaScript)中实现深浅拷贝的几种方式(详细阅读 非常重要)。
  17. Wavesurfer.js 生成音频波形图
  18. html5学习笔记——三、图文展示网页设计
  19. elment文件上传 展示点击下载
  20. Material Design之定制动画--触摸反馈,循环揭露,转场动画,共享元素和曲线运动

热门文章

  1. 判断java日期跨月_18 个 Java8 日期处理的实践,太有用了!
  2. python绘制相关性矩阵_python seaborn heatmap可视化相关性矩阵实例
  3. 树结构遍历节点名字提取,这里提取的是el-tree数据结构,封装成函数
  4. 【Java】Struts2配置默认Action和配置Action的默认处理类
  5. 讲述Sagit.Framework解决:双向引用导致的IOS内存泄漏(上)
  6. MessagePack 学习笔记
  7. [异常解决] android studio检测不到手机的解决办法——ADB驱动自己安装
  8. 如何用堆栈和循环结构代替递归调用--递归转换为非递归的10条军规
  9. 对linux lvm怎么分区,linux下的LVM分区管理
  10. java中鼠标事件_java中检测鼠标事件动作