© 作者|都一凡

机构|中国人民大学高瓴人工智能学院

研究方向 | 预训练模型

ICLR是人工智能领域顶级会议之一,会议主题包括深度学习、统计和数据科学,以及一些重要的应用,例如:计算机视觉、计算生物学、语音识别、文本理解、游戏和机器人等。

ICLR 2023将于2023年5月1日至5月5日在卢旺达基加利举行。由于官方的论文接受列表尚未公开,因此本文从投稿论文中选取了与自然语言处理相关的100多篇论文,按照不同的研究主题进行了分类整理,以供参考。

ICLR 2023投稿论文链接如下:https://openreview.net/group?id=ICLR.cc/2023/Conference。

目录

  • 模型

  • 文本生成

  • 机器翻译

  • 对话与问答

  • 知识与推理

  • 多模态

  • 信息检索

  • 代码

  • 数学

  • 知识蒸馏

  • 表示学习

  • 可解释性

  • 鲁棒性

  • 其他任务

  • Benchmark

1. 模型

1.1 模型结构

  • EIT: Enhanced Interactive Transformer for Sequence Generation

  • Transformers with Multiresolution Attention Heads

  • SaMoE: Parameter Efficient MoE Language Models via Self-Adaptive Expert Combination

  • Sparse MoE with Random Routing as the New Dropout: Training Bigger and Self-Scalable Models

1.2 模型训练

  • Guess the Instruction! Making Language Models Stronger Zero-Shot Learners

  • LEXA: Language-agnostic Cross-consistency Training for Question Answering Tasks

  • CCT: Cross-consistency training for Clone Detection and Code Search Tasks

  • Large Language Models Can Self-improve

  • Self-Guided Noise-Free Data Generation for Efficient Zero-Shot Learning

  • PMixUp: Simultaneous Utilization of Part-of-Speech Replacement and Feature Space Interpolation for Text Data Augmentation

  • Self-Consistent Learning: Cooperation between Generators and Discriminators

  • Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning

  • Toward Adversarial Training on Contextualized Language Representation

  • ContraGen: Effective Contrastive Learning For Causal Language Model

  • Language Model Pre-training with Linguistically Motivated Curriculum Learning

  • MLM with Global Co-occurrence

  • Improving Language Model Pretraining with Text Structure Information

  • Learning by Distilling Context

  • MAT: Mixed-Strategy Game of Adversarial Training in Fine-tuning

  • Sub-Task Decomposition Enables Learning in Sequence to Sequence Tasks

1.3 模型使用

  • Prompt Injection: Parameterization of Fixed Inputs

  • Meta-Weighted Language Model Tuning for Augmentation-Enhanced Few-Shot Learning

  • Pre-trained Language Models can be Fully Zero-Shot Learners

  • KnowDA: All-in-One Knowledge Mixture Model for Data Augmentation in Low-Resource NLP

  • Contrastive Novelty Learning: Anticipating Outliers with Large Language Models

  • Model ensemble instead of prompt fusion: a sample-specific knowledge transfer method for few-shot prompt tuning

  • Mass-Editing Memory in a Transformer

  • Zemi: Learning Zero-Shot Semi-Parametric Language Models from Multiple Tasks

  • Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models

  • Selective Annotation Makes Language Models Better Few-Shot Learners

  • Generate rather than Retrieve: Large Language Models are Strong Context Generators

  • Ahead-of-Time P-Tuning

  • Can discrete information extraction prompts generalize across language models?

2. 文本生成

  • Dynamic Scheduled Sampling with Imitation Loss for Neural Text Generation

  • DiffusER: Diffusion via Edit-based Reconstruction

  • MVP: Multi-task Supervised Pre-training for Natural Language Generation

  • Penalizing the High-likelihood: A Novel Sampling Method for Open-ended Neural Text Generation via Inverse Probability Weighting

  • RainProof: An Umbrella to Shield Text Generator from Out-Of-Distribution Data

  • A Non-monotonic Self-terminating Language Model

  • PromptSum: Planning with Mixed Prompts for Parameter-Efficient Controllable Abstractive Summarization

  • On the Usefulness of Embeddings, Clusters and Strings for Text Generation Evaluation

  • Joint Generator-Ranker Learning for Natural Language Generation

  • Calibrating Sequence likelihood Improves Conditional Language Generation

  • Sequence to sequence text generation with diffusion models

  • Tailoring Language Generation Models under Total Variation Distance

  • Language Models Can See: Plugging Visual Controls in Text Generation

  • Distribution Aware Metrics for Conditional Natural Language Generation

  • PEER: A Collaborative Language Model

3. 机器翻译

  • Seq2Seq Pre-training with Dual-channel Recombination for Translation

  • Simple and Scalable Nearest Neighbor Machine Translation

  • Fuzzy Alignments in Directed Acyclic Graph for Non-Autoregressive Machine Translation

4. 对话与问答

  • Towards Boosting the Open-Domain Chatbot with Human Feedback

  • Learning Locality and Isotropy in Dialogue Modeling

  • Knowledge-Consistent Dialogue Generation with Language Models and Knowledge Graphs

  • Complex-Target-Guided Open-Domain Conversation based on offline reinforcement learning

5. 知识与推理

  • ReAct: Synergizing Reasoning and Acting in Language Models

  • Language model with Plug-in Knowldge Memory

  • Thrust: Adaptively Propels Large Language Models with External Knowledge

  • Self-Consistency Improves Chain of Thought Reasoning in Language Models

  • DecAF: Joint Decoding of Answers and Logical Forms for Question Answering over Knowledge Bases

  • Least-to-Most Prompting Enables Complex Reasoning in Large Language Models

  • Neuro-Symbolic Procedural Planning with Commonsense Prompting

  • Multimodal Analogical Reasoning over Knowledge Graphs

  • ThinkSum: Probabilistic reasoning over sets using large language models

  • Joint Representations of Text and Knowledge Graphs for Retrieval and Evaluation

  • Rethinking Identity in Knowledge Graph Embedding

  • gGN: learning to represent nodes in directed graphs as low-rank Gaussian distributions

  • Don't Throw Your Old Policies Away: Knowledge-based Policy Recycling Protects Against Adversarial Attacks

  • Measuring and Narrowing the Compositionality Gap in Language Models

6. 多模态

  • CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers

  • CLIP model is an Efficient Continual Learner

  • Language Modelling with Pixels

  • Visual Classification via Description from Large Language Models

  • Contrastive Alignment of Vision to Language Through Parameter-Efficient Transfer Learning

  • RelationCLIP: Training-free Fine-grained Visual and Language Concept Matching

  • Contrastive Prompt Tuning Improves Generalization in Vision-Language Models

  • Masked Vision and Language Modeling for Multi-modal Representation Learning

  • UNIFIED-IO: A Unified Model for Vision, Language, and Multi-modal Tasks

  • Visually-augmented pretrained language models for NLP Tasks without Images

  • Music-to-Text Synaesthesia: Generating Descriptive Text from Music Recordings

  • VLG: General Video Recognition with Web Textual Knowledge

  • Dynamic Historical Adaptation for Continual Image-Text Modeling

  • From Images to Textual Prompts: Zero-shot VQA with Frozen Large Language Models

  • NÜWA-LIP: Language-guided Image Inpainting with Defect-free VQGAN

  • Universal Vision-Language Dense Retrieval: Learning A Unified Representation Space for Multi-Modal Retrieval

  • Socratic Models: Composing Zero-Shot Multimodal Reasoning with Language

  • Language-Guided Artistic Style Transfer Using the Latent Space of DALL-E

  • Unified Vision and Language Prompt Learning

  • DrML: Diagnosing and Rectifying Vision Models using Language

  • MaPLe: Multi-modal Prompt Learning

  • Prefix Conditioning Unifies Language and Label Supervision

  • Domain-Unified Prompt Representations for Source-Free Domain Generalization

  • Learning to Decompose Visual Features with Latent Textual Prompts

  • Delving into the Openness of CLIP

  • Cali-NCE: Boosting Cross-modal Video Representation Learning with Calibrated Alignment

  • Dynamic Historical Adaptation for Continual Image-Text Modeling

  • Design of the topology for contrastive visual-textual alignment

7. 信息检索

  • Multi-Vector Retrieval as Sparse Alignment

  • Augmenting Zero-shot Dense Retrievers With Plug-in Mixture-of-memories

  • CAMVR: Context-Adaptive Multi-View Representation Learning for Dense Retrieval

8. 代码

  • Language Models Can Teach Themselves to Program Better

  • Repository-Level Prompt Generation for Large Language Models of Code

  • NAPG: Non-Autoregressive Program Generation for Hybrid Tabular-Textual Question Answering

  • A Simple, Yet Effective Approach to Finding Biases in Code Generation

  • Deep Learning-based Source Code Complexity Prediction

  • FixEval: Execution-based Evaluation of Program Fixes for Competitive Programming Problems

  • InCoder: A Generative Model for Code Infilling and Synthesis

  • Code Translation with Compiler Representations

  • CodeT: Code Generation with Generated Tests

  • Multi-lingual Evaluation of Code Generation Models

9. 数学

  • Learning Math Reasoning from Self-Sampled Correct and Partially-Correct Solutions

  • Dynamic Prompt Learning via Policy Gradient for Semi-structured Mathematical Reasoning

10. 知识蒸馏

  • Speed Up Iterative Non-Autoregressive Transformers by Distilling Multiple Steps

  • A comparison of dataset distillation and active learning in text classification

  • Less is More: Task-aware Layer-wise Distillation for Language Model Compression

  • Distilling Text-Image Foundation Models

11. 表示学习

  • RankCSE: Unsupervised Sentence Representations Learning via Learning to Rank

  • Neural Embeddings for Text

  • Ranking-Enhanced Unsupervised Sentence Representation Learning

  • Neural Topic Modeling with Embedding Clustering Regularization

  • Counterfactual Contrastive Learning for Robust Text Classification

  • On The Inadequacy of Optimizing Alignment and Uniformity in Contrastive Learning of Sentence Representations

12. 可解释性

  • ORCA: Interpreting Prompted Language Models via Locating Supporting Evidence in the Ocean of Pretraining Data

  • ContraSim -- A Similarity Measure Based on Contrastive Learning

13. 鲁棒性

  • Learning from Others: Similarity-based Regularization for Mitigating Artifacts

  • Randomized Smoothing with Masked Inference for Adversarially Robust NLP Systems

14. 其他任务

  • Exploring Methods for Parsing Movie Scripts - Feature Extraction for Further Social Injustice Analysis

  • MSQ-BioBERT: Ambiguity Resolution to Enhance BioBERT Medical Question-Answering

  • Compositional Semantic Parsing with Large Language Models

  • AxBERT: An Explainable Chinese Spelling Correction Method Driven by Associative Knowledge Network

  • BED: Boundary-Enhanced Decoder for Chinese Word Segmentation

  • Semi-connected Joint Entity Recognition and Relation Extraction of Contextual Entities in Family History Records

15. Benchmark

  • GuoFeng: A Discourse-aware Evaluation Benchmark for Language Understanding, Translation and Generation

一起交流

想和你一起学习进步!『NewBeeNLP』目前已经建立了多个不同方向交流群(机器学习 / 深度学习 / 自然语言处理 / 搜索推荐 / 图网络 / 面试交流 / 等),名额有限,赶紧添加下方微信加入一起讨论交流吧!(注意一定o要备注信息才能通过)

ICLR 2023(投稿)|自然语言处理相关论文分类整理相关推荐

  1. SIGIR 2021 | 推荐系统相关论文分类整理

    © 作者|范欣妍 机构|中国人民大学高瓴人工智能学院 导师|赵鑫教授 研究方向 | 推荐系统 导读 ACM SIGIR 2021是CCF A类会议,人工智能领域智能信息检索( Information ...

  2. KBQA相关论文分类整理:简单KBQA和复杂KBQA

    ©作者 | 蒋锦昊 学校 | 中国人民大学博士生 研究方向 | 知识推理和问答系统 引言 基于知识图谱的问答系统(Knowledge Based Question Answering, KBQA)目标 ...

  3. KBQA相关论文分类整理

    © 作者|蒋锦昊 机构|中国人民大学高瓴人工智能学院直博生 导师|赵鑫教授 研究方向 | 知识推理和问答系统 1.引言 基于知识图谱的问答系统(Knowledge Based Question Ans ...

  4. CIKM 2021 | 推荐系统相关论文分类整理

    © 作者|孙文奇 机构|中国人民大学高瓴人工智能学院 研究方向|推荐系统 本文选取了CIKM2021中85篇长文.15篇应用文和29篇短文,重点对推荐系统相关论文(76篇)按不同的任务场景和研究话题进 ...

  5. 论文清单:SIGIR 2021推荐系统相关论文分类整理

    © 作者|范欣妍 机构|中国人民大学高瓴人工智能学院 研究方向 | 推荐系统 文章来源 | RUC AI Box 导读 ACM SIGIR 2021是CCF A类会议,人工智能领域智能信息检索( In ...

  6. SIGIR 2022 | 推荐系统相关论文分类整理

    大家好,我是对白. ACM SIGIR 2022是CCF A类会议,人工智能领域智能信息检索( Information Retrieval,IR)方向最权威的国际会议.会议专注于信息的存储.检索和传播 ...

  7. (zhuan) 126 篇殿堂级深度学习论文分类整理 从入门到应用

    126 篇殿堂级深度学习论文分类整理 从入门到应用 | 干货 雷锋网  作者: 三川 2017-03-02 18:40:00 查看源网址 阅读数:66 如果你有非常大的决心从事深度学习,又不想在这一行 ...

  8. 国家集训队论文分类整理[转]

    国家集训队论文分类整理 转自这里 dalao写的东西,是非常有学习价值的.反正noip不一定会用,但是对以后肯定有用的. 组合数学 计数与统计 2001 - 符文杰:<Pólya原理及其应用&g ...

  9. 国家集训队论文分类整理

    国家集训队论文分类整理 组合数学 计数与统计 2001 - 符文杰:<Pólya原理及其应用> 2003 - 许智磊:<浅谈补集转化思想在统计问题中的应用> 2007 - 周冬 ...

最新文章

  1. jenkins配置git
  2. c语言一输入运算就显示错误,大咖都会犯的18个c语言入门错误
  3. SilverLight是什么?(来自Csdn)
  4. java基础—Objcet中的equals方法重写
  5. Spring Cloud Eureka 属性配置中文说明文档
  6. layui 行变灰_layui table设置某一行的字体颜色方法
  7. 从前端到后端的跨域攻击与防御
  8. linux忘记开机密码解决方案
  9. Scrapy 中的 Request 对象和 Respionse 对象
  10. python一对一参数_Python数值操作1
  11. 潍坊学院c语言上机题库,参编学校_C语言程序设计上机指导与同步训练(刘迎春、张艳霞)_pdf_大学课件预览_高等教育资讯网...
  12. mysql报错无效默认值1067_Mysql 报错:#1067
  13. Science Word安装教程附下载链接
  14. Android studio 设置签名
  15. 虚拟运营商出牌:流量不清零难成杀手锏
  16. uni-app开发App上架Apple Store流程(简略版)
  17. 软件应用无所不在 正吞噬整个世界
  18. 区别:千赫kHz、兆赫MHz、吉赫GHz、太赫THz、拍赫PHz、艾赫EHz每两级之间的换算以及之间的关系
  19. winform ui 设计参考
  20. centos服务器操作命令

热门文章

  1. vivo手机助手新手使用图文教程
  2. 第一次到这里,踩个脚印
  3. 聊聊最近的面试感受,几点忠告
  4. 基于flask-socketio的web聊天室
  5. 跑马灯抽奖js php,文字跑马灯效果的完成
  6. jQuery仿天猫翻牌抽奖代码
  7. Hanselminutes播客1
  8. ubuntu18.04 kubeadm 安装kubernetes v1.18.3
  9. 安防监控实现之从网页上控制A9的LED灯
  10. 在IIS环境之外运行ASPNET程序