机器学习 声音 分角色

Bart. Elmo. Bert. Kermit. Marge. What do they have in common?

巴特 埃尔莫 伯特 柯米特 玛格 他们有什么共同点?

They’re all beloved fictional characters from TV shows many of us watched when we were young. But that’s not all — they’re also all AI models.

它们都是我们许多人在年轻时观看的电视节目中最受欢迎的虚构人物。 但这还不是全部-它们都是AI模型。

In 2018, researchers at the Allen Institute published the language model ELMo. The lead author, Matt Peters, said the team brainstormed many acronyms for their model, and ELMo instantly stuck as a “whimsical but memorable” choice.

2018年,艾伦研究所(Allen Institute)的研究人员发布了语言模型ELMo 。 主要作者马特·彼得斯(Matt Peters)说,该团队对其模型进行了许多头脑风暴,而ELMo立即成为“异想天开但令人难忘”的选择。

What started out as an inside joke has become a full-blown trend.

最初由内心的笑话变成了一种成熟的趋势。

Google AI followed with BERT, an incredibly powerful and now widely used Transformer-based language model. Then, many more: ERNIE, KERMIT, BART, GROVER, etc. OpenAI almost named GPT-2 “Snuffleupagus, or Snuffy for short.” Just last month, Facebook AI published MARGE.

Google AI紧随其后的是BERT ,这是一种功能强大且现已广泛使用的基于Transformer的语言模型。 然后,还有更多: ERNIE , KERMIT , BART , GROVER等。OpenAI 几乎将GPT-2命名为“ Snuffleupagus或简称Snuffy”。 就在上个月,Facebook AI发布了MARGE 。

This article gives an overview on all of these models and, of course, their character inspirations below:

本文概述了所有这些模型,以及下面它们的特征灵感:

ELMo(2018):语言模型的嵌入 (ELMo (2018): Embeddings from Language Models)

The one that started this trend, ELMo is a deep contextualized word representation approach that is able to capture more characteristics about words (syntax, semantics, and more).

引发这一趋势的一种方法是ELMo , 它是一种深层的上下文化词表示方法,能够捕获有关词的更多特征(语法,语义等)。

A big challenge in representing words as vectors (“word embeddings”) was that a word would be represented by the same vector no matter what context it was used in. However, “current” has different meanings in “what’s your current job?” versus “that is a strong river current” — we can’t just use one fixed representation for both “current”s!

将单词表示为矢量(“单词嵌入”)的一大挑战是,无论使用何种上下文,一个单词都将由相同的矢量表示。但是,“当前”在“您目前的工作是什么?”中具有不同的含义。 而不是“那是强劲的河流潮流 ” —我们不能仅对两个“ 潮流”使用一个固定的表示!

Thus, contextualized word embeddings were created to capture a word’s context in its representation as well. Rather than only reading a word at a time, ELMo reads the entire sentence for context before assigning every word its embedding, which is done using a bi-directional LSTM.

因此,创建情境字的嵌入到捕捉词的上下文中的表现也是如此。 ELMo不仅一次读取一个单词,还为上下文分配了整个句子,然后才为每个单词分配嵌入,这是使用双向LSTM完成的。

ELMo was a great step forward in natural language processing (NLP) research using language modeling. For an illustrated explanation of ELMo, I would highly recommend this resource.

ELMo在使用语言建模的自然语言处理(NLP)研究方面迈出了一大步。 对于ELMo的图解说明,我强烈推荐该资源 。

Outside machine learning and NLP, Elmo is an adorable furry red Muppet on the children’s show Sesame Street. Elmo likes surprises, pizza, and bubble baths and won the longest giggle award on the Golden Grover Awards. It first appeared on screen in 1980.

除了机器学习和NLP外,Elmo是儿童表演芝麻街上的可爱毛茸茸的红色布偶。 Elmo喜欢惊喜,披萨和泡泡浴,并获得了Golden Grover奖中最长的傻笑奖 。 它于1980年首次出现在屏幕上。

Image for post
ELMoELMo

BERT(2019):变压器的双向编码器表示 (BERT (2019): Bidirectional Encoder Representations from Transformers)

Google further transformed pre-training in NLP by introducing BERT, a new Transformer-based language model that was the first to allow deeply bidirectional and unsupervised representation.

Google通过引入 BERT (一种基于Transformer的新语言模型)进一步转变了NLP中的预培训,这是第一个允许深度 双向和无监督表示的模型。

Deeply bidirectional means in capturing contextual embeddings, BERT represents each word using both its previous and next context. (In comparison, ELMo is shallowly bidirectional.) Simply conditioning each word on its previous and next words can be problematic, so BERT randomly masks some of the words and conditions each word bidirectionally to predict the masked words.

BERT是捕获上下文嵌入的一种深度双向方法,它使用上一个和下一个上下文来表示每个单词。 (相比之下,ELMo是浅双向的。)仅对每个单词的前一个单词和下一个单词进行条件调整可能会出现问题,因此BERT会随机屏蔽某些单词,并双向地调整每个单词以预测被屏蔽的单词。

In its initial release, BERT already achieved impressive results in question-and-answering and natural language understanding tasks. BERT and other Transformer-based architectures have been a cornerstone of NLP research in the past year.

在最初的发行版中,BERT在问答和自然语言理解任务中已经取得了令人瞩目的成果。 过去一年,BERT和其他基于Transformer的体系结构一直是NLP研究的基石。

Outside machine learning and NLP, Bert is a friendly yellow character on Sesame Street. In his free time, he likes to read Boring Stories, eat oatmeal, and study pigeons.

除了机器学习和NLP外,Bert在芝麻街上是一个友好的黄色人物。 在业余时间,他喜欢阅读《 无聊的故事》 ,吃燕麦片和研究鸽子。

Image for post
BERTBERT

ERNIE(2019):双重麻烦 (ERNIE (2019): Double Trouble)

You can’t have Bert without his best buddy Ernie — good thing researchers developed ERNIE (Sun et al.), ERNIE (Zhang et al.), and even ERNIE 2.0!

没有他最好的伙伴Ernie,您将无法拥有Bert,好东西研究人员开发了ERNIE (Sun等人), ERNIE (Zhang等人),甚至ERNIE 2.0 !

The first ERNIE (Enhanced Representation through Knowledge Integration) presented a language model that extended BERT’s word masking strategy to entity-level and phrase-level masking. In doing so, this ERNIE can implicitly learn the prior knowledge of phrases, entities, and relationships between them during the training process.

第一个ERNIE (通过知识集成增强表示)提出了一种语言模型,该模型将BERT的单词屏蔽策略扩展到实体级和短语级屏蔽 。 这样,该ERNIE可以在训练过程中隐式学习短语,实体以及它们之间的关系的先验知识。

Less than two months later, a second ERNIE (Enhanced Language Representation with Informative Entities) was published. This ERNIE presented a language model that incorporated knowledge graphs to optimize for having as much information as possible. Knowledge graphs are a powerful way to represent data points and relations that link them together.

不到两个月后,第二本ERNIE (信息实体的增强语言表示)发布了。 该ERNIE提出了一种语言模型,该模型合并了知识图以进行优化以获取尽可能多的信息。 知识图是表示数据点和将它们链接在一起的关系的有效方法。

Outside machine learning and NLP, Ernie is a troublemaker whose life mission is to annoy Bert on Sesame Street. He is extremely fond of his rubber duckie and once famously said “I can’t hear you, I have a banana in my ear!”

除了机器学习和NLP之外,Ernie是一个麻烦制造者,其人生使命是惹恼芝麻街上的 Bert 他非常喜欢他的橡皮鸭,并且曾经有句著名的话:“我听不到您的声音,我的耳朵里有香蕉!”

Image for post
ERNIEERNIE

KERMIT(2019):插入变换制作的Kontextuell编码器表示形式 (KERMIT (2019): Kontextuell Encoder Representations Made by Insertion Transformations)

KERMIT is an insertion-based generative architecture that models the joint distribution, decompositions (i.e. its marginals), and conditionals together. The researchers found that KERMIT worked well for a few select tasks, including machine translation.

KERMIT是基于插入的生成体系结构,可对关节分布,分解(即其边际)和条件一起建模。 研究人员发现,KERMIT在某些选定的任务(包括机器翻译)中表现良好。

In case you were wondering like I was, “kontextuell” is Swedish for “contextual”. We know the researchers were quite intentional on this naming because later in the paper, they write “Then, like its friends ELMo (Peters et al., 2018), BERT (Devlin et al., 2019), and ERNIE (Sun et al., 2019), we can also use KERMIT…”

如果您像我一样想知道,“ kontextuell”是瑞典语的“上下文”。 我们知道研究人员在命名时非常有意,因为他们在论文的后面写道:“然后, 像它的朋友 ELMo(Peters等,2018),BERT(Devlin等,2019)和ERNIE(Sun等, 。,2019),我们也可以使用KERMIT…”

Outside machine learning and NLP, Kermit is an iconic singing frog that has graced many Muppet productions (Sam and Friends, Sesame Street, The Muppet Show, etc.) and memes (But That’s None of My Business, Evil Kermit, etc.). Introduced in 1955, Kermit is the oldest TV character on this list.

在机器学习和NLP之外,Kermit是一个标志性的唱歌青蛙,它使Muppet的许多作品( Sam和Friends,芝麻街,The Muppet Show等)和模因( 但不关我的事 , Evil Kermit等)得到了赞誉 。 Kermit于1955年推出,是该列表中最古老的电视角色。

Image for post
KERMITKERMIT

BART(2019):双向和自回归变压器 (BART (2019): Bidirectional and Auto-Regressive Transformers)

Facebook AI built on BERT, GPT, and previous NLP pre-training work to create BART, a new pretrained model for text generation and comprehension that combines both bidirectional and auto-regressive Transformers.

基于BERT,GPT和以前的NLP预训练工作的Facebook AI创建了BART ,这是一种用于文本生成和理解的新的预训练模型,该模型结合了双向和自回归变压器。

BERT used masking in its bidirectional encoder, which meant that masked/missing words were predicted independently. As a result, BERT can’t be used for text generation.

BERT在其双向编码器中使用了屏蔽,这意味着屏蔽/缺失字是独立预测的。 结果,BERT不能用于文本生成。

In contrast, GPT is auto-regressive, which means that it predicts a future word from a set of words given a context in a forward direction. As a result, GPT can’t learn bidirectional interactions.

相反,GPT是自回归的,这意味着GPT从给定上下文中向前的一组单词中预测未来的单词。 结果,GPT无法学习双向交互。

BART combines these fundamental ideas together: 1) a document with random spans of text replaced with masks is encoded bidirectionally, then 2) the likelihood of the unmodified document is predicted with an autoregressive decoder.

BART将这些基本思想结合在一起:1)对用文本掩码替换的随机文本跨度的文档进行双向编码,然后2)用自回归解码器预测未修改文档的可能性。

Outside machine learning and NLP, Bart is one of TV’s most well-known young rebels. You can find Bart and his endless prank calls on The Simpsons.

除了机器学习和NLP外,Bart是电视界最知名的年轻叛军之一。 您可以在《辛普森一家》中找到Bart和他无休止的恶作剧电话。

Image for post
BARTBART

格罗夫(2019) (GROVER (2019))

The NLP field has progressed very quickly in recent years to achieve highly believable summarization and translation. However, these technologies can also be used for less-than-positive purposes, such as AI-generated fake news and propaganda.

近年来,NLP领域取得了飞速发展,以实现高度可信的摘要和翻译。 但是,这些技术也可以用于不太积极的目的,例如AI生成的假新闻和宣传。

To combat this, researchers created GROVER, a publicly available generator that can write realistic and controlled fake news. GROVER’s purpose is so others could practice against it to develop better techniques for classifying AI-generated fake news from real, human-written news. At the time, the best discriminator could only distinguish AI-generated fake news from real news with 73% accuracy.

为了解决这个问题,研究人员创建了GROVER , GROVER是一种公开可用的生成器,可以编写真实且受控的假新闻 。 GROVER的目的是让其他人可以实践它,以开发出更好的技术,以将AI生成的假新闻从真实的人工编写的新闻中分类。 当时,最佳判别器只能以73%的准确度区分AI生成的假新闻和真实新闻。

(In a baffling naming decision, there’s no acronym here — it’s called GROVER just because.)

(在一个令人困惑的命名决定中,这里没有首字母缩写,仅因为它而被称为GROVER。)

Outside machine learning and NLP, Grover is a furry, blue Sesame Street character who loves to help (or tries to help) others. Hopefully GROVER can help the AI world in addressing disinformation as much as Grover helps his friends.

在机器学习和NLP之外,格罗弗是一位毛茸茸的蓝色芝麻街角色,他喜欢帮助(或尝试帮助)他人。 希望GROVER可以像Grover帮助他的朋友一样,帮助AI世界解决虚假信息。

Image for post
GROVERGROVER

MARGE(2020):检索和生成的多语言自动编码器 (MARGE (2020): Multilingual Autoencoder that Retrieves and Generates)

Published very recently by Facebook AI, MARGE is a new “a pre-trained sequence-to-sequence model learned with an unsupervised multilingual multi-document paraphrasing objective.”

MARGE由Facebook AI最近发布,是一种新的“预先训练的序列到序列模型,它是在无监督的多语言多文档措辞目标下学习的。”

In simpler words, it’s a pre-trained language model that generates text by 1) retrieving related text in other languages, 2) reconstructing the original text by finding patterns within them.

用简单的话来说,它是一个经过训练的语言模型,它通过以下步骤生成文本:1) 检索其他语言的相关文本,2)通过在其中查找模式来重建原始文本。

The researchers found that MARGE can successfully perform paraphrasing, translation, multi-document summarization, and information retrieval tasks all without any fine tuning. MARGE scored up to 35.8 on BLEU (Bilingual Evaluation Understudy), a metric for language translations, which is considered quite high for a model without fine tuning.

研究人员发现,MARGE可以成功执行释义,翻译,多文档摘要和信息检索任务,而无需进行任何微调。 MARGE在BLEU (双语评估研究)(语言翻译指标)上的得分高达35.8,对于没有微调的模型来说,这是相当高的。

Outside machine learning and NLP, Marge is a fictional character on The Simpsons. Mother of three kids (including Bart), Marge is also well-known for her tall, blue (bleu?

机器学习 声音 分角色_机器学习对儿童电视节目角色的痴迷相关推荐

  1. 机器学习线性回归学习心得_机器学习中的线性回归

    机器学习线性回归学习心得 机器学习中的线性回归 (Linear Regression in Machine Learning) There are two types of supervised ma ...

  2. 机器学习中qa测试_机器学习项目测试怎么做?(看实例)

    机器学习交付项目通常包含两部分产物,一部分是机器学习模型,另一部分是机器学习应用系统.机器学习模型是嫁接在应用之上产生价值的.比如:一款预测雷雨天气的APP,它的雷雨预测功能就是由机器学习模型完成的. ...

  3. oracle查询用户权限和角色_详解jenkins配置用户角色权限的实现方法

    概述 今天介绍下jenkins应该怎么去配置用户角色权限,注意jenkins 配置用户角色权限需要安装插件 Role Strategy Plugin. 1.安装 Role Strategy Plugi ...

  4. java 绘制角色_利用scratch绘图编辑器绘制角色和角色的加工

    同学们,前面我们学习了矢量图的绘制编辑工具,那么我们这节课就来学习利用绘制编辑工具绘制角色和修改加工角色. 一.微课视频分享 1.绘制角色 在编辑区有两种模式,位图与矢量图模式,因为矢量图放大后不会模 ...

  5. python卡方分箱_机器学习(十六)特征工程之数据分箱

    1 分箱简介 数据分箱(也称为离散分箱或分段)是一种数据预处理技术,用于减少次要观察误差的影响,是一种将多个连续值分组为较少数量的"分箱"的方法. 例如,例如我们有一组关于人年龄的 ...

  6. 机器学习解决什么问题_机器学习帮助解决水危机

    机器学习解决什么问题 According to Water.org and Lifewater International, out of 57 million people in Tanzania, ...

  7. 机器学习回归模型精度_机器学习模型实施精度召回和概率截止。

    机器学习回归模型精度 Congratulations on completing your Machine Learning (ML) pipeline! In the second part of ...

  8. 机器学习朴素贝叶斯_机器学习基础朴素贝叶斯分类

    机器学习朴素贝叶斯 In the previous stories, I had given an explanation of the program for implementation of v ...

  9. 机器学习模型 知乎_机器学习:模型评估之评估方法

    ​机器学习已经成为了人工智能的核心研究领域之一,它的研究动机就是为了让计算机系统具有人的学习能力以便实现人工智能.目前,关于机器学习定义的说法比较多,而被广泛采用的定义是"利用经验来改善计算 ...

  10. 机器学习中qa测试_机器学习自动化单元测试平台

    机器学习自动化单元测试平台.零代码.全方位.自动化测试方法/函数的正确性和可用性. 原理 后端不需要写任何单元测试代码(逻辑代码.注解代码等全都不要), 这个工具会自动生成测试参数,并执行方法,拿到返 ...

最新文章

  1. git checkout .的用法
  2. linux系统切换用户
  3. golang 并发实践
  4. Ubantu18.04安装Vivado
  5. android15版本号手机,Lineage OS 15.1正式发布:基于Android 8.1
  6. SQLServer2008设置开启INTERNET远程连接
  7. php 判断昨天_PHP 判断时间在今天、昨天、前天、几天前几点
  8. 基于matlab人脸识别论文,基于matlab的人脸识别系统设计 毕业论文
  9. 小程序 获取小程序链接
  10. python英文词频统计软件_英语词频统计软件功能介绍
  11. 需求变更对软件质量的影响
  12. BAT解密:互联网技术发展之路(2)- 业务如何驱动技术发展
  13. Mac解压rar文件工具
  14. 解决ios的webview中上/下拉露出黑灰色背景问题
  15. php环境下,两种汇率接口,已经测试通过
  16. opengl碗状_OpenGLES简单入门
  17. 响应式织梦模板行业资讯网类网站
  18. 极智Coding | C 和 C++ 读存 bin 文件方法
  19. IBM:云计算 SKC
  20. springmvc官网jar下载

热门文章

  1. 论SetItemData和GetItemData
  2. ZOJ 1111 Poker Hands
  3. PageRank与社交网络模型评估
  4. IP网络中的路由聚合的解析
  5. .NET编译工具(每日构建或自动化编译)
  6. 我靠! 算你运气好 a joke
  7. conda创建环境速度慢解决办法
  8. 2020-08-17每日一句
  9. 190520每日一句
  10. 190519每日一句