An overview of the main types of neural network architecture 神经网络结构的主要类型

什么是神经网络的结构?

神经网络中神经元的组织方式。

1. 目前应用领域中最常用的结构类型是feet-forward 神经网络, 信息来自输入单元,并且以一个方向流动,通过隐含层,直到所有信息到达输出单元。

2 .一种非常有趣的结构类型是recurrent神经网络,information can flow round in cycles. 这种网络能够记住信息一段时间,They can exhibit all sorts of interesting oscillations but they are much more difficult to train in part because they are so much more complicated in what they can do. 然而最近,人们在训练recurrent神经网络上有了很大进展,他们现在能做令人印象深刻的事。

3. 最后一种结构是symmetrically-connected 网络,在这种网络中,权值是相同的in both directions between two units.

feet-forward 神经网络

These are the commonest type of neural network in practical applications. 最常用
– The first layer is the input and the last layer is the output. 第一层是输入,最后一层是输出
– If there is more than one hidden layer, we call them “deep” neural networks. 如果多于一层的隐藏层,我们就称为“深度”神经网络
• They compute a series of transformations between their input an output. So at each layer, you get a new representation of the input in which things that were similar in the previous layer may have become less similar, or things that were dissimilar in the previous layer may become more similar. So in speech recongntion, for example, we'd like the same thing said by different speakers to become more similar, and different thing said by the same speaker to be less similar as we go up through the layers of the network.
– In order to achieve this, we need the activities of the neurons in each layer to be a non-linear function of the activities in the layer below.

Recurrent networks
他比前向网络强大得多They have directed cycles in their connection graph.在他们的连接图中是有方向的循环
That means you can sometimes get back to where you started by following the arrows. 这意味着你有时想回到起点,沿着这些箭头即可
They can have complicated dynamics and this can make them very difficult to train.
– There is a lot of interest at present in finding efficient ways of training recurrent nets. 目前在寻找训练recurren网络高效方法上引起了很多的兴趣

They are more biologically realistic. 同时他们是基于生物学现实的

Recurrent nets with multiple hidden layers are just a special case of a general recurrent neural net that has some of its hidden to hidden connections missing.

Recurrent neural networks for modeling sequences
Recurrent neural networks are a very naturalway to model sequential data: 非常适合对序列数据建模, 我们需要做的是在隐含单元之间建立联系
– They are equivalent to very deep nets withone hidden layer per time slice. 隐含单元表现的像网络,very deep in time.

所以每个时间步长隐含单元的状态决定了下一时刻时间步长隐含单元状态

– Except that they use the same weights at every time slice and they get input at every time slice.

他们区别与前向网络的一个方面是我们在每个时间步长使用相同的权值。所以下图红色箭头,隐含单元正在决定下时刻的隐含单元。红色箭头描述权值矩阵,在每个时间步长都是相同的

• They have the ability to remember information in their hidden state for a long time.

他们在每个时间也获得输入,在每个时间戳也给出输出,这些也都使用相同的权值矩阵。他们能在隐藏单元储存信息很长一段时间  
– But its very hard to train them to use this potential.

An example of what recurrent neural nets can now do
Ilya Sutskever (2011) trained a special type of recurrent neural net to predict the next character in a sequence. So llya trained it on lots and lots of strings from English Wikipedia. It's seeing English characters and trying to predict the next English character. He actually used 86 different characters to allow for punctuation, and digits , and capital letters and so on. After you trained it, one way of seeing how well it can do is to see whether it assigns high probability to the next character that actually occurs. Another way of seeing get it to generate text. So what you do is you give it a string of characters and get it to predit probabilitites for the next character.Then you pick the next character from that probability distribution. It's no use picking the most likely character. If you do that after a while it starts saying the United States of the United States of the United States of the United States of the United States. That tells you something about Wikipedia.

Some text generated one character at a time by Ilya Sutskever’s recurrent neural network
In 1974 Northern Denver had been overshadowed by CNL, and several Irish intelligence agencies in the Mediterranean
region. However, on the Victoria, Kings Hebrew stated that Charles decided to
escape during an alliance. The mansion house was completed in 1882, the second in
its bridge are omitted, while closing is the proton reticulum composed below it aims,
such that it is the blurring of appearing on any well-paid type of box printer.

symmetrically-connected 网络

与recurrent网络相似,These are like recurrent networks, but the connections between units are symmetrical (they have the same weight in both directions)
– John Hopfield (and others) realized that symmetric networks are much easier to analyze than recurrent networks.
– They are also more restricted in what they can do. because they obey an energy function
   For example, they cannot model cycles.
• Symmetrically connected nets without hidden units are called “Hopfield nets”
Symmetrically connected networks with hidden units
• These are called “Boltzmann machines”.
– They are much more powerful models than Hopfield nets.
– They are less powerful than recurrent neural networks.
– They have a beautifully simple learning algorithm.
• We will cover Boltzmann machines towards the end of the course

转载于:https://www.cnblogs.com/jinee/p/4473984.html

机器学习中使用的神经网络(六) --第二课相关推荐

  1. 机器学习中使用的神经网络(三)

    神经元的简单模型 Idealized neurons • To model things we have to idealize them (e.g. atoms) – Idealization re ...

  2. 机器学习中的数学(全集)

    在当下,机器学习.人工智能领域吸引了许多同学投身其中,其中包含了大量非科班出身或者从其他行业切换赛道转行而来的朋友们,大家在学习的过程中发现学习曲线陡峭.难度较大,普遍的心声就是:机器学习难,首要就是 ...

  3. 1.3 机器学习基础-深度学习第二课《改善深层神经网络》-Stanford吴恩达教授

    ←上一篇 ↓↑ 下一篇→ 1.2 偏差/方差 回到目录 1.4 正则化 机器学习基础 (Basic "Recipe" for Machine Learning) 上节课我们讲的是如 ...

  4. Javascript第六章JavaScript中构造器创建对象第二课

    Javascript第六章JavaScript用new创建对象第一课 https://blog.csdn.net/qq_30225725/article/details/89304586 Javasc ...

  5. Coursera公开课笔记: 斯坦福大学机器学习第二课“单变量线性回归(Linear regression with one variable)”

    Coursera公开课笔记: 斯坦福大学机器学习第二课"单变量线性回归(Linear regression with one variable)" 发表于 2012年05月6号 由 ...

  6. 0.0 目录-深度学习第二课《改善神经网络》-Stanford吴恩达教授

    文章目录 第五课 第四课 第三课 第二课 第一课 第五课 <序列模型> 笔记列表 Week 1 循环序列模型 Week 1 传送门 -> 1.1 为什么选择序列模型 1.2 数学符号 ...

  7. 吴恩达deeplearning.ai系列课程笔记+编程作业(6)第二课 改善深层神经网络-第二周:优化算法 (Optimization algorithms)

    第二门课 改善深层神经网络:超参数调试.正则化以及优化(Improving Deep Neural Networks:Hyperparameter tuning, Regularization and ...

  8. 2.4 理解指数加权平均-深度学习第二课《改善深层神经网络》-Stanford吴恩达教授

    ←上一篇 ↓↑ 下一篇→ 2.3 指数加权平均 回到目录 2.5 指数加权平均的偏差修正 理解指数加权平均 (Understanding Exponentially Weighted Averages ...

  9. 1.8 其他正则化方法-深度学习第二课《改善深层神经网络》-Stanford吴恩达教授

    ←上一篇 ↓↑ 下一篇→ 1.7 理解 Dropout 回到目录 1.9 归一化输入 其他正则化方法 (Other Regularization Methods) 除了 L2L2L2 正则化和随机失活 ...

最新文章

  1. CSS3秘笈第三版涵盖HTML5学习笔记13~17章
  2. 周一02.1编程语言介绍
  3. 取代C语言标准输入输出:cin 和 cout【C++标准输入输出】
  4. 电脑账户头像怎么删掉_情侣头像丨情侣头像一男一女背影
  5. 【Qt】简单Qt文本查找功能
  6. Caffe阅读代码并修改
  7. Codeforces Round #281 (Div. 2) A. Vasya and Football 模拟
  8. JDK源码学习之一lang包
  9. Bootstrap列表组禁用和活动状态
  10. NATSserver配置具体解释
  11. GlusterFS更换故障Brick
  12. IBatisNet -- 保护你的配置文件及映射文件信息
  13. 响应式高端大气的模板源码图库素材资源下载平台网站源码,无加密无后门
  14. speedoffice(PPT)怎么将背景设置为渐变的背景
  15. 华为手机企业邮箱无法连接到服务器,华为手机邮箱无法登录
  16. Hadoop Yarn Linux Container Executor配置
  17. function func() { for(var i=0; i<3; i++) { setTimeout(()=>{ consol
  18. 高仿知乎android客户端,仿知乎分享界面
  19. BOM:窗口位置、页面视口大小、window.open
  20. 【python脚本系列】gif录制png序列帧

热门文章

  1. 职场干货——职场多年,一些对付小人的经验之谈
  2. 打开计算机首先映入眼帘的一整屏称之为桌面,利用“活动桌面” 让桌面也“闪”起来...
  3. 一个简单的页面加载管理类(包含加载中,加载失败,数据为空,加载成功)
  4. HBase进化之从NoSQL到NewSQL,凤凰涅槃成就Phoenix
  5. NO.2 微信第三方平台(小程序)授权流程技术说明
  6. (USB:VCP+HID复合设备与系统配置)
  7. android 仿微信demo————注册功能实现(服务端)
  8. visio的图在word中空白处太大的问题
  9. 谈学习中的改变——有病要诊断,有药要服用
  10. html页面如何获取已经存在的token,移动端通过携带token访问html页面