Deep Learning经典论文列表(Reading List)

文章目录

  • Review Papers
  • Computer Vision
  • NLP and Speech
  • Disentangling Factors and Varitions with Depth
  • Transfer Learning and domain adaptation
  • Practical Tricks and Guides
  • Sparse Coding
  • Foundation Theory and Motivation
  • Classification
  • Large Scale Deep Learning
  • Recurrent Networks
  • Hyper Parameters
  • Optimization
  • Unsupervised Feature Learning
  • Miscellaneous

Reading List

List of reading lists and survey papers:

  • Review Papers

    • Representation Learning: A Review and New Perspectives, Yoshua Bengio, Aaron Courville, Pascal Vincent, Arxiv, 2012.
    • The monograph or review paper Learning Deep Architectures for AI (Foundations & Trends in Machine Learning, 2009).
    • Deep Machine Learning – A New Frontier in Artificial Intelligence Research – a survey paper by Itamar Arel, Derek C. Rose, and Thomas P. Karnowski.
    • Graves, A. (2012). Supervised sequence labelling with recurrent neural networks(Vol. 385). Springer.
  • Computer Vision

    • ImageNet Classification with Deep Convolutional Neural Networks, Alex Krizhevsky, Ilya Sutskever, Geoffrey E Hinton, NIPS 2012.
    • Learning Hierarchical Features for Scene Labeling, Clement Farabet, Camille Couprie, Laurent Najman and Yann LeCun, IEEE Transactions on Pattern Analysis and Machine Intelligence, 2013.
    • Learning Convolutional Feature Hierachies for Visual Recognition, Koray Kavukcuoglu, Pierre Sermanet, Y-Lan Boureau, Karol Gregor, Michaël Mathieu and Yann LeCun, Advances in Neural Information Processing Systems (NIPS 2010), 23, 2010.
    • Graves, Alex, et al. “A novel connectionist system for unconstrained handwriting recognition.” Pattern Analysis and Machine Intelligence, IEEE Transactions on 31.5 (2009): 855-868.
    • Cireşan, D. C., Meier, U., Gambardella, L. M., & Schmidhuber, J. (2010). Deep, big, simple neural nets for handwritten digit recognition. Neural computation, 22(12), 3207-3220.
    • Ciresan, Dan, Ueli Meier, and Jürgen Schmidhuber.“Multi-column deep neural networks for image classification.” Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on. IEEE, 2012.
    • Ciresan, D., Meier, U., Masci, J., & Schmidhuber, J. (2011, July). A committee of neural networks for traffic sign classification. In Neural Networks (IJCNN), The 2011 International Joint Conference on (pp. 1918-1921). IEEE.
  • NLP and Speech

    • Joint Learning of Words and Meaning Representations for Open-Text Semantic Parsing, Antoine Bordes, Xavier Glorot, Jason Weston and Yoshua Bengio (2012), in: Proceedings of the 15th International Conference on Artificial Intelligence and Statistics (AISTATS)
    • Dynamic pooling and unfolding recursive autoencoders for paraphrase detection. Socher, R., Huang, E. H., Pennington, J., Ng, A. Y., and Manning, C. D. (2011a).  In NIPS’2011.
    • Semi-supervised recursive autoencoders for predicting sentiment distributions. Socher, R., Pennington, J., Huang, E. H., Ng, A. Y., and Manning, C. D. (2011b).  In EMNLP’2011.
    • Mikolov Tomáš: Statistical Language Models based on Neural Networks. PhD thesis, Brno University of Technology, 2012.
    • Graves, Alex, and Jürgen Schmidhuber. “Framewise phoneme classification with bidirectional LSTM and other neural network architectures.“ Neural Networks 18.5 (2005): 602-610.
  • Disentangling Factors and Varitions with Depth

    • Goodfellow, Ian, et al. “Measuring invariances in deep networks.” Advances in neural information processing systems 22 (2009): 646-654.
    • Bengio, Yoshua, et al. “Better Mixing via Deep Representations.” arXiv preprint arXiv:1207.4404(2012).
    • Xavier Glorot, Antoine Bordes and Yoshua Bengio, Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach, in: Proceedings of the Twenty-eight International Conference on Machine Learning (ICML’11), pages 97-110, 2011.
  • Transfer Learning and domain adaptation

    • Raina, Rajat, et al. “Self-taught learning: transfer learning from unlabeled data.” Proceedings of the 24th international conference on Machine learning. ACM, 2007.
    • Xavier Glorot, Antoine Bordes and Yoshua Bengio, Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach, in: Proceedings of the Twenty-eight International Conference on Machine Learning (ICML’11), pages 97-110, 2011.
    • R. Collobert, J. Weston, L. Bottou, M. Karlen, K. Kavukcuoglu and P. Kuksa. Natural Language Processing (Almost) from ScratchJournal of Machine Learning Research, 12:2493-2537, 2011.
    • Mesnil, Grégoire, et al. “Unsupervised and transfer learning challenge: a deep learning approach.”Unsupervised and Transfer Learning Workshop, in conjunction with ICML. 2011.
    • Ciresan, D. C., Meier, U., & Schmidhuber, J. (2012, June). Transfer learning for Latin and Chinese characters with deep neural networks. In Neural Networks (IJCNN), The 2012 International Joint Conference on (pp. 1-6). IEEE.
  • Practical Tricks and Guides

    • “Improving neural networks by preventing co-adaptation of feature detectors.” Hinton, Geoffrey E., et al.  arXiv preprint arXiv:1207.0580 (2012).
    • Practical recommendations for gradient-based training of deep architectures, Yoshua Bengio, U. Montreal, arXiv report:1206.5533, Lecture Notes in Computer Science Volume 7700, Neural Networks: Tricks of the Trade Second Edition, Editors: Grégoire Montavon, Geneviève B. Orr, Klaus-Robert Müller, 2012.
    • A practical guide to training Restricted Boltzmann Machines, by Geoffrey Hinton.
  • Sparse Coding

    • Emergence of simple-cell receptive field properties by learning a sparse code for natural images, Bruno Olhausen, Nature 1996.
    • Kavukcuoglu, Koray, Marc’Aurelio Ranzato, and Yann LeCun. “Fast inference in sparse coding algorithms with applications to object recognition.“ arXiv preprint arXiv:1010.3467 (2010).
    • Goodfellow, Ian, Aaron Courville, and Yoshua Bengio. “Large-Scale Feature Learning With Spike-and-Slab Sparse Coding.” ICML 2012.
    • Efficient sparse coding algorithms. Honglak Lee, Alexis Battle, Raina Rajat and Andrew Y. Ng. In NIPS 19, 2007. pdf
    • “Sparse coding with an overcomplete basis set: A strategy employed by VI?.” . Olshausen, Bruno A., and David J. Field. Vision research 37.23 (1997): 3311-3326.
  • Foundation Theory and Motivation

    • Hinton, Geoffrey E. “Deterministic Boltzmann learning performs steepest descent in weight-space.” Neural computation 1.1 (1989): 143-150.
    • Bengio, Yoshua, and Samy Bengio. “Modeling high-dimensional discrete data with multi-layer neural networks.” Advances in Neural Information Processing Systems 12 (2000): 400-406.
    • Bengio, Yoshua, et al. “Greedy layer-wise training of deep networks.” Advances in neural information processing systems 19 (2007): 153.
    • Bengio, Yoshua, Martin Monperrus, and Hugo Larochelle. “Nonlocal estimation of manifold structure.” Neural Computation 18.10 (2006): 2509-2528.
    • Hinton, Geoffrey E., and Ruslan R. Salakhutdinov. “Reducing the dimensionality of data with neural networks.” Science 313.5786 (2006): 504-507.
    • Marc’Aurelio Ranzato, Y., Lan Boureau, and Yann LeCun. “Sparse feature learning for deep belief networks.” Advances in neural information processing systems 20 (2007): 1185-1192.
    • Bengio, Yoshua, and Yann LeCun. “Scaling learning algorithms towards AI.” Large-Scale Kernel Machines34 (2007).
    • Le Roux, Nicolas, and Yoshua Bengio. “Representational power of restricted boltzmann machines and deep belief networks.” Neural Computation 20.6 (2008): 1631-1649.
    • Sutskever, Ilya, and Geoffrey Hinton. “Temporal-Kernel Recurrent Neural Networks.” Neural Networks23.2 (2010): 239-243.
    • Le Roux, Nicolas, and Yoshua Bengio. “Deep belief networks are compact universal approximators.”Neural computation 22.8 (2010): 2192-2207.
    • Bengio, Yoshua, and Olivier Delalleau. “On the expressive power of deep architectures.” Algorithmic Learning Theory. Springer Berlin/Heidelberg, 2011.
    • Montufar, Guido F., and Jason Morton. “When Does a Mixture of Products Contain a Product of Mixtures?.”arXiv preprint arXiv:1206.0387 (2012).
  • Classification

    • The Manifold Tangent Classifier, Salah Rifai, Yann Dauphin, Pascal Vincent, Yoshua Bengio and Xavier Muller, in: NIPS’2011.
    • “Discriminative Learning of Sum-Product Networks.“, Gens, Robert, and Pedro Domingos, NIPS 2012 Best Student Paper.
  • Large Scale Deep Learning

    • Building High-level Features Using Large Scale Unsupervised Learning Quoc V. Le, Marc’Aurelio Ranzato, Rajat Monga, Matthieu Devin, Kai Chen, Greg S. Corrado, Jeffrey Dean, and Andrew Y. Ng, ICML 2012.
    • Bengio, Yoshua, et al. “Neural probabilistic language models.“ Innovations in Machine Learning (2006): 137-186. Specifically Section 3 of this paper discusses the asynchronous SGD.
    • Goodfellow, I., Warde-Farley, D., Mirza, M., Courville, A., and Bengio, Y. (2013). Maxout networks. Technical Report, Universite de Montreal.
  • Recurrent Networks

    • Training Recurrent Neural Networks, Ilya Sutskever, PhD Thesis, 2012.
    • Bengio, Yoshua, Patrice Simard, and Paolo Frasconi.“Learning long-term dependencies with gradient descent is difficult.” Neural Networks, IEEE Transactions on 5.2 (1994): 157-166.
    • Mikolov Tomáš: Statistical Language Models based on Neural Networks. PhD thesis, Brno University of Technology, 2012.
    • Hochreiter, Sepp, and Jürgen Schmidhuber.“Long short-term memory.” Neural computation 9.8 (1997): 1735-1780.
    • Hochreiter, S., Bengio, Y., Frasconi, P., & Schmidhuber, J. (2001). Gradient flow in recurrent nets: the difficulty of learning long-term dependencies.
    • Schmidhuber, J. (1992). Learning complex, extended sequences using the principle of history compression. Neural Computation, 4(2), 234-242.
    • Graves, A., Fernández, S., Gomez, F., & Schmidhuber, J. (2006, June). Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd international conference on Machine learning (pp. 369-376). ACM.
  • Hyper Parameters

    • “Practical Bayesian Optimization of Machine Learning Algorithms”, Jasper Snoek, Hugo Larochelle, Ryan Adams, NIPS 2012.
    • Random Search for Hyper-Parameter Optimization, James Bergstra and Yoshua Bengio (2012), in: Journal of Machine Learning Research, 13(281–305).
    • Algorithms for Hyper-Parameter Optimization, James Bergstra, Rémy Bardenet, Yoshua Bengio and Balázs Kégl, in: NIPS’2011, 2011.
  • Optimization

    • Training Deep and Recurrent Neural Networks with Hessian-Free Optimization, James Martens and Ilya Sutskever, Neural Networks: Tricks of the Trade, 2012.
    • Schaul, Tom, Sixin Zhang, and Yann LeCun. “No More Pesky Learning Rates.” arXiv preprint arXiv:1206.1106 (2012).
    • Le Roux, Nicolas, Pierre-Antoine Manzagol, and Yoshua Bengio. “Topmoumoute online natural gradient algorithm.” Neural Information Processing Systems (NIPS). 2007.
    • Bordes, Antoine, Léon Bottou, and Patrick Gallinari. “SGD-QN: Careful quasi-Newton stochastic gradient descent.” The Journal of Machine Learning Research 10 (2009): 1737-1754.
    • Glorot, Xavier, and Yoshua Bengio. “Understanding the difficulty of training deep feedforward neural networks.” Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS’10). Society for Artificial Intelligence and Statistics. 2010.
    • Glorot, Xavier, Antoine Bordes, and Yoshua Bengio. “Deep Sparse Rectifier Networks.“ Proceedings of the 14th International Conference on Artificial Intelligence and Statistics. JMLR W&CP Volume. Vol. 15. 2011.
    • “Deep learning via Hessian-free optimization.”Martens, James. Proceedings of the 27th International Conference on Machine Learning (ICML). Vol. 951. 2010.
    • Hochreiter, Sepp, and Jürgen Schmidhuber. “Flat minima.” Neural Computation, 9.1 (1997): 1-42.
  • Unsupervised Feature Learning

    • Salakhutdinov, Ruslan, and Geoffrey E. Hinton.“Deep boltzmann machines.” Proceedings of the international conference on artificial intelligence and statistics. Vol. 5. No. 2. Cambridge, MA: MIT Press, 2009.
    • Scholarpedia page on Deep Belief Networks.
    • Deep Boltzmann Machines

      • An Efficient Learning Procedure for Deep Boltzmann Machines, Ruslan Salakhutdinov and Geoffrey Hinton, Neural Computation August 2012, Vol. 24, No. 8: 1967 — 2006.
      • Montavon, Grégoire, and Klaus-Robert Müller. “Deep Boltzmann Machines and the Centering Trick.“ Neural Networks: Tricks of the Trade (2012): 621-637.
      • Salakhutdinov, Ruslan, and Hugo Larochelle. “Efficient learning of deep boltzmann machines.“ International Conference on Artificial Intelligence and Statistics. 2010.
      • Salakhutdinov, Ruslan. Learning deep generative models. Diss. University of Toronto, 2009.
    • RBMs

      • Large-Scale Feature Learning With Spike-and-Slab Sparse Coding, Ian Goodfellow, Aaron Courville and Yoshua Bengio, in: ICML’2012
      • Unsupervised Models of Images by Spike-and-Slab RBMs, Aaron Courville, James Bergstra and Yoshua Bengio, in: ICML’2011
  • Autoencoders

    • Regularized Auto-Encoders Estimate Local Statistics, Guillaume Alain, Yoshua Bengio and Salah Rifai, Université de Montréal, arXiv report 1211.4246, 2012
    • A Generative Process for Sampling Contractive Auto-Encoders, Salah Rifai, Yoshua Bengio, Yann Dauphin and Pascal Vincent, in: ICML’2012, Edinburgh, Scotland, U.K., 2012
    • Contracting Auto-Encoders: Explicit invariance during feature extraction, Salah Rifai, Pascal Vincent, Xavier Muller, Xavier Glorot and Yoshua Bengio, in: ICML’2011
    • Disentangling factors of variation for facial expression recognition, Salah Rifai, Yoshua Bengio, Aaron Courville, Pascal Vincent and Mehdi Mirza, in: ECCV’2012.
    • Vincent, Pascal, et al. “Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion.“ The Journal of Machine Learning Research 11 (2010): 3371-3408.
    • Vincent, Pascal. “A connection between score matching and denoising autoencoders.” Neural computation 23.7 (2011): 1661-1674.
    • Chen, Minmin, et al. “Marginalized denoising autoencoders for domain adaptation.“ arXiv preprint arXiv:1206.4683 (2012).
  • Miscellaneous

    • The ICML 2009 Workshop on Learning Feature Hierarchies webpage has a reading list.
    • Stanford’s UFLDL Recommended Readings.
    • The LISApublic wiki has a reading list and abibliography.
    • Geoff Hinton has readings NIPS 2007 tutorial.
    • The LISA publications database contains a deep architectures category.
    • A very brief introduction to AI, Machine Learning, and Deep Learning in Yoshua Bengio‘s IFT6266 graduate class
Last modified on October 10, 2013, at 11:07 am by Caglar Gulcehre
本文转载自:deeplearning.net
» 本文链接:http://www.52ml.net/4102.html
» 转载请注明来源:我爱机器学习(52ml.net) » 《Deep Learning经典论文列表(Reading List)》

Deep Learning经典论文列表(Reading List)相关推荐

  1. 机器学习——深度学习(Deep Learning)经典资料

    Deep Learning是机器学习中一个非常接近AI的领域,其动机在于建立.模拟人脑进行分析学习的神经网络,最近研究了机器学习中一些深度学习的相关知识,本文给出一些很有用的资料和心得. Key Wo ...

  2. TopPaper:AI 初学者经典论文列表

    无论你是学习机器学习还是深度学习,看学术论文必不可少.作为 AI 初学者而言,如何挑选更适合自己.更容易学习的 AI 论文尤为重要!真的会起到事半功倍的效果. 今天给大家推荐一个非常不错的 GitHu ...

  3. 【学术相关】TopPaper:AI 初学者经典论文列表

    无论你是学习机器学习还是深度学习,看学术论文必不可少.作为 AI 初学者而言,如何挑选更适合自己.更容易学习的 AI 论文尤为重要!真的会起到事半功倍的效果. 今天给大家推荐一个非常不错的 GitHu ...

  4. scDeepCluster:Clustering single-cell RNA-seq data with a model-based deep learning approach论文解读

    这是2019年发表于nature子刊machine intelligence的一篇论文,作者是Tian Tian , Ji Wan, Qi Song and Zhi Wei.论文主要是提出了一个新的框 ...

  5. Prediction of Multidrug-Resistant TB from CT Pulmonary Images Based on Deep Learning Techniques论文总结

    Prediction of Multidrug-Resistant TB from CT Pulmonary Images Based on Deep Learning Techniques 解决的问 ...

  6. 《Medical image classification using synergic deep learning》论文笔记

    利用协同深度学习进行医学图像分类 0 Abstract 医学图像分类在计算机辅助诊断.医学图像检索和医学图像挖掘中是一个非常重要的任务.尽管深度学习相对于传统的手工标注特征的方法有明显的优势,但是因为 ...

  7. 论文阅读(3):Image-Based 3D Object Reconstruction:State-of-the-Art and Trends in the Deep Learning Era

    论文标题:Image-Based 3D Object Reconstruction:State-of-the-Art and Trends in the Deep Learning Era 论文类型: ...

  8. Deep learning 资源、NLP词向量和语言模型

    转自:http://blog.csdn.net/xceman1997/article/details/38423541 在网上总能发现一些感兴趣的东西,从前是直接转载过来,现在发现太多了,还是汇总ur ...

  9. (转)Deep Learning深度学习相关入门文章汇摘

    转自:http://farmingyard.diandian.com/post/2013-04-07/40049536511 摘: 应用实例 计算机视觉. ImageNet Classificatio ...

  10. 机器学习——深度学习(Deep Learning)

    Deep Learning是机器学习中一个非常接近AI的领域,其动机在于建立.模拟人脑进行分析学习的神经网络,近期研究了机器学习中一些深度学习的相关知识,本文给出一些非常实用的资料和心得. Key W ...

最新文章

  1. Confluence 6 用户目录图例 - 可读写连接 LDAP
  2. 《机器学习》 第 5 章 神经网络
  3. 浏览器加载和渲染html的顺序
  4. maven 根据P参数值打包动态修改properties文件中值或一定properties
  5. Struts 1基础入门
  6. Vue学习(动态组件、组件路由缓存keepalive)-学习笔记
  7. win8: 清除iframe的缓存
  8. ext的另一个table布局的例子
  9. url上接收到 el表达式 不渲染_一文摸透从输入URL到页面渲染的过程
  10. GRPC: 如何优雅关闭进程(graceful shutdown)
  11. python字符串函数的find方法_Python string模块 字符串函数方法操作教程
  12. Go语言的变量、函数、Socks5代理服务器
  13. 计算机硬件维修教材,计算机硬件组装与维护教程(21世纪高等教育计算机规划教材)...
  14. 个人工作用SQL短句,不定时更新
  15. 小程序云开发实现上传图片功能
  16. 74HC595芯片应用
  17. Ubuntu关机重启卡死在关机动画界面
  18. 中间件常见安全漏洞(转载)
  19. jfreechart折线图+柱状图、柱状图(堆叠)+折线图、饼状图、环形图
  20. Echarts实现省级地图的两种方法(以浙江省为例)

热门文章

  1. MySQL 刷脏页问题
  2. Redis 配置文件详解
  3. Spring之切入点表达式
  4. javaSE基础知识 知识点(一)
  5. python多进程编程实例_Python多进程并发(multiprocessing)用法实例详解
  6. ftp 上传文件夹_8uftp上传工具,8uftp上传工具的使用方法
  7. linux拨号日志,Linux系统日志管理:(1)连接时间日志
  8. 事务对性能影响_mysql之事务
  9. Java基础:接口多态的综合案例 —— 笔记本电脑
  10. 开源API网关系统:Kong简介