Reading lists for new LISA students(转)
Research in General
Basics of machine learning
Basics of deep learning
Practical recommendations for gradient-based training of deep architectures
Quick’n’dirty introduction to deep learning: Advances in Deep Learning
Contractive auto-encoders: Explicit invariance during feature extraction
An Analysis of Single Layer Networks in Unsupervised Feature Learning
The importance of Encoding Versus Training With Sparse Coding and Vector Quantization
Feedforward nets
“Improving Neural Nets with Dropout” by Nitish Srivastava
“What is the best multi-stage architecture for object recognition?”
MCMC
Radford Neal’s Review Paper (old but still very comprehensive)
Restricted Boltzmann Machines
Unsupervised learning of distributions of binary vectors using 2-layer networks
Training restricted Boltzmann machines using approximations to the likelihood gradient
Tempered Markov Chain Monte Carlo for training of Restricted Boltzmann Machine
Enhanced Gradient for Training Restricted Boltzmann Machines
Using fast weights to improve persistent contrastive divergence
Training Products of Experts by Minimizing Contrastive Divergence
Boltzmann Machines
Deep Boltzmann Machines (Salakhutdinov & Hinton)
A Two-stage Pretraining Algorithm for Deep Boltzmann Machines
Regularized Auto-Encoders
Regularization
Stochastic Nets & GSNs
Others
Slow, Decorrelated Features for Pretraining Complex Cell-like Networks
What Regularized Auto-Encoders Learn from the Data Generating Distribution
Recurrent Nets
Learning long-term dependencies with gradient descent is difficult
Learning recurrent neural networks with Hessian-free optimization
On the importance of momentum and initialization in deep learning,
Long short-term memory (Hochreiter & Schmidhuber)
Long Short-Term Memory in Echo State Networks: Details of a Simulation Study
The "echo state" approach to analysing and training recurrent neural networks
Backpropagation-Decorrelation: online recurrent learning with O(N) complexity
New results on recurrent network training:Unifying the algorithms and accelerating convergence
Convolutional Nets
ImageNet Classification with Deep Convolutional Neural Networks, Alex Krizhevsky, Ilya Sutskever, Geoffrey E Hinton, NIPS 2012.
Optimization issues with DL
Knowledge Matters: Importance of Prior Information for Optimization
Practical recommendations for gradient-based training of deep architectures
Hessian Free
Natural Gradient (TONGA)
NLP + DL
Distributed Representations of Words and Phrases and their Compositionality
Dynamic Pooling and Unfolding Recursive Autoencoders for Paraphrase Detection
CV+RBM
CV + DL
Scaling Up
DL + Reinforcement learning
Graphical Models Background
An Introduction to Graphical Models (Mike Jordan, brief course notes)
A View of the EM Algorithm that Justifies Incremental, Sparse and Other Variants (Neal & Hinton, important paper to the modern understanding of Expectation-Maximization)
A Unifying Review of Linear Gaussian Models (Roweis & Ghahramani, ties together PCA, factor analysis, hidden Markov models, Gaussian mixtures, k-means, linear dynamical systems)
An Introduction to Variational Methods for Graphical Models (Jordan et al, mean-field, etc.)
Writing
Software documentation
Python, Theano, Pylearn2, Linux (bash) (at least the 5 first sections), git (5 first sections), github/contributing to it (Theano doc), vim tutorial or emacs tutorial
Software lists of built-in commands/functions
Other Software stuff to know about:
screen
ssh
ipython
matplotlib
Reading lists for new LISA students(转)的更多相关文章
- Reading Lists
* Non-academic 1. Slowing Down to the Speed of Life, by Richard Carlson and Joseph Bailey.2. Your Mo ...
- deep learning 的综述
从13年11月初开始接触DL,奈何boss忙or 各种问题,对DL理解没有CSDN大神 比如 zouxy09等 深刻,主要是自己觉得没啥进展,感觉荒废时日(丢脸啊,这么久....)开始开文,即为记录自 ...
- 深度学习阅读列表 Deep Learning Reading List
Reading List List of reading lists and survey papers: Books Deep Learning, Yoshua Bengio, Ian Goodfe ...
- Reading With Purpose: A grand experiment
Reading With Purpose: A grand experiment This is the preface to a set of notes I'm writing for a sem ...
- Deep Learning 和 Knowledge Graph howto
领军大家: Geoffrey E. Hinton http://www.cs.toronto.edu/~hinton/ 阅读列表: reading lists and survey papers fo ...
- Courses on Turbulence
Courses on Turbulence Table of Contents 1. Lecture 1.1. UIUC Renewable energy and turbulent environm ...
- The Ph.D. Grind
The Ph.D. Grind A Ph.D. Student Memoir Summary The Ph.D. Grind, a 122-page e-book, is the first know ...
- QuantStart量化交易文集
Over the last seven years more than 200 quantitative finance articles have been written by members o ...
- Teen Readers【青少年读者】
Teen Readers Teens and younger children are reading a lot less for fun, according to a Common Sense ...
随机推荐
- POJ 3164 Command Network ( 最小树形图 朱刘算法)
题目链接 Description After a long lasting war on words, a war on arms finally breaks out between littlek ...
- Linux基础-简单的进程操作
任务:查找一个名为firewall的进程,并且将其强制杀死 首先要使用ps -aux来查询firewall的进程ID(|grep firewall) 这样我们就得到了firewall的进程ID是653 ...
- Linux input子系统学习总结(一)---- 三个重要的结构体
一 . 总体架构 图 上层是图形界面和应用程序,通过监听设备节点,获取用户相应的输入事件,根据输入事件来做出相应的反应:eventX (X从0开始)表示 按键事件,mice 表示鼠标事件 Input ...
- 利用Mysql5.7的新特性实现多机房高可用架构【转】
再牛逼的架构也敌不过挖掘机,无论单机房内你的架构多么的高可用,多么的完善,当挖掘机挖下去那一瞬间,都是扯蛋,楼主所在的公司也被挖掘机挖断过光纤.电力线. 为什么大家都在谈论服务冗余,缓存击穿等高可用时 ...
- rsync本地及远程复制备份【原创】
1.安装rsyncyum instsall rsync 2.本地复制 rsync -auq --progress --delete /tongbu1/ /tongbu2/ rsync -auq --p ...
- 2018Java开发面经(持续更新)
不要给自己挖坑!!!不要给自己挖坑!!!不要给自己挖坑!!!如果面试官只是问你了解xxx吗,如果不是很了解,就直接说不知道,不要说知道,不然面试官深问再不知道就印象很不好! 处女面送给了头条(北京)日 ...
- Python的日志记录-logging模块的使用
一.日志 1.1什么是日志 日志是跟踪软件运行时所发生的事件的一种方法,软件开发者在代码中调用日志函数,表明发生了特定的事件,事件由描述性消息描述,同时还包含事件的重要性,重要性也称为级别或严重性. ...
- spring学习之三 数据库操作jdbcTemplate
概念 jdbcTemplate就Spring对数据库持久化技术的实现,通过它可以对数据库进行CRUD等操作. JDBCTemplate和代码实现 public void jdbcadd() { Dri ...
- poj1095
题意:给出n,要求输出第n个二叉树,二叉树编号规则如下图所示: 分析:g[i]表示有i个节点的二叉树,有多少种.f[i][j]表示有i个节点,且左子树有j个节点的树有多少种. sumg[i]表示g数组 ...
- win10 操作配置备忘
让程序自动启动 如果想要实现应用程序在所有的用户登录系统后都能自动启动,就把该应用程序的快捷方式放到"系统启动文件夹"里: C:\ProgramData\Microsoft\Win ...