- Feature Learning
- Learning Feature Representations with K-means by Adam Coates and Andrew Y. Ng
- The devil is in the details: an evaluation of recent feature encoding methods by Chatfield et. al.
- Emergence of Object-Selective Features in Unsupervised Feature Learning by Coates, Ng
- Scaling Learning Algorithms towards AI Benjio & LeCun
- Deep Neural Nets
- Dropout: A Simple Way to Prevent Neural Networks from Overfitting by Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever and Ruslan Salakhutdinov
- Understanding the difficulty of training deep feedforward neural networks by Xavier Glorot and Yoshua Bengio
- On the difficulty of training Recurrent Neural Networks by Razvan Pascanu, Tomas Mikolov and Yoshua Bengio
- Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift by Sergey Ioffe and Christian Szegedy
- Deep Learning in Neural Networks: An Overview by Jurgen Schmidhuber
- Stochastic Gradient Descent Tricks by L´eon Bottou
- Qualitatively characterizing neural network optimization problems by Ian J. Goodfellow, Oriol Vinyals
- On Recurrent and Deep Neural Networks Phd thesis of Razvan Pascanu
- Scaling Learning Algorithms towards AI by Yann LeCun and Yoshua Benjio
- Efficient Backprop by LeCun, Bottou et al
- Towards Biologically Plausible Deep Learning by Yoshua Bengio, Dong-Hyun Lee, Jorg Bornschein, Zhouhan Lin
- Training Recurrent Neural Networks Phd thesis of Ilya Sutskever