Skip to content

Instantly share code, notes, and snippets.

Show Gist options
  • Save srogovtsev/dc819112fa14427f80b9250a697eefb0 to your computer and use it in GitHub Desktop.
Save srogovtsev/dc819112fa14427f80b9250a697eefb0 to your computer and use it in GitHub Desktop.

Sources:

Dropout

  • Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2014). Dropout: A simple way to prevent neural networks from overfitting. Journal of Machine Learning Research, 15 , 1929–1958: http://jmlr.org/papers/v15/srivastava14a.html

Initialization

Optimization

Hyperparameter tuning

Batch Norm

  • Ioffe, S. and Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift: https://arxiv.org/abs/1502.03167

Sources:

Classic Networks

Resnets

Networks in Networks and 1x1 Convolutions

Inception Networks

Convolutional Implementation of Sliding Windows

  • Pierre Sermanet, David Eigen, Xiang Zhang, Michael Mathieu, Rob Fergus, Yann LeCun, "OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks": https://arxiv.org/abs/1312.6229

Bounding Box Predictions

Region Proposals

Siamese Network

Triplet Loss

What are deep ConvNets learning?

Neural Style

Sources

RNN

Word Embeddings

  • Andrej Karpathy. The Unreasonable Effectiveness of Recurrent Neural Networks. http://karpathy.github.io/2015/05/21/rnn-effectiveness/. Published 2015.

  • Pennington J, Socher R, Manning C. Glove: Global Vectors for Word Representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Doha, Qatar: Association for Computational Linguistics; 2014:1532-1543: https://nlp.stanford.edu/pubs/glove.pdf

  • Mikolov T, Sutskever I, Chen K, Corrado G, Dean J. Distributed Representations of Words and Phrases and their Compositionality. arXiv:13104546. October 2013: http://arxiv.org/abs/1310.4546.

  • Mikolov T, Chen K, Corrado G, Dean J. Efficient Estimation of Word Representations in Vector Space. arXiv:13013781. January 2013: http://arxiv.org/abs/1301.3781.

  • Mikolov T, Yih W, Zweig G. Linguistic Regularities in Continuous Space Word Representations. In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics; 2013:746–751: http://aclweb.org/anthology/N13-1090.

  • Maaten L van der, Hinton GE. Visualizing Data using t-SNE. In: ; 2008.

  • Bengio Y, Ducharme R, Vincent P, Jauvin C. A Neural Probabilistic Language Model. 2003:19.

  • Bolukbasi T, Chang K-W, Zou J, Saligrama V, Kalai A. Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings. July 2016: https://arxiv.org/abs/1607.06520v1.

Sequence to Sequence Models

  • Cho K, van Merrienboer B, Gulcehre C, et al. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. arXiv:14061078. June 2014: http://arxiv.org/abs/1406.1078.

  • Sutskever I, Vinyals O, Le QV. Sequence to Sequence Learning with Neural Networks. In: Proc. NIPS. Montreal, CA; 2014. http://arxiv.org/abs/1409.3215.

  • Papineni K, Roukos S, Ward T, Zhu W-J. Bleu: a Method for Automatic Evaluation of Machine Translation. In: Proceedings of 40th Annual Meeting of the Association for Computational Linguistics. Philadelphia, Pennsylvania, USA: Association for Computational Linguistics; 2002:311–318. doi:10.3115/1073083.1073135: https://www.aclweb.org/anthology/P02-1040.pdf

Image Captioning

  • Mao J, Xu W, Yang Y, Wang J, Huang Z, Yuille A. Deep Captioning with Multimodal Recurrent Neural Networks (m-RNN). arXiv:14126632. December 2014: http://arxiv.org/abs/1412.6632.

  • Karpathy A, Fei-Fei L. Deep Visual-Semantic Alignments for Generating Image Descriptions. December 2014: https://arxiv.org/abs/1412.2306v2.

  • Vinyals O, Toshev A, Bengio S, Erhan D. Show and Tell: A Neural Image Caption Generator. arXiv:14114555. November 2014: http://arxiv.org/abs/1411.4555.

Attention-based Models

  • Bahdanau D, Cho K, Bengio Y. Neural Machine Translation by Jointly Learning to Align and Translate. arXiv:14090473. September 2014: http://arxiv.org/abs/1409.0473.

  • Xu K, Ba J, Kiros R, et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention. arXiv:150203044. February 2015: http://arxiv.org/abs/1502.03044.

Speech Recognition

  • Graves A, Fernandez S, Gomez F, Schmidhuber J. Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks. 2006:8: ftp://ftp.idsia.ch/pub/juergen/icml2006.pdf
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment