Skip to content

Instantly share code, notes, and snippets.

@sherlock42
sherlock42 / smallberta_pretraining.ipynb
Created February 27, 2020 13:53 — forked from aditya-malte/smallberta_pretraining.ipynb
smallBERTa_Pretraining.ipynb
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
@sherlock42
sherlock42 / fast_ai_mooc_important_points.md
Created March 5, 2019 09:38 — forked from bhavikngala/fast_ai_mooc_important_points.md
This gist contains a list of important points from fast.ai "practical deep learning for coders" and "cutting edge deep learning for coders" MOOC

This gist contains a list of points I found very useful while watching the fast.ai "Practical deep learning for coders" and "Cutting edge deep learning for coders" MOOC by Jeremy Howard and team. This list may not be complete as I watched the video at 1.5x speed on marathon but I did write down as many things I found to be very useful to get a model working. A fair warning the points are in no particular order, you may find the topics are all jumbled up.

Before beginning, I want to thank Jeremy Howard, Rachel Thomas, and the entire fast.ai team in making this awesome practically oriented MOOC.

  1. Progressive image resolution training: Train the network on lower res first and then increase the resolution to get better performance. This can be thought of as transfer learning from the same dataset but at a different resolution. There is one paper by NVIDIA as well that used such an approach to train GANs.

  2. Cyclical learning rates: Gradually increasing the learning rate initially helps to avoid getting stuc

@sherlock42
sherlock42 / twitter_sentiment_analysis_convnet.py
Created October 11, 2018 16:53 — forked from giuseppebonaccorso/twitter_sentiment_analysis_convnet.py
Twitter Sentiment Analysis with Gensim Word2Vec and Keras Convolutional Networks
import keras.backend as K
import multiprocessing
import tensorflow as tf
from gensim.models.word2vec import Word2Vec
from keras.callbacks import EarlyStopping
from keras.models import Sequential
from keras.layers.core import Dense, Dropout, Flatten
from keras.layers.convolutional import Conv1D
@sherlock42
sherlock42 / log.md
Created August 24, 2018 06:41 — forked from abhinishetye/100daysLog.md
100DaysofMLCode

100 Days Of ML Code

Hi! I am Abhini, a Machine Learning Enthusiast and this is my log for the 100DaysOfMLCode Challenge

Day 1: July 08, 2018

Today's Progress: Understood the basics of Neural Network and how to build ANN. Also practiced Python on Hackerrank.

Thoughts: Cleared up my concepts on ANN in which I had earlier found confusing like Activation and Cost functions, Batch and Stochastic Gradient Descent and Backpropagation.