Skip to content

Instantly share code, notes, and snippets.

@sidharthshah
Last active September 3, 2017 07:27
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save sidharthshah/82b840136e7c92a7ea5676c062324286 to your computer and use it in GitHub Desktop.
Save sidharthshah/82b840136e7c92a7ea5676c062324286 to your computer and use it in GitHub Desktop.
Notes from DeepLearning.TV's Youtube Videso
  • GEOFF HILTON {Father of DNN}
  • MLP {Multi Layered Perceptron} -> Vanialla Neural Network
  • Problem of Vanishing Gradient {This is what RBM Solves}

Unsupervised

RBM {Extract features and reconstruct inputs}

  • Forward/Backward Phase

Autoencoders {Family of algo to which RBM belongs to}

  • Idenfity patterns in dataset
  • Denoising
  • Contractive
  • Unlabelled Input -> Try to Reconstruct Input as Possible
  • Feature Extraction Engine
  • Shallow
  • Backprogation with Lost
  • Deep Auto-Encoder [28 x 28] -> 30
  • Better than PCA

Supervised

DBN {Feed Forward}

  • Combine multiple RBMs
  • Hidden Layer of one RBM is Visible Layer of another RBM
  • Useful in scenarios where we have small sample of Trained Data
  • Training process get completed in reasonable amount of time

CNN {Feed Forward}

  • Machine Vision is where CNN is used
  • ImageNet competition
  • Andrej Karpathy's Note {http://cs231n.github.io/}
  • Layers:
    • Convolution Layer
      • Technical Operation {Convolution} to search for pattern
    • Relu {Rectified Linear Unit}
      • Used for reducing training time for early layers
    • Pooling
      • Used for dimensionality reduction
    • Fully Connected Layer
      • Allows network to classify data from samples based on patterns discovered by Polling layer
  • Since this is supervised method, take large sample of labeled data

RNN

  • If patterns change over time is where RNN is useful
  • Application {Speech Recognization, Driverless Car}
  • Output of layer is fed back to the same layer along with inputs
  • Applications
    • Sequences of Input to Produce Sequences of Output
    • Image Captioning {Singel Input, Multiple Output}
    • Document Classification {Single Input, Single Output}
    • Video Classification {Multiple Input, Multiple Output}
    • Forecasting {with time Delays}
  • Stacking
  • Difficult to train because of Vanishing Gradients
  • Solution:
    • Gating {Method of deciding for network when to forget current input or consider current input}
      • LSTM
      • GRU
    • Gradient Clipping
    • Better Optimizers
    • Steeper Gates
  • GPUs prefered methods of training
  • RNN suited for timeseries data
  • Feedforward: Classification/Regression
  • Recurrent Neural Network: Forecaster

RNTN {Recursive Neural Tensor Nets}

  • For applications like Parsing
  • 3 Basic Component
    • Root {Fires out Class and Score}
    • Child Nodes {Leafs}
  • Collection of Recursive Binary Tress
  • Score repesent quality
  • Text
    • Parsing
    • Sentiement Analysis
  • Image Parsing {Break Image into many different components}
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment