Skip to content

Instantly share code, notes, and snippets.

View amoudgl's full-sized avatar

Abhinav Moudgil amoudgl

View GitHub Profile
@amoudgl
amoudgl / tx_lopt.py
Last active January 3, 2024 14:38
Code for our ICML23W paper "Learning to Optimize with Recurrent Hierarchical Transformers"
# Copyright 2021 DeepMind Technologies Limited. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
@amoudgl
amoudgl / dummyfig.tex
Created May 12, 2021 03:40 — forked from dpgettings/dummyfig.tex
Fancy placeholder figures in LaTeX
%% This part goes in preamble
\newcommand{\dummyfig}[1]{
\centering
\fbox{
\begin{minipage}[c][0.33\textheight][c]{0.5\textwidth}
\centering{#1}
\end{minipage}
}
}
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
@amoudgl
amoudgl / setup_visdom.md
Last active July 4, 2023 11:27
Setup Visdom on Remote Server

Setup Visdom on Remote Server

Install visdom on your local system and remote server.

pip3 install visdom

On remote server, do:

@amoudgl
amoudgl / README.md
Last active April 17, 2024 06:25
TLP benchmark
@amoudgl
amoudgl / deep-qlearning.py
Created March 30, 2017 09:16
Deep Q-learning for Frozenlake OpenAI gym environment
# simple neural network implementation of qlearning
import gym
from gym import wrappers
import numpy as np
import tensorflow as tf
# build environment
env = gym.make("FrozenLake-v0")
env = wrappers.Monitor(env, '/tmp/frozenlake-qlearning', force=True)
n_obv = env.observation_space.n
# monte carlo policy gradient algorithm
# use neural network to decide the policy
# from observations and rewards, update the parameters of the neural networks to optimize the policy
import numpy as np
import tensorflow as tf
import gym
from gym import wrappers
# initialize constants
@amoudgl
amoudgl / random_guessing.py
Created January 16, 2017 06:17
Random Guessing Algorithm for Cartpole Environment
# random guessing algorithm
# generate 10000 random configurations of the model's parameters and pick the one that achieves the best cumulative reward.
# optimize it for weighted sum
import gym
from gym import wrappers
import numpy as np
env = gym.make('CartPole-v0')
env = wrappers.Monitor(env, '/tmp/cartpole-random-guessing', force=True)
@amoudgl
amoudgl / hill_climbing.py
Created January 16, 2017 06:16
Hill Climbing Algorithm for Cartpole Environment
# hill climbing algorithm
# generate a random configuration of the parameters, add small amount of noise to the parameters and evaluate the new parameter configuration
# if new configuration is better than old one, discard the old one and accept the new one
# optimize it for weighted sum
# returns the net episode reward
def get_episode_reward(env, observation, params):
t = 0
net_reward = 0
while (t < 1000):
@amoudgl
amoudgl / policy_gradient.py
Created January 16, 2017 06:15
Policy Gradient Algorithm for Cartpole Environment
# monte carlo policy gradient algorithm
# use neural network to decide the policy
# from observations and rewards, update the parameters of the neural networks to optimize the policy
import numpy as np
import tensorflow as tf
import gym
from gym import wrappers
# initialize constants