Created
October 15, 2016 19:30
-
-
Save brotherofken/74b7994b2f607e73a19f09efb7d62aa5 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# In[] | |
class DiscretePolicy(object): | |
def __init__(self, env): | |
if not issubclass(type(env), gym.envs.toy_text.discrete.DiscreteEnv): | |
raise Exception('env should be subclass of gym.envs.toy_text.' | |
'discrete.DiscreteEnv') | |
self.env = env | |
self.policy = np.array([env.action_space.sample() for i in | |
range(self.env.nS)], | |
dtype=int) | |
def action(self, state): | |
return self.policy[state] | |
# In[] Monte Carlo Discrete Model Free Predictor | |
class MonteCarloDMFPredictor(object): | |
def __init__(self, env): | |
if not issubclass(type(env), gym.envs.toy_text.discrete.DiscreteEnv): | |
raise Exception('env should be subclass of gym.envs.toy_text.' | |
'discrete.DiscreteEnv') | |
self.env = env | |
def evaluate(self, policy, iterations=1000, discount=1., | |
every_visit=True): | |
if not isinstance(policy, DiscretePolicy): | |
raise Exception('policy should have type DiscretePolicy') | |
counts = np.zeros((self.env.nS)) | |
values = np.zeros((self.env.nS)) | |
max_time_steps = 500 | |
for i_episode in xrange(iterations): | |
# reset environment to beginning | |
observations = np.zeros((max_time_steps), dtype=np.int) | |
rewards = np.zeros((max_time_steps)) | |
# generate an episode using policy | |
observations[0] = env.reset() | |
steps = 0 | |
for t in xrange(1, max_time_steps): | |
# sample a random action | |
action = policy.action(observations[t - 1]) | |
# observe next step and get reward | |
observations[t], rewards[t - 1], done, info = env.step(action) | |
#observations[t] = observation | |
if done: | |
steps = t + 1 | |
break | |
if steps <= 1: | |
continue | |
if i_episode % 1000 == 0: | |
print 'Episode {} finished in {} steps.'.format(i_episode, | |
steps) | |
observations = observations[:steps] | |
rewards = rewards[:steps] | |
returns = np.zeros((steps)) | |
returns[-1] = rewards[-1] | |
for t in reversed(xrange(steps - 1)): | |
returns[t] = discount * returns[t + 1] + rewards[t] | |
visited = np.zeros((self.env.nS), dtype=np.bool) | |
for t in xrange(steps): | |
s = observations[t] | |
if every_visit or not visited[s]: | |
counts[s] += 1. | |
values[s] += (returns[t] - values[s]) / counts[s] | |
visited[s] = True | |
return values |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment