mirror of
https://github.com/Farama-Foundation/Gymnasium.git
synced 2025-08-29 01:27:29 +00:00
67 lines
1.9 KiB
Python
67 lines
1.9 KiB
Python
"""
|
|
Simple environment with known optimal policy and value function.
|
|
|
|
Action 0 then 0 yields randomly -1 or 1 reward and terminates the session.
|
|
Action 0 then 1 yields randomly 0, 0, or 9 reward and terminates the session.
|
|
Action 0 then 0 yields randomly 0 or 2 reward and terminates the session.
|
|
Action 1 then 1 yields randomly 2 or 3 reward and terminates the session.
|
|
|
|
Optimal policy: action 0 then 1.
|
|
|
|
Optimal value function v(observation): (this is a fully observable MDP so observation==state)
|
|
|
|
v(0)= 3 (you get observation 0 after taking action 0)
|
|
v(1)= 2.5 (you get observation 1 after taking action 1)
|
|
v(2)= 3 (you get observation 2 in the starting state)
|
|
"""
|
|
|
|
import gym
|
|
import random
|
|
from gym import spaces
|
|
from gym.utils import seeding
|
|
|
|
class TwoRoundNondeterministicRewardEnv(gym.Env):
|
|
def __init__(self):
|
|
self.action_space = spaces.Discrete(2)
|
|
self.observation_space = spaces.Discrete(3)
|
|
self._reset()
|
|
|
|
def _step(self, action):
|
|
rewards = [
|
|
[
|
|
[-1, 1], #expected value 0
|
|
[0, 0, 9] #expected value 3. This is the best path.
|
|
],
|
|
[
|
|
[0, 2], #expected value 1
|
|
[2, 3] #expected value 2.5
|
|
]
|
|
]
|
|
|
|
assert self.action_space.contains(action)
|
|
|
|
if self.firstAction is None:
|
|
self.firstAction = action
|
|
reward = 0
|
|
done = False
|
|
else:
|
|
reward = random.choice(rewards[self.firstAction][action])
|
|
done = True
|
|
|
|
return self._get_obs(), reward, done, {}
|
|
|
|
def _get_obs(self):
|
|
if self.firstAction is None:
|
|
return 2
|
|
else:
|
|
return self.firstAction
|
|
|
|
def _reset(self):
|
|
self.firstAction = None
|
|
return self._get_obs()
|
|
|
|
def _seed(self, seed=None):
|
|
seed = seed if seed else seeding.hash_seed(seed) % 2**32
|
|
random.seed(seed)
|
|
return [seed]
|