import logging import numpy as np from gym import envs logger = logging.getLogger(__name__) class ClipTo01ThenAverage(object): def __init__(self, num_episodes=100): self.num_episodes = num_episodes def score_evaluation(self, benchmark, env_id, episode_lengths, episode_rewards, episode_types, timestamps, initial_reset_timestamp): tasks = benchmark.task_groups[env_id] spec = envs.spec(env_id) (t_idx,) = np.where([t == 't' for t in episode_types]) # training episodes (e_idx,) = np.where([t == 'e' for t in episode_types]) # evaluation episodes if len(e_idx) == 0: # If no episodes marked for evaluation, consider # everything both a training and evaluation episode. (t_idx,) = np.where([True for t in episode_types]) (e_idx,) = np.where([True for t in episode_types]) training_lengths = np.array(episode_lengths)[t_idx] training_rewards = np.array(episode_rewards)[t_idx] evaluation_lengths = np.array(episode_lengths)[e_idx] evaluation_rewards = np.array(episode_rewards)[e_idx] # How many training timesteps have elapsed by the end of each # episode. Not to be confused with Unix timestamps. elapsed_timesteps = np.cumsum(training_lengths) scores = [] solves = [] rewards = [] _timestamps = [] for task in tasks: # Find the first episode where we're over the allotted # training timesteps. (cutoff,) = np.where(elapsed_timesteps > task.timesteps) if len(cutoff) > 0: cutoff_idx = cutoff[-1] orig_cutoff_idx = t_idx[cutoff_idx] # cutoff index in the original (allowed_e_idx,) = np.where(e_idx < orig_cutoff_idx) # restrict to earlier episodes else: # All episodes are fair game allowed_e_idx = e_idx if len(allowed_e_idx) > 0: last_timestamp = timestamps[allowed_e_idx[-1]] else: # If we don't have any evaluation episodes, then the # last valid timestamp is when we started. last_timestamp = initial_reset_timestamp # Grab the last num_episodes evaluation episodes from # before the cutoff (at which point we've gathered too # much experience). # # This probably won't work long-term but is fine for now. allowed_episode_rewards = np.array(episode_rewards)[allowed_e_idx] reward = allowed_episode_rewards[-self.num_episodes:] if len(reward) == 0: logger.info('No rewards for %s', env_id) scores.append(0) return floor = task.reward_floor ceiling = task.reward_ceiling # Grab the indexes where we reached the ceiling solved = reward >= ceiling # Linearly rescale rewards to between 0 and 1 clipped = np.clip((reward - floor) / (ceiling - floor), 0, 1) # Take the mean rescaled score score = np.sum(clipped) / self.num_episodes scores.append(score) # Record the list of solved episodes solves.append(solved) # Record the list of rewards rewards.append(reward) # Record the timestamp of the last episode timestamp _timestamps.append(last_timestamp) return { 'rewards': rewards, 'scores': scores, 'solves': solves, 'timestamps': _timestamps, } def score_benchmark(self, benchmark, episode_scores): all_scores = [] for env_id, scores in episode_scores.items(): all_scores += scores return np.mean(all_scores)