Files
Gymnasium/docs/environments/classic_control/mountain_car.md
Manuel Goulão 91769fc862 Add docs (#13)
2022-09-13 20:27:34 +01:00

3.3 KiB

AUTOGENERATED, title
AUTOGENERATED title
DO NOT EDIT FILE DIRECTLY Mountain Car

Mountain Car

:width: 200px
:name: mountain_car

This environment is part of the Classic Control environments. Please read that page first for general information.

Action Space Discrete(3)
Observation Shape (2,)
Observation High [0.6 0.07]
Observation Low [-1.2 -0.07]
Import gymnasium.make("MountainCar-v0")

Description

The Mountain Car MDP is a deterministic MDP that consists of a car placed stochastically at the bottom of a sinusoidal valley, with the only possible actions being the accelerations that can be applied to the car in either direction. The goal of the MDP is to strategically accelerate the car to reach the goal state on top of the right hill. There are two versions of the mountain car domain in gymnasium: one with discrete actions and one with continuous. This version is the one with discrete actions.

This MDP first appeared in Andrew Moore's PhD Thesis (1990)

@TECHREPORT{Moore90efficientmemory-based,
    author = {Andrew William Moore},
    title = {Efficient Memory-based Learning for Robot Control},
    institution = {University of Cambridge},
    year = {1990}
}

Observation Space

The observation is a ndarray with shape (2,) where the elements correspond to the following:

Num Observation Min Max Unit
0 position of the car along the x-axis -Inf Inf position (m)
1 velocity of the car -Inf Inf position (m)

Action Space

There are 3 discrete deterministic actions:

Num Observation Value Unit
0 Accelerate to the left Inf position (m)
1 Don't accelerate Inf position (m)
2 Accelerate to the right Inf position (m)

Transition Dynamics:

Given an action, the mountain car follows the following transition dynamics:

velocityt+1 = velocityt + (action - 1) * force - cos(3 * positiont) * gravity

positiont+1 = positiont + velocityt+1

where force = 0.001 and gravity = 0.0025. The collisions at either end are inelastic with the velocity set to 0 upon collision with the wall. The position is clipped to the range [-1.2, 0.6] and velocity is clipped to the range [-0.07, 0.07].

Reward:

The goal is to reach the flag placed on top of the right hill as quickly as possible, as such the agent is penalised with a reward of -1 for each timestep.

Starting State

The position of the car is assigned a uniform random value in [-0.6 , -0.4]. The starting velocity of the car is always assigned to 0.

Episode End

The episode ends if either of the following happens:

  1. Termination: The position of the car is greater than or equal to 0.5 (the goal position on top of the right hill)
  2. Truncation: The length of the episode is 200.

Arguments

gymnasium.make('MountainCar-v0')

Version History

  • v0: Initial versions release (1.0.0)