* initial implementaion of ppo2_rnn. * set lstm memory as tf.GraphKeys.LOCAL_VARIABLES. * replace dones with tf.placeholder_with_default. * improves for 'play' option. * removed unnecessary TODO . * improve lstm code. * move learning rate placeholer to optimizer scope. * support the microbatched model. * sync cnn lstm layer with originals. * add cnn_lnlstm layer. * fix a case when `states` is None. * add initial_state variable to help test. * make ppo2 rnn test available. * rename 'obs' with 'observations'. rename 'transition' with 'transitions'. fix forgetting `dones` in the replay buffer. fix a misuse of `states` and `next_states` in the replay buffer. * make initialization once. make `test_fixed_sequence` compatible with ppo2. * adjust input shape. * fix checking of a model input args in `simple_test` function. * disable warning on purpose. * support the play. * improve scopes to compatible with multiple models (i.e, other tensorflow global/local variables) * clean the scope of ppo2 policy model. * name the memory variable of PPO RNNs more describly * wrap the initializations in ppo2. * remove redundant lines. * update `REAMD.md`. * add RNN layers. * add the result of HalfCheeta-v2 env experiment. * correct a typo. * add RNN class. * rename `nlstm` with `num_units` in RNN builder functions. * remove state saving. * reuse RNNs in a2c.utils. * revert baselines/run.py. * replace `ppo2.step()` with original interface. * revert `baselines/common/tests/util.py`. * remove redundant lines. * revert `baselines/common/test/util.py` tob875fb7
. * remove `states` variable. * move RNN class to `baselines/ppo2/layers.py' and revert `baselines/common/models.py` to858afa8
. * rename `model.step_as_dict` with `model.step_with_dict`. * removed `ppo_lstm_mlp`. * fix02e26fd
.
2 lines
119 B
Python
2 lines
119 B
Python
from baselines.ppo2.layers import ppo_lstm, ppo_cnn_lstm, ppo_cnn_lnlstm # pylint: disable=unused-import # noqa: F401
|