Note
Ray 2.40 uses RLlib’s new API stack by default. The Ray team has mostly completed transitioning algorithms, example scripts, and documentation to the new code base.
If you’re still using the old API stack, see New API stack migration guide for details on how to migrate.
Offline RL API#
Configuring Offline RL#
Sets the config's offline data settings. |
|
Sets LearnerGroup and Learner worker related configurations. |
Configuring Offline Recording EnvRunners#
Sets the rollout worker configuration. |
Constructing a Recording EnvRunner#
The environment runner to record the single agent case. |
Constructing OfflineData#
PublicAPI (alpha): This API is in alpha and may change before becoming stable. |
|
Sampling from Offline Data#
Constructing an OfflinePreLearner#
Class that coordinates data transformation from dataset to learner. |
|
Transforming Data with an OfflinePreLearner#
This is the default schema used if no |
|
Prepares plain data batches for training with |
|
Maps a batch of data to episodes. |
|
Maps an old stack |
|
Checks which modules in a MultiRLModule should be updated. |
|
Sets the default replay buffer. |
|
Sets the default arguments for the replay buffer. |