ray.rllib.utils.replay_buffers.multi_agent_replay_buffer.MultiAgentReplayBuffer#
- class ray.rllib.utils.replay_buffers.multi_agent_replay_buffer.MultiAgentReplayBuffer(capacity: int = 10000, storage_unit: str = 'timesteps', num_shards: int = 1, replay_mode: str = 'independent', replay_sequence_override: bool = True, replay_sequence_length: int = 1, replay_burn_in: int = 0, replay_zero_init_states: bool = True, underlying_buffer_config: dict | None = None, **kwargs)[source]#
Bases:
ReplayBufferA replay buffer shard for multiagent setups.
This buffer is meant to be run in parallel to distribute experiences across
num_shardsshards. Unlike simpler buffers, it holds a set of buffers - one for each policy ID.DeveloperAPI: This API may change across minor Ray releases.
Methods
Initializes a MultiAgentReplayBuffer instance.
Adds a batch to the appropriate policy's replay buffer.
Calls the given function with this Actor instance.
Returns the computer's network name.
Returns all local state.
Ping the actor.
Deprecated in favor of new ReplayBuffer API.
Samples a MultiAgentBatch of
num_itemsper one policy's buffer.Restores all local state to the provided
state.Returns the stats of this buffer and all underlying buffers.