Back to Ray

Replay Buffer API

doc/source/rllib/package_ref/replay-buffers.rst

1.13.11.8 KB
Original Source

.. include:: /_includes/rllib/we_are_hiring.rst

.. _replay-buffer-api-reference-docs:

Replay Buffer API

.. include:: /_includes/rllib/new_api_stack.rst

The following classes don't take into account the separation of experiences from different policies, multi-agent replay buffers will be explained further below.

Replay Buffer Base Classes

.. currentmodule:: ray.rllib.utils.replay_buffers

.. autosummary:: :nosignatures: :toctree: doc/

~replay_buffer.StorageUnit
~replay_buffer.ReplayBuffer
~prioritized_replay_buffer.PrioritizedReplayBuffer
~reservoir_replay_buffer.ReservoirReplayBuffer

Public Methods

.. currentmodule:: ray.rllib.utils.replay_buffers.replay_buffer

.. autosummary:: :nosignatures: :toctree: doc/

~ReplayBuffer.sample
~ReplayBuffer.add
~ReplayBuffer.get_state
~ReplayBuffer.set_state

Multi Agent Buffers

The following classes use the above, "single-agent", buffers as underlying buffers to facilitate splitting up experiences between the different agents' policies. In multi-agent RL, more than one agent exists in the environment and not all of these agents may utilize the same policy (mapping M agents to N policies, where M <= N). This leads to the need for MultiAgentReplayBuffers that store the experiences of different policies separately.

.. currentmodule:: ray.rllib.utils.replay_buffers

.. autosummary:: :nosignatures: :toctree: doc/

~multi_agent_replay_buffer.MultiAgentReplayBuffer
~multi_agent_prioritized_replay_buffer.MultiAgentPrioritizedReplayBuffer

Utility Methods

.. autosummary:: :nosignatures: :toctree: doc/

~utils.update_priorities_in_replay_buffer
~utils.sample_min_n_steps_from_buffer