diff --git a/docs/02_notebooks/L2_Buffer.ipynb b/docs/02_notebooks/L2_Buffer.ipynb index 65a23626b..892aaff4f 100644 --- a/docs/02_notebooks/L2_Buffer.ipynb +++ b/docs/02_notebooks/L2_Buffer.ipynb @@ -22,7 +22,7 @@ }, "source": [ "# Buffer\n", - "Replay Buffer is a very common module in DRL implementations. In Tianshou, you can consider Buffer module as as a specialized form of Batch, which helps you track all data trajectories and provide utilities such as sampling method besides the basic storage.\n", + "Replay Buffer is a very common module in DRL implementations. In Tianshou, the Buffer module can be viewed as a specialized form of Batch, designed to track all data trajectories and offering utilities like sampling methods beyond basic storage.\n", "\n", "There are many kinds of Buffer modules in Tianshou, two most basic ones are ReplayBuffer and VectorReplayBuffer. The later one is specially designed for parallelized environments (will introduce in tutorial [Vectorized Environment](https://tianshou.readthedocs.io/en/master/02_notebooks/L3_Vectorized__Environment.html)). In this tutorial, we will focus on ReplayBuffer." ] @@ -43,7 +43,7 @@ }, "source": [ "### Basic usages as a batch\n", - "Usually a buffer stores all the data in a batch with circular-queue style." + "Typically, a buffer stores all data in batches, employing a circular-queue mechanism." ] }, { @@ -164,7 +164,7 @@ }, "source": [ "### Data sampling\n", - "We keep a replay buffer in DRL for one purpose: sample data from it for training. `ReplayBuffer.sample()` and `ReplayBuffer.split(..., shuffle=True)` can both fulfill this need." + "The primary purpose of maintaining a replay buffer in DRL is to sample data for training. `ReplayBuffer.sample()` and `ReplayBuffer.split(..., shuffle=True)` can both fulfill this need." ] }, { @@ -395,7 +395,7 @@ "* CachedReplayBuffer, one main buffer with several cached buffers (higher sample efficiency in some scenarios)\n", "* ReplayBufferManager, A base class that can be inherited (may help you manage multiple buffers).\n", "\n", - "Check the documentation and the source code for more details.\n", + "Refer to the documentation and source code for further details.\n", "\n", "### Support for steps stacking to use RNN in DRL.\n", "There is an option called `stack_num` (default to 1) when initializing the ReplayBuffer, which may help you use RNN in your algorithm. Check the documentation for details." diff --git a/docs/02_notebooks/L3_Vectorized__Environment.ipynb b/docs/02_notebooks/L3_Vectorized__Environment.ipynb index a83b0bd9c..374ee2ba8 100644 --- a/docs/02_notebooks/L3_Vectorized__Environment.ipynb +++ b/docs/02_notebooks/L3_Vectorized__Environment.ipynb @@ -7,7 +7,7 @@ }, "source": [ "# Vectorized Environment\n", - "In reinforcement learning, the agent interacts with environments to improve itself. In this tutorial we will concentrate on the environment part. Although there are many kinds of environments or their libraries in DRL research, Tianshou chooses to keep a consistent API with [OPENAI Gym](https://gym.openai.com/).\n", + "In reinforcement learning, an agent engages with environments to enhance its performance. In this tutorial we will concentrate on the environment part. Although there are many kinds of environments or their libraries in DRL research, Tianshou chooses to keep a consistent API with [OPENAI Gym](https://gym.openai.com/).\n", "\n", "
\n", "\n", @@ -200,7 +200,7 @@ "Check the [documentation](https://tianshou.org/en/master/03_api/env/venvs.html) for details.\n", "\n", "### Difference between synchronous and asynchronous mode (How to choose?)\n", - "Explanation can be found at the [Parallel Sampling](https://tianshou.org/en/master/01_tutorials/07_cheatsheet.html#parallel-sampling) tutorial." + "For further insights, refer to the [Parallel Sampling](https://tianshou.org/en/master/01_tutorials/07_cheatsheet.html#parallel-sampling) tutorial." ] } ],