diff --git a/img/markov.png b/img/markov.png new file mode 100644 index 0000000..5fa1ae0 Binary files /dev/null and b/img/markov.png differ diff --git a/img/the_agent.png b/img/the_agent.png new file mode 100644 index 0000000..f38a01c Binary files /dev/null and b/img/the_agent.png differ diff --git a/tutorial.ipynb b/tutorial.ipynb index 4eeba7a..2586b69 100644 --- a/tutorial.ipynb +++ b/tutorial.ipynb @@ -138,6 +138,36 @@ "```" ] }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "slideshow": { + "slide_type": "slide" + } + }, + "outputs": [], + "source": [ + "# Importing the required packages\n", + "from time import sleep\n", + "\n", + "import matplotlib.pyplot as plt\n", + "import names\n", + "import numpy as np\n", + "from gymnasium.wrappers import RescaleAction\n", + "from IPython.display import clear_output, display\n", + "from stable_baselines3 import PPO\n", + "\n", + "from utils.helpers import (\n", + " evaluate_ares_ea_agent,\n", + " plot_ares_ea_training_history,\n", + " show_video,\n", + ")\n", + "from utils.train import ARESEACheetah, make_env, read_from_yaml\n", + "from utils.train import train as train_ares_ea\n", + "from utils.utils import NotVecNormalize" + ] + }, { "cell_type": "markdown", "metadata": { @@ -159,17 +189,7 @@ } }, "source": [ - "

Formulating the RL problem

\n", - "

Refresher from the lecture

\n", - "\n", - "We need to define:\n", - "\n", - "- Actions\n", - "- Observations\n", - "- Reward\n", - "- Environment\n", - "- Agent\n", - " \n" + "\n" ] }, { @@ -181,7 +201,6 @@ }, "source": [ "

Formulating the RL problem

\n", - "

Refresher from the lecture

\n", "\n", "We need to define:\n", "\n", @@ -201,17 +220,7 @@ } }, "source": [ - "

Formulating the RL problem

\n", - "

Refresher from the lecture

\n", - "\n", - "We need to define:\n", - "\n", - "- Actions\n", - "- Observations\n", - "- Reward\n", - "- Environment\n", - "- Agent\n", - " \n" + "\n" ] }, { @@ -222,17 +231,18 @@ } }, "source": [ - "

Formulating the RL problem

\n", - "

Refresher from the lecture

\n", - "\n", - "We need to define:\n", - "\n", - "- Actions\n", - "- Observations\n", - "- Reward\n", - "- Environment\n", - "- Agent\n", - " \n" + "\n" + ] + }, + { + "cell_type": "markdown", + "metadata": { + "slideshow": { + "slide_type": "slide" + } + }, + "source": [ + "\n" ] }, { @@ -284,7 +294,7 @@ "\n", "

Discussion

\n", "

$\\implies$ Is the action space continuous or discrete?

\n", - "

$\\implies$ Is the problem deterministic or stochastic?

\n" + "

$\\implies$ Is the problem fully observable or partially observable?

\n" ] }, { @@ -578,36 +588,6 @@ "You can find more information in the [paper](https://arxiv.org/abs/2401.05815) and the [code repository](https://github.com/desy-ml/cheetah).\n" ] }, - { - "cell_type": "code", - "execution_count": null, - "metadata": { - "slideshow": { - "slide_type": "slide" - } - }, - "outputs": [], - "source": [ - "# Importing the required packages\n", - "from time import sleep\n", - "\n", - "import matplotlib.pyplot as plt\n", - "import names\n", - "import numpy as np\n", - "from gymnasium.wrappers import RescaleAction\n", - "from IPython.display import clear_output, display\n", - "from stable_baselines3 import PPO\n", - "\n", - "from utils.helpers import (\n", - " evaluate_ares_ea_agent,\n", - " plot_ares_ea_training_history,\n", - " show_video,\n", - ")\n", - "from utils.train import ARESEACheetah, make_env, read_from_yaml\n", - "from utils.train import train as train_ares_ea\n", - "from utils.utils import NotVecNormalize" - ] - }, { "cell_type": "markdown", "metadata": {