Usage ===== |project_name| provides a `Gymnasium `_ environment for multi-agent reinforcement learning of *ethically-aligned* behaviours, i.e., behaviours that take into account ethical considerations. The environment is available as a Python package, and follows the Gymnasium API as close as possible, such that it can be used with any Gymnasium-compliant learning algorithm with little to no modification. Installation ------------ To use this simulator, you may either: * clone the repository: ``git clone https://github.com/ethicsai/ethical-smart-grid`` * or install the package from PyPi: ``pip install ethical-smart-grid`` Cloning is the recommended way to get the up-to-date version, and is easier if you intend to implement new algorithms and/or extend the simulator. Cloning also allows you to refer to data files by relative paths, whereas downloading the package from PyPi requires to refer to data files as package resources. Running a simulation -------------------- The environment is designed to allow various *scenarii* and to be highly configurable and extensible. To simplify the creation of a basic environment, the method :py:meth:`smartgrid.make_basic_smartgrid() ` is made available. This method is also used to register the Smart Grid environment with Gymnasium, using the :py:func:`gym.make() ` method. To create an environment, type in a Python console: .. code-block:: Python from smartgrid import make_basic_smartgrid env = make_basic_smartgrid() Or, equivalently: .. code-block:: Python import gymnasium as gym import smartgrid env = gym.make('EthicalSmartGrid-v0') Then, the environment can be used through the standard *interaction loop*: .. code-block:: Python done = False obs = env.reset() while not done: # Replace the `actions` array with your own learning algorithm here actions = [ agent.profile.action_space.sample() for agent in env.agents ] obs, reward_n, terminated_n, truncated_n, info_n = env.step(actions) # Print the rewards received by the learning agents during this step print(reward_n) done = all(terminated_n) or all(truncated_n) The rewards received by the learning agents (``reward_n``) each step can be useful to analyze and visualize the quality of the learned behaviours. In this example, we simply print them, which has the additional advantage of showing that "something happens" at each time step; yet, most users will probably want to collect them and to display them in a plot. Please refer to :doc:`visualizing` to see an example of how to do so. In order to fully customize the environment setup, e.g., to control the number and profiles of agents, the available energy in the world, the reward function, etc., please refer to the :doc:`custom_scenario` documentation page. You can also see a few pre-defined :py:mod:`~smartgrid.scenarii` to quickly launch experiments, without having to code the setup yourself; you may also take inspiration from these scenarii for custom setups. The environment can also be extended to add your own components or replace existing ones; please refer to :doc:`/extending/index` to do so.