diff --git a/.buildinfo b/.buildinfo index fb2539f..9d3705e 100644 --- a/.buildinfo +++ b/.buildinfo @@ -1,4 +1,4 @@ # Sphinx build info version 1 # This file hashes the configuration used when building these files. When it is not found, a full rebuild will be done. -config: 029408ec80ac705c56ad18ad01221a7f +config: 472fe710f589a6183296fabf4e8ea525 tags: d77d1c0d9ca2f4c8421862c7c5a0d620 diff --git a/release_notes/index.html b/release_notes/index.html index 7353cfc..0413ce9 100644 --- a/release_notes/index.html +++ b/release_notes/index.html @@ -462,7 +462,7 @@
With MORL, users have the flexibility to determine the compromises they desire based on their preferences for each objective. Consequently, the environments in MO-Gymnasium do not have predefined weights. Thus, MO-Gymnasium extends the capabilities of Gymnasium to the multi-objective setting, where the agents receives a vectorial reward.
For example, here is an illustration of the multiple policies learned by an MORL agent for the mo-halfcheetah
domain, balancing between saving battery and speed:
This release marks the first mature version of MO-Gymnasium within Farama, indicating that the API is stable, and we have achieved a high level of quality in this library.