diff --git a/main/.buildinfo b/main/.buildinfo index ca908cff..d7c0c66b 100644 --- a/main/.buildinfo +++ b/main/.buildinfo @@ -1,4 +1,4 @@ # Sphinx build info version 1 # This file hashes the configuration used when building these files. When it is not found, a full rebuild will be done. -config: 4e1b7ae49d50fe19c88b0a78195833c0 +config: e8169fef498febe3262dcf2f4cfaa3ba tags: d77d1c0d9ca2f4c8421862c7c5a0d620 diff --git a/main/release_notes/index.html b/main/release_notes/index.html index 8d85c270..07d4402b 100644 --- a/main/release_notes/index.html +++ b/main/release_notes/index.html @@ -356,7 +356,7 @@

MO-Gymnasium 1.0.0 Release Notes

reward = forward_reward - ctrl_cost

With MORL, users have the flexibility to determine the compromises they desire based on their preferences for each objective. Consequently, the environments in MO-Gymnasium do not have predefined weights. Thus, MO-Gymnasium extends the capabilities of Gymnasium to the multi-objective setting, where the agents receives a vectorial reward.

For example, here is an illustration of the multiple policies learned by an MORL agent for the mo-halfcheetah domain, balancing between saving battery and speed:

-

+

This release marks the first mature version of MO-Gymnasium within Farama, indicating that the API is stable, and we have achieved a high level of quality in this library.

API