Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
28 commits
Select commit Hold shift + click to select a range
33e6865
Fixes deprecation warning for pxr.Semantics (#2721)
kellyguo11 Jun 17, 2025
91ad494
Fixes visual prims handling during texture randomization. (#2476)
KumoLiu Jun 17, 2025
cf61e98
Fixes link in `training_jetbot_gt.rst` (#2699)
fan-ziqi Jun 23, 2025
68d96a5
Fixes walkthrough path to in text for jetbot env (#2770)
mpgussert Jun 24, 2025
9980e66
Adds print info in disassembly direct environment (#2750)
yijieg Jun 24, 2025
ad14a67
Adds optimizations and additional training configs for SB3 (#2022)
araffin Jun 25, 2025
ea717fa
Integrates `NoiseModel` to manager-based workflows (#2755)
ozhanozen Jun 25, 2025
c75bc5c
Fixes inconsistent data reading in body, link, com for RigidObject, R…
ooctipus Jun 25, 2025
0b377f6
Updates documentation for conda, fabric, and inferencing (#2772)
kellyguo11 Jun 25, 2025
33b4973
Resets joint position/velocity targets in reset_scene_to_default() (#…
wghou Jun 25, 2025
887342a
Updates locomotion configs to fix body_com error (#2655)
kellyguo11 Jun 25, 2025
9655e62
Adds env_cfg and agent_cfg to wandb in rl_games (#2771)
ooctipus Jun 25, 2025
4f25c9d
Adds digit locomotion examples (#1892)
lgulich Jun 26, 2025
34cda4f
Fixes typo in reset_scene_to_default (#2778)
kellyguo11 Jun 26, 2025
1ad83e1
Updates `NoiseModelWithAdditiveBias` to apply per-feature bias sampli…
ozhanozen Jun 26, 2025
25f7a5d
Removes redundant contact termination assignment in `H1RoughEnvCfg` (…
louislelay Jun 26, 2025
d7fac05
Improves the implementation of euler_xyz_from_quat (#2365)
ShaoshuSu Jun 27, 2025
3d82782
Fixes typo in the docs for adding your own library (#2520)
norbertcygiert Jun 27, 2025
fa8612f
Fixes memory leak in SDF calculation - warp doesn't free memory (#2680)
leondavi Jun 27, 2025
394a162
Modifies `update_class_from_dict()` to wholesale replace flat Iterabl…
ozhanozen Jun 27, 2025
05c22be
Removes protobuf upper version pin (#2726)
kwlzn Jun 27, 2025
66cc743
Updates play script for SB3 and RL library benchmarks (#2789)
kellyguo11 Jun 27, 2025
4a7d15d
Updates training_jetbot_reward_exploration.rst (#2788)
bsb808 Jun 27, 2025
3d6f55b
Fixes the implementation of `quat_inv()` (#2797)
ozhanozen Jun 28, 2025
752be19
Sets robomimic version to v0.4.0 (#2814)
masoudmoghani Jul 1, 2025
367925c
Allows slicing to be processed from lists (#2853)
kellyguo11 Jul 7, 2025
8d589fc
Fixes numpy ver to <2 for isaaclab_rl and isaaclab_tasks (#2866)
ooctipus Jul 7, 2025
75824e8
Updates gymnasium to v1.2.0 (#2852)
kellyguo11 Jul 8, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 6 additions & 0 deletions CONTRIBUTORS.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,11 +40,13 @@ Guidelines for modifications:
* Amr Mousa
* Andrej Orsula
* Anton Bjørndahl Mortensen
* Antonin Raffin
* Arjun Bhardwaj
* Ashwin Varghese Kuruttukulam
* Bikram Pandit
* Bingjie Tang
* Brayden Zhang
* Brian Bingham
* Cameron Upright
* Calvin Yu
* Cheng-Rong Lai
Expand Down Expand Up @@ -74,6 +76,7 @@ Guidelines for modifications:
* Jinqi Wei
* Johnson Sun
* Kaixi Bao
* Kris Wilson
* Kourosh Darvish
* Kousheek Chakraborty
* Lionel Gulich
Expand All @@ -86,6 +89,7 @@ Guidelines for modifications:
* Miguel Alonso Jr
* Muhong Guo
* Nicola Loi
* Norbert Cygiert
* Nuoyan Chen (Alvin)
* Nuralem Abizov
* Ori Gadot
Expand All @@ -102,6 +106,7 @@ Guidelines for modifications:
* Rosario Scalise
* Ryley McCarroll
* Shafeef Omar
* Shaoshu Su
* Shundo Kishi
* Stefan Van de Mosselaer
* Stephan Pleines
Expand All @@ -119,6 +124,7 @@ Guidelines for modifications:
* Zhengyu Zhang
* Ziqi Fan
* Zoe McCarthy
* David Leon

## Acknowledgements

Expand Down
11 changes: 6 additions & 5 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -93,12 +93,13 @@ or opening a question on its [forums](https://forums.developer.nvidia.com/c/agx-

## Connect with the NVIDIA Omniverse Community

Have a project or resource you'd like to share more widely? We'd love to hear from you! Reach out to the
NVIDIA Omniverse Community team at OmniverseCommunity@nvidia.com to discuss potential opportunities
for broader dissemination of your work.
Do you have a project or resource you'd like to share more widely? We'd love to hear from you!
Reach out to the NVIDIA Omniverse Community team at OmniverseCommunity@nvidia.com to explore opportunities
to spotlight your work.

Join us in building a vibrant, collaborative ecosystem where creativity and technology intersect. Your
contributions can make a significant impact on the Isaac Lab community and beyond!
You can also join the conversation on the [Omniverse Discord](https://discord.com/invite/nvidiaomniverse) to
connect with other developers, share your projects, and help grow a vibrant, collaborative ecosystem
where creativity and technology intersect. Your contributions can make a meaningful impact on the Isaac Lab community and beyond!

## License

Expand Down
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
2 changes: 1 addition & 1 deletion docs/source/how-to/add_own_library.rst
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ Integrating a new library

Adding a new library to Isaac Lab is similar to using a different version of a library. You can install the library
in your Python environment and use it in your experiments. However, if you want to integrate the library with
Isaac Lab, you can will first need to make a wrapper for the library, as explained in
Isaac Lab, you will first need to make a wrapper for the library, as explained in
:ref:`how-to-env-wrappers`.

The following steps can be followed to integrate a new library with Isaac Lab:
Expand Down
1 change: 1 addition & 0 deletions docs/source/overview/core-concepts/actuators.rst
Original file line number Diff line number Diff line change
Expand Up @@ -35,6 +35,7 @@ maximum effort:
.. math::

\tau_{j, computed} & = k_p * (q_{des} - q) + k_d * (\dot{q}_{des} - \dot{q}) + \tau_{ff} \\
\tau_{j, max} & = \gamma \times \tau_{motor, max} \\
\tau_{j, applied} & = clip(\tau_{computed}, -\tau_{j, max}, \tau_{j, max})


Expand Down
38 changes: 36 additions & 2 deletions docs/source/overview/environments.rst
Original file line number Diff line number Diff line change
Expand Up @@ -214,6 +214,12 @@ We provide environments for both disassembly and assembly.
wget https://developer.download.nvidia.com/compute/cuda/12.8.0/local_installers/cuda_12.8.0_570.86.10_linux.run
sudo sh cuda_12.8.0_570.86.10_linux.run

When using conda, cuda toolkit can be installed with:

.. code-block:: bash

conda install cudatoolkit

For addition instructions and Windows installation, please refer to the `CUDA installation page <https://developer.nvidia.com/cuda-12-8-0-download-archive>`_.

* |disassembly-link|: The plug starts inserted in the socket. A low-level controller lifts the plug out and moves it to a random position. This process is purely scripted and does not involve any learned policy. Therefore, it does not require policy training or evaluation. The resulting trajectories serve as demonstrations for the reverse process, i.e., learning to assemble. To run disassembly for a specific task: ``python source/isaaclab_tasks/isaaclab_tasks/direct/automate/run_disassembly_w_id.py --assembly_id=ASSEMBLY_ID --disassembly_dir=DISASSEMBLY_DIR``. All generated trajectories are saved to a local directory ``DISASSEMBLY_DIR``.
Expand Down Expand Up @@ -288,6 +294,12 @@ Environments based on legged locomotion tasks.
+------------------------------+----------------------------------------------+------------------------------------------------------------------------------+
| |velocity-rough-g1| | |velocity-rough-g1-link| | Track a velocity command on rough terrain with the Unitree G1 robot |
+------------------------------+----------------------------------------------+------------------------------------------------------------------------------+
| |velocity-flat-digit| | |velocity-flat-digit-link| | Track a velocity command on flat terrain with the Agility Digit robot |
+------------------------------+----------------------------------------------+------------------------------------------------------------------------------+
| |velocity-rough-digit| | |velocity-rough-digit-link| | Track a velocity command on rough terrain with the Agility Digit robot |
+------------------------------+----------------------------------------------+------------------------------------------------------------------------------+
| |tracking-loco-manip-digit| | |tracking-loco-manip-digit-link| | Track a root velocity and hand pose command with the Agility Digit robot |
+------------------------------+----------------------------------------------+------------------------------------------------------------------------------+

.. |velocity-flat-anymal-b-link| replace:: `Isaac-Velocity-Flat-Anymal-B-v0 <https://github.com/isaac-sim/IsaacLab/blob/main/source/isaaclab_tasks/isaaclab_tasks/manager_based/locomotion/velocity/config/anymal_b/flat_env_cfg.py>`__
.. |velocity-rough-anymal-b-link| replace:: `Isaac-Velocity-Rough-Anymal-B-v0 <https://github.com/isaac-sim/IsaacLab/blob/main/source/isaaclab_tasks/isaaclab_tasks/manager_based/locomotion/velocity/config/anymal_b/rough_env_cfg.py>`__
Expand Down Expand Up @@ -318,6 +330,9 @@ Environments based on legged locomotion tasks.
.. |velocity-flat-g1-link| replace:: `Isaac-Velocity-Flat-G1-v0 <https://github.com/isaac-sim/IsaacLab/blob/main/source/isaaclab_tasks/isaaclab_tasks/manager_based/locomotion/velocity/config/g1/flat_env_cfg.py>`__
.. |velocity-rough-g1-link| replace:: `Isaac-Velocity-Rough-G1-v0 <https://github.com/isaac-sim/IsaacLab/blob/main/source/isaaclab_tasks/isaaclab_tasks/manager_based/locomotion/velocity/config/g1/rough_env_cfg.py>`__

.. |velocity-flat-digit-link| replace:: `Isaac-Velocity-Flat-Digit-v0 <https://github.com/isaac-sim/IsaacLab/blob/main/source/isaaclab_tasks/isaaclab_tasks/manager_based/locomotion/velocity/config/digit/flat_env_cfg.py>`__
.. |velocity-rough-digit-link| replace:: `Isaac-Velocity-Rough-Digit-v0 <https://github.com/isaac-sim/IsaacLab/blob/main/source/isaaclab_tasks/isaaclab_tasks/manager_based/locomotion/velocity/config/digit/rough_env_cfg.py>`__
.. |tracking-loco-manip-digit-link| replace:: `Isaac-Tracking-Flat-Digit-v0 <https://github.com/isaac-sim/IsaacLab/blob/main/source/isaaclab_tasks/isaaclab_tasks/manager_based/locomotion/tracking/config/digit/loco_manip_env_cfg.py>`__

.. |velocity-flat-anymal-b| image:: ../_static/tasks/locomotion/anymal_b_flat.jpg
.. |velocity-rough-anymal-b| image:: ../_static/tasks/locomotion/anymal_b_rough.jpg
Expand All @@ -336,6 +351,9 @@ Environments based on legged locomotion tasks.
.. |velocity-rough-h1| image:: ../_static/tasks/locomotion/h1_rough.jpg
.. |velocity-flat-g1| image:: ../_static/tasks/locomotion/g1_flat.jpg
.. |velocity-rough-g1| image:: ../_static/tasks/locomotion/g1_rough.jpg
.. |velocity-flat-digit| image:: ../_static/tasks/locomotion/agility_digit_flat.jpg
.. |velocity-rough-digit| image:: ../_static/tasks/locomotion/agility_digit_rough.jpg
.. |tracking-loco-manip-digit| image:: ../_static/tasks/locomotion/agility_digit_loco_manip.jpg

Navigation
~~~~~~~~~~
Expand Down Expand Up @@ -554,6 +572,10 @@ Environments based on fixed-arm manipulation tasks.
Comprehensive List of Environments
==================================

For environments that have a different task name listed under ``Inference Task Name``, please use the Inference Task Name
provided when running ``play.py`` or any inferencing workflows. These tasks provide more suitable configurations for
inferencing, including reading from an already trained checkpoint and disabling runtime perturbations used for training.

.. list-table::
:widths: 33 25 19 25

Expand Down Expand Up @@ -761,6 +783,10 @@ Comprehensive List of Environments
-
- Manager Based
-
* - Isaac-Tracking-LocoManip-Digit-v0
- Isaac-Tracking-LocoManip-Digit-Play-v0
- Manager Based
- **rsl_rl** (PPO)
* - Isaac-Navigation-Flat-Anymal-C-v0
- Isaac-Navigation-Flat-Anymal-C-Play-v0
- Manager Based
Expand Down Expand Up @@ -869,6 +895,10 @@ Comprehensive List of Environments
- Isaac-Velocity-Flat-Cassie-Play-v0
- Manager Based
- **rsl_rl** (PPO), **skrl** (PPO)
* - Isaac-Velocity-Flat-Digit-v0
- Isaac-Velocity-Flat-Digit-Play-v0
- Manager Based
- **rsl_rl** (PPO)
* - Isaac-Velocity-Flat-G1-v0
- Isaac-Velocity-Flat-G1-Play-v0
- Manager Based
Expand All @@ -884,7 +914,7 @@ Comprehensive List of Environments
* - Isaac-Velocity-Flat-Unitree-A1-v0
- Isaac-Velocity-Flat-Unitree-A1-Play-v0
- Manager Based
- **rsl_rl** (PPO), **skrl** (PPO)
- **rsl_rl** (PPO), **skrl** (PPO), **sb3** (PPO)
* - Isaac-Velocity-Flat-Unitree-Go1-v0
- Isaac-Velocity-Flat-Unitree-Go1-Play-v0
- Manager Based
Expand Down Expand Up @@ -913,6 +943,10 @@ Comprehensive List of Environments
- Isaac-Velocity-Rough-Cassie-Play-v0
- Manager Based
- **rsl_rl** (PPO), **skrl** (PPO)
* - Isaac-Velocity-Rough-Digit-v0
- Isaac-Velocity-Rough-Digit-Play-v0
- Manager Based
- **rsl_rl** (PPO)
* - Isaac-Velocity-Rough-G1-v0
- Isaac-Velocity-Rough-G1-Play-v0
- Manager Based
Expand All @@ -924,7 +958,7 @@ Comprehensive List of Environments
* - Isaac-Velocity-Rough-Unitree-A1-v0
- Isaac-Velocity-Rough-Unitree-A1-Play-v0
- Manager Based
- **rsl_rl** (PPO), **skrl** (PPO)
- **rsl_rl** (PPO), **skrl** (PPO), **sb3** (PPO)
* - Isaac-Velocity-Rough-Unitree-Go1-v0
- Isaac-Velocity-Rough-Unitree-Go1-Play-v0
- Manager Based
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -187,7 +187,7 @@ Stable-Baselines3

- Training an agent with
`Stable-Baselines3 <https://stable-baselines3.readthedocs.io/en/master/index.html>`__
on ``Isaac-Cartpole-v0``:
on ``Isaac-Velocity-Flat-Unitree-A1-v0``:

.. tab-set::
:sync-group: os
Expand All @@ -200,14 +200,13 @@ Stable-Baselines3
# install python module (for stable-baselines3)
./isaaclab.sh -i sb3
# run script for training
# note: we set the device to cpu since SB3 doesn't optimize for GPU anyway
./isaaclab.sh -p scripts/reinforcement_learning/sb3/train.py --task Isaac-Cartpole-v0 --headless --device cpu
./isaaclab.sh -p scripts/reinforcement_learning/sb3/train.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --headless
# run script for playing with 32 environments
./isaaclab.sh -p scripts/reinforcement_learning/sb3/play.py --task Isaac-Cartpole-v0 --num_envs 32 --checkpoint /PATH/TO/model.zip
./isaaclab.sh -p scripts/reinforcement_learning/sb3/play.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --num_envs 32 --checkpoint /PATH/TO/model.zip
# run script for playing a pre-trained checkpoint with 32 environments
./isaaclab.sh -p scripts/reinforcement_learning/sb3/play.py --task Isaac-Cartpole-v0 --num_envs 32 --use_pretrained_checkpoint
./isaaclab.sh -p scripts/reinforcement_learning/sb3/play.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --num_envs 32 --use_pretrained_checkpoint
# run script for recording video of a trained agent (requires installing `ffmpeg`)
./isaaclab.sh -p scripts/reinforcement_learning/sb3/play.py --task Isaac-Cartpole-v0 --headless --video --video_length 200
./isaaclab.sh -p scripts/reinforcement_learning/sb3/play.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --headless --video --video_length 200

.. tab-item:: :icon:`fa-brands fa-windows` Windows
:sync: windows
Expand All @@ -217,14 +216,13 @@ Stable-Baselines3
:: install python module (for stable-baselines3)
isaaclab.bat -i sb3
:: run script for training
:: note: we set the device to cpu since SB3 doesn't optimize for GPU anyway
isaaclab.bat -p scripts\reinforcement_learning\sb3\train.py --task Isaac-Cartpole-v0 --headless --device cpu
isaaclab.bat -p scripts\reinforcement_learning\sb3\train.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --headless
:: run script for playing with 32 environments
isaaclab.bat -p scripts\reinforcement_learning\sb3\play.py --task Isaac-Cartpole-v0 --num_envs 32 --checkpoint /PATH/TO/model.zip
isaaclab.bat -p scripts\reinforcement_learning\sb3\play.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --num_envs 32 --checkpoint /PATH/TO/model.zip
:: run script for playing a pre-trained checkpoint with 32 environments
isaaclab.bat -p scripts\reinforcement_learning\sb3\play.py --task Isaac-Cartpole-v0 --num_envs 32 --use_pretrained_checkpoint
isaaclab.bat -p scripts\reinforcement_learning\sb3\play.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --num_envs 32 --use_pretrained_checkpoint
:: run script for recording video of a trained agent (requires installing `ffmpeg`)
isaaclab.bat -p scripts\reinforcement_learning\sb3\play.py --task Isaac-Cartpole-v0 --headless --video --video_length 200
isaaclab.bat -p scripts\reinforcement_learning\sb3\play.py --task Isaac-Velocity-Flat-Unitree-A1-v0 --headless --video --video_length 200

All the scripts above log the training progress to `Tensorboard`_ in the ``logs`` directory in the root of
the repository. The logs directory follows the pattern ``logs/<library>/<task>/<date-time>``, where ``<library>``
Expand Down
11 changes: 6 additions & 5 deletions docs/source/overview/reinforcement-learning/rl_frameworks.rst
Original file line number Diff line number Diff line change
Expand Up @@ -71,17 +71,18 @@ Training Performance
--------------------

We performed training with each RL library on the same ``Isaac-Humanoid-v0`` environment
with ``--headless`` on a single RTX 4090 GPU
with ``--headless`` on a single RTX PRO 6000 GPU using 4096 environments
and logged the total training time for 65.5M steps for each RL library.


+--------------------+-----------------+
| RL Library | Time in seconds |
+====================+=================+
| RL-Games | 203 |
| RL-Games | 207 |
+--------------------+-----------------+
| SKRL | 204 |
| SKRL | 208 |
+--------------------+-----------------+
| RSL RL | 207 |
| RSL RL | 199 |
+--------------------+-----------------+
| Stable-Baselines3 | 6320 |
| Stable-Baselines3 | 322 |
+--------------------+-----------------+
4 changes: 4 additions & 0 deletions docs/source/setup/installation/binaries_installation.rst
Original file line number Diff line number Diff line change
Expand Up @@ -451,6 +451,10 @@ On Windows machines, please terminate the process from Command Prompt using

If you see this, then the installation was successful! |:tada:|

If you see an error ``ModuleNotFoundError: No module named 'isaacsim'``, ensure that the conda environment is activated
and ``source _isaac_sim/setup_conda_env.sh`` has been executed.


Train a robot!
~~~~~~~~~~~~~~~

Expand Down
2 changes: 1 addition & 1 deletion docs/source/setup/walkthrough/api_env_design.rst
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ the prim path with ``/World/envs/env_.*/Robot`` we are implicitly saying that ev
The Environment
-----------------

Next, let's take a look at the contents of the other python file in our task directory: ``isaac_lab_tutorial_env_cfg.py``
Next, let's take a look at the contents of the other python file in our task directory: ``isaac_lab_tutorial_env.py``

.. code-block:: python

Expand Down
2 changes: 1 addition & 1 deletion docs/source/setup/walkthrough/training_jetbot_gt.rst
Original file line number Diff line number Diff line change
Expand Up @@ -104,7 +104,7 @@ Next, we need to expand the initialization and setup steps to construct the data

Most of this is setting up the book keeping for the commands and markers, but the command initialization and the yaw calculations are worth diving into. The commands
are sampled from a multivariate normal distribution via ``torch.randn`` with the z component fixed to zero and then normalized to unit length. In order to point our
command markers along these vectors, we need to rotate the base arrow mesh appropriately. This means we need to define a `quaternion <https://en.wikipedia.org/wiki/Quaternion>`_` that will rotate the arrow
command markers along these vectors, we need to rotate the base arrow mesh appropriately. This means we need to define a `quaternion <https://en.wikipedia.org/wiki/Quaternion>`_ that will rotate the arrow
prim about the z axis by some angle defined by the command. By convention, rotations about the z axis are called a "yaw" rotation (akin to roll and pitch).

Luckily for us, Isaac Lab provides a utility to generate a quaternion from an axis of rotation and an angle: :func:`isaaclab.utils.math.quat_from_axis_angle`, so the only
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -75,6 +75,8 @@ from linear algebra! Replace the contents of ``_get_observations`` with the foll
observations = {"policy": obs}
return observations

We also need to **edit the ``IsaacLabTutorialEnvCfg`` to set the observation space back to 3** which includes the dot product, the z component of the cross product, and the forward speed.

The dot or inner product tells us how aligned two vectors are as a single scalar quantity. If they are very aligned and pointed in the same direction, then the inner
product will be large and positive, but if they are aligned and in opposite directions, it will be large and negative. If two vectors are
perpendicular, the inner product is zero. This means that the inner product between the forward vector and the command vector can tell us
Expand Down
2 changes: 1 addition & 1 deletion scripts/reinforcement_learning/rl_games/play.py
Original file line number Diff line number Diff line change
Expand Up @@ -188,7 +188,7 @@ def main():
s[:, dones, :] = 0.0
if args_cli.video:
timestep += 1
# Exit the play loop after recording one video
# exit the play loop after recording one video
if timestep == args_cli.video_length:
break

Expand Down
3 changes: 2 additions & 1 deletion scripts/reinforcement_learning/rl_games/train.py
Original file line number Diff line number Diff line change
Expand Up @@ -194,10 +194,11 @@ def main(env_cfg: ManagerBasedRLEnvCfg | DirectRLEnvCfg | DirectMARLEnvCfg, agen
entity=args_cli.wandb_entity,
name=experiment_name,
sync_tensorboard=True,
config=agent_cfg,
monitor_gym=True,
save_code=True,
)
wandb.config.update({"env_cfg": env_cfg.to_dict()})
wandb.config.update({"agent_cfg": agent_cfg})

if args_cli.checkpoint is not None:
runner.run({"train": True, "play": False, "sigma": train_sigma, "checkpoint": resume_path})
Expand Down
Loading