Closed vassil-atn closed 2 months ago
What happens if you enable this flag?
Unfortunately it doesn't seem to have much of an effect: Where the gray and cyan are two consecutive runs with enable_enhanced_determinism=True and magenta and yellow are two consecutive runs with enable_enhanced_determinism=False
I realized one more change that might affect you:
Can you change this to the following and see if that improves things?
# set seed for torch and other libraries
return torch_utils.set_seed(seed, torch_deterministic=True)
Interestingly, calling torch.use_deterministic_algorithms(True)
seems to break it with the following error:
2024-03-11 10:02:35 [95,796ms] [Error] [__main__] linearIndex.numel()*sliceSize*nElemBefore == expandedValue.numel() INTERNAL ASSERT FAILED at "../aten/src/ATen/native/cuda/Indexing.cu":389, please report a bug to PyTorch. number of flattened indices did not match number of elements in the value tensor: 102 vs 2
2024-03-11 10:02:35 [95,797ms] [Error] [__main__] Traceback (most recent call last):
File "/home/user/orbit/source/standalone/workflows/rsl_rl/train.py", line 142, in <module>
main()
File "/home/user/orbit/source/standalone/workflows/rsl_rl/train.py", line 133, in main
runner.learn(num_learning_iterations=agent_cfg.max_iterations, init_at_random_ep_len=True)
File "/home/user/miniconda3/envs/orbit/lib/python3.10/site-packages/rsl_rl/runners/on_policy_runner.py", line 112, in learn
obs, rewards, dones, infos = self.env.step(actions)
File "/home/user/orbit/source/extensions/omni.isaac.orbit_tasks/omni/isaac/orbit_tasks/utils/wrappers/rsl_rl/vecenv_wrapper.py", line 161, in step
obs_dict, rew, terminated, truncated, extras = self.env.step(actions)
File "/home/user/miniconda3/envs/orbit/lib/python3.10/site-packages/gymnasium/wrappers/order_enforcing.py", line 56, in step
return self.env.step(action)
File "/home/user/orbit/source/extensions/omni.isaac.orbit/omni/isaac/orbit/envs/rl_task_env.py", line 192, in step
self._reset_idx(reset_env_ids)
File "/home/user/orbit/source/extensions/omni.isaac.orbit/omni/isaac/orbit/envs/rl_task_env.py", line 311, in _reset_idx
self.scene.reset(env_ids)
File "/home/user/orbit/source/extensions/omni.isaac.orbit/omni/isaac/orbit/scene/interactive_scene.py", line 222, in reset
articulation.reset(env_ids)
File "/home/user/orbit/source/extensions/omni.isaac.orbit/omni/isaac/orbit/assets/articulation/articulation.py", line 143, in reset
super().reset(env_ids)
File "/home/user/orbit/source/extensions/omni.isaac.orbit/omni/isaac/orbit/assets/rigid_object/rigid_object.py", line 114, in reset
self._external_force_b[env_ids] = 0.0
RuntimeError: linearIndex.numel()*sliceSize*nElemBefore == expandedValue.numel() INTERNAL ASSERT FAILED at "../aten/src/ATen/native/cuda/Indexing.cu":389, please report a bug to PyTorch. number of flattened indices did not match number of elements in the value tensor: 102 vs 2
Seems like it's an issue with the slicing that only pops up when running in deterministic mode.
For reference I am running with torch == 2.0.1+cu118
Hi,
Have there been any developments regarding this? Is the issue reproducible on your end @Mayankm96 ?
Seems like the issue of using torch.use_deterministic_algorithms(True)
has been fixed in this pytorch PR, which is included in pytorch v2.1.0
It seems difficult to upgrade the pytorch as the it's shipped with isaac sim 2023.1.1. This is my workaround to avoid the error without upgrading pytorch: change self._external_force_b[env_ids] = 0.0
to self._external_force_b[env_ids].zero_()
. However, the results are still non-deterministic :(
Maybe you wanna check: https://github.com/isaac-sim/IsaacLab/issues/904
Closing this issue as #904 raises the some concern. The fix is under review: https://github.com/isaac-sim/IsaacLab/pull/940
Describe the bug
The training is non-deterministic even with the same random seed.
Steps to reproduce
If you then rerun the same command and train for some episodes, the following two plots can be seen:
Despite nothing changing in between the two runs and using the same random seed, the resulting plots differ. Similarly, if you run the Flat terrain environment, the same can be observed:
Overall the same trends can be observed (which is good), but there is some stochasticity in each training run which makes hyperparameter and reward tuning problematic.
From my experience Isaac gym also had the same problem (https://github.com/NVIDIA-Omniverse/IsaacGymEnvs/issues/189). Interestingly, in Isaac Gym this was only an issue when training on trimesh terrains, but on plane terrains it was deterministic.
Based on the documentation, this can be somewhat expected (https://isaac-orbit.github.io/orbit/source/refs/issues.html#non-determinism-in-physics-simulation) when randomising physical properties at run-time. However, the rough_env_cfg does not do this to my knowledge - friction and mass are only randomised at startup. In any case, I tested it with both
physics_material
andbase_mass
commented out in theRandomizationCfg
and it was still non-deterministic.Is there something that I'm missing or is it an inherent issue of the GPU-based training?
System Info
Checklist
[x] I have checked that there is no similar issue in the repo (required)
[x] I have checked that the issue is not in running Isaac Sim itself and is related to the repo
Acceptance Criteria