Cuncurrent training of supervised network during policy training #3208
Unanswered
giulioturrisi
asked this question in
Q&A
Replies: 1 comment
-
Solution found: with torch.inference_mode(False): By adding the above lines, it trains. I will check if this mess up with the rl training for some reason. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, I'm trying to train a supervised state estimator together with the RL policy, inside a direct env (see https://arxiv.org/pdf/2202.05481). The only problem is that, when I train the supervised network during rl policy training (the training method is called inside _get_observations), I obtain:
RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
So it seems that the main functions are wrapped inside a no_grad routine of PyTorch. Note that outside IsaacLab it trains normally, so the problem is the embedding inside the exposed function of IsaacLab.
Has anyone found this problem before? Is there a solution? (other than completely separated training routine)
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions