How I enable to get trained policy of Isaac-Repose-Cube-Allegro-v0 task? #1442
Replies: 3 comments 2 replies
-
|
Thanks for posting this. By expert policy, do you mean the final trained policy? You may find them in this directory: |
Beta Was this translation helpful? Give feedback.
-
|
I think the first step is for us to reproduce what you're seeing. Can you give me reproduction steps for your two policies so I can see if I get the same results? Are you using the samples found here directly: https://isaac-sim.github.io/IsaacLab/main/source/overview/environments.html Are these also the shadow hand examples you're looking for? |
Beta Was this translation helpful? Give feedback.
-
|
Hi @cold-young Were you able to determine how to get the Allegro-v0 task to give a good policy? I have run into the same issue, where the policy does not perform well with the default training parameters. I increased the timesteps and the mean reward is not any better. As seen in the image below
|
Beta Was this translation helpful? Give feedback.



Uh oh!
There was an error while loading. Please reload this page.
-
Hi there,
I have been trying to train the Isaac-Repose-Cube-Allegro-v0 task using the skrl PPO algorithm.
(num_envs 2048, <1 hour)
Even though the reinforcement learning reward graph has converged, I cannot get an expert policy.
I think I need well defined shaped rewards or other learning approaches like imitation learning.
By the way, I found a good demonstration of in-hand manipulation with shadow hand. on this page
To acquire expert policy, what i do ?
Thanks
Beta Was this translation helpful? Give feedback.
All reactions