-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About the training curve #6
Comments
为啥a1.urdf 中的limit effort 会被修改啊? |
Please post the learning rate for a look. Visual inspection suggests that there are too many iterations. 1500 iterations is enough. |
Sorry for the late reply. We did not train for such a long time. However, the significant decrease in the mean reward is probably due to the curriculum of commands and terrains. I suggest training for at most 2000 iterations. We also updated the config for a1 and go1, high speed is a bit of a conflict with the ability to cross difficult terrains for small dogs, so we limited the highest command for small dogs to 2m/s. |
We use the a1.urdf from the original repo of legged_gym. There is also a file named a1.urdf.origin from unitree_ros, which seems hard to train. But the deployment is built upon the control loop of joint position and unitree SDK seems to handle this issue well. Therefore, don't worry about this. Position limits, velocity limits and power penalization are enough. |
In other words, the current environment is not suitable for small dogs, but in order not to change the environment, the model is changed. Am i right? |
@Junfeng-Long I ran the If I want to compare different methods in simulation, can I directly compare them with these curves? |
Hello, I have also encountered the same problem. Do you know where the problem lies? Thank you for your help |
Sorry for the late reply. There are some bugs and improper configs in the code. Already fixed, please try the new one. |
Thank you for your impressive work on this project.I used this project to train a policy and wanted to simulate it in Gazebo, but I found that the policy performed well in isaac. However, when I used Gazebo, the robot shook violently and could not stand properly. Have you done similar work before? Can you give me some suggestions? Thank you for your help |
We have done this test with Aliengo in gazebo, it works well but is still worse than Isaac. I would like to help if you can offer more information. For example, video, inference output, or the code. You can send me directly or post them here. |
|
We use pytorch since there are cuda on dog's ob-board computer. |
It seems that you accidentally open an issue under my homepage repo. Sorry for not noticing that. But happy to see that you figured out the problem:) |
|
I think this is due to improper target height configuration. Try a lower target height, for example, 0.25m for a1. |
Thank you for your reply. I noticed that your code has set the leg lifting height, but the weight setting is very low and the leg lifting height after training is not ideal, which is much different from the effect in the video you posted. I would like to ask if you have any other methods to improve this issue. |
I am using the code to train A1 on RTX 4090. However, I've noticed a significant decrease in the mean reward around the 2800th iteration. Is everything right? Should I continue training?
The text was updated successfully, but these errors were encountered: