-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Request for Detailed Hyperparameters in LOTUS Experiments #6
Comments
Hi @BrightMoonStar, thanks for your interest in our work! For the hyperparameters you mentioned, you can find them in configs, we use these provided training scripts in our paper experiments. |
Thank you for your reply. I strictly followed the instructions of each step in your project and repeated it many times. I found that all evaluation indicators including success rate and Aoc are always 0. I also found that LIBERO has this problem , as shown below Lifelong-Robot-Learning/LIBERO#21 . I really can't find where the problem is. |
At first I thought n_epochs: 50 might not be enough in |
Dear Weikang, Thank you for open-sourcing this great repo! I'm wondering how did you tackle the challenge when opening hdf5 files with multiprocessing for LIBERO? It seems many people encountered this same issue: Lifelong-Robot-Learning/LIBERO#19 (comment). May we have your suggestions? @wkwan7 Thank you for your attention and precious time. Look forward to your reply!! Best regards, |
Hi @BrightMoonStar , can you try the default parameters (e.g., n_epochs: 50) and post the your output log shere? btw, I recommend to use wandb which can show more detailed logs. |
Hi @pengzhi1998, if the default dataloader setting not works for you, you can try this:
I don't think this will significantly increase the training time. |
Thank you Weikang for your reply! Yes I have tried it and worked well when training and evaluating on the first task in LIBERO. However, when training the second task, the same problem occurred but due to a different reason:
It seems when running experience reply algorithm, this Besides, I noticed Robomimic also made use of multiprocessing for training with data as hdf5 files. While their implementations are very similar to LIBERO's, but didn't encounter this error, which is also confusing. May I have some of your insights about these issues? Thank you so much again!! |
Hi @pengzhi1998, when I run ER using the Lotus codebase, it doesn't seem to have the issue you mentioned. The command I used is as follows:
Maybe you can try using Lotus codebase to see if you still have the issue. |
Hi , this is the wandb log with n_epoch=50 |
@BrightMoonStar Hi, did you solve this problem (success rate is always around 0) at last? |
Dear Dr. Weikang Wan and Team,
I recently came across your fascinating work on the LOTUS algorithm, as detailed in your paper "LOTUS: Continual Imitation Learning for Robot Manipulation Through Unsupervised Skill Discovery."
Your approach to lifelong robot learning through unsupervised skill discovery is truly impressive and offers significant insights into continual imitation learning for robot manipulation. I am particularly interested in replicating and building upon your experiments as part of my research.
However, I noticed that the paper does not provide specific details on some of the experimental hyperparameters, such as the learning rate, number of epochs, and batch size used during training. These details are crucial for ensuring that my replication is as accurate as possible.
Could you kindly provide the following details:
The learning rate(s) used for training the models.
The number of epochs each model was trained for.
The batch size used during training.
Any other relevant hyperparameters or settings that were critical to the performance of the LOTUS algorithm.
I greatly appreciate your time and assistance. Your work is a significant contribution to the field, and having these details would be immensely helpful for my research.
Thank you very much for your support and I look forward to your response.
Best regards
The text was updated successfully, but these errors were encountered: