-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to reproduce the values in the paper #8
Comments
Hi, Unfortunately, the training under the pytorch framework is non-deterministic. A relevant issue is here: https://discuss.pytorch.org/t/random-seed-initialization/7854/18 Even though we re-run the code, there still exists some fluctuation, but it is not difficult to get a value higher than 44% (with batch size = 2). To your guess,
The potential reason could be the version of libraries. I am re-running my code with batch size = 1. I hope I can come up with some good results and random seed to help you reproduce the performance. --- edit --- Cheers, |
Thank you for all the suggestions. I will try them out and see if I can improve the values.Please let me know if you are able to get to the value of 44% with batch size = 1. |
hello,i run the code,but my device is two 1080GPU,i want to know that what are your device? and how long you cost to train the model? @subeeshvasu |
@crazygirl1992 I was using a single GTX TitanX (12 GB). With this settings, for batch size = 1, training cost was approximately: 2hrs, 20 minutes per 1000 iterations. |
thank you very much,and can you achive the paper's result now? the training almost 250000 iterations in his paper,and 20*250mins,not 2hrs |
I couldn't get those values. With batch size = 4, one could reproduce the values I guess!. |
Hello,
I run the code "train_dise_gta2city.py" following the procedure explained in this project page. The only change I have done was to keep the batch_size as 1, to reduce the memory requirement. I got 38.4% mIoU on val set. This is a big difference as compared to the value of 45.4% reported in the paper. Can you please help me to understand the potential reasons behind this performance drop.
Some of the possible reasons which I could guess are the following.
The mIoU scores are computed at a resolution of 512 x 1024, while the original images are of size 1024 x 2048. In the paper, the resolution used to report the values are not mentioned. May be authors have reported the values at a resolution of 1024 x 2048? Just to check if this is the reason, I used the pretrained weights provided by the authors and got a score of 44.2% for images at resolution 512 x 1024. Therefore, I am assuming that the resolution of test images is not the reason behind performance drop
As per the paper, authors have used pretrained weights from PASCAL VOC dataset to initialise the encoder. This can also be the reason behind performance drop. However, even when I start the training scheme with the pretrained weights provided by the authors, the performance goes down eventually and will start to fluctuate around 38-39%.
Has anyone succeeded to get values around 44 % up on experimentation with this code?
Regards,
Subeesh
The text was updated successfully, but these errors were encountered: