-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Testing Using Pretrained Model Results in Different Performance Across Multiple Runs #12
Comments
Hi @YunzeMan , We didn't meet nondeterministic issues on ScanNet. Can you please share |
This comment was marked as outdated.
This comment was marked as outdated.
Regarding randomness, I strictly followed your steps. I checked torch randomness guide but didn't find useful directions. Setting I also trained the model following your steps. Here is the log file. The [email protected] and [email protected] are both lower than your reported value. Have you altered the codebase or parameters a little bit without noticing it? |
Btw, i think i understand this little randomness in test stage. Here in SparseTensor construction the default |
Thanks for pointing that out. However, after changing Run 1: ([email protected]: 0.7068, [email protected]: 0.5702) What's more frustrating than the little randomness is the marginally lower performance. But since the gap isn't very large, I can perhaps work with the current version. |
Hi, Thanks for sharing your great work. As indicated in the title, I not only got lower performance ([email protected]: 0.7069, [email protected]: 0.5699), but also different performance across multiple runs.
These are several runs using the following command:
python tools/test.py configs/tr3d/tr3d_scannet-3d-18class.py ./tr3d_scannet.pth --eval mAP
(Yes, I'm using the provided pretrained model on scannet)Run 1: ([email protected]: 0.7069, [email protected]: 0.5699)
Run 2: ([email protected]: 0.7068, [email protected]: 0.5716)
Run 3: ([email protected]: 0.7069, [email protected]: 0.5710)
I'm using Pytorch 1.12, CUDA 11.3, CUDNN 8.
I cannot figure out where the stochasticity may come from, especially during the evaluation (test). Could you shed some lights on the possible reasons of this scenario?
Here is one of my outputs during the testing.
test_log.txt
The text was updated successfully, but these errors were encountered: