You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
hello, when i run the code,i find that i can't get the same loss or f1 score with fixed seed.
for example i set seed==0,and the f1 result will get 43.25, 43.20, 42.89 etc.
i print the output and find that first finetune round the model get the fixed bert out and loss,then loss.backward(),the second round the bert out or loss will change slightly.
it's so wired that every time rerun the scripts, the result always change even fix the seed.
whether the loss function is too complex and The model accuracy is not enough
The text was updated successfully, but these errors were encountered:
Regarding model performance: Can you be more specific about which dataset, which script, and which support/test set you are using? I think I got a similar performance in repeated runs.
Hi, I get a similar issue on few_nerd inter dataset provided by here when I tried to train from scratch. I run:
sh exec_container.sh inter 0 5 5, I finally get different F1-score of 55.17 and 45.28 across different runs (seed=1, gpu=0 for both). Is this variance a normal case? (I set batch size=16 because of limited GPU memory).
And for simplicity, I just evaluate on 'support_test_5_5/0', and test on 'query_test_5_5/0'
In exec_container.sh, I adjust my code line21 to:
hello, when i run the code,i find that i can't get the same loss or f1 score with fixed seed.
for example i set seed==0,and the f1 result will get 43.25, 43.20, 42.89 etc.
i print the output and find that first finetune round the model get the fixed bert out and loss,then loss.backward(),the second round the bert out or loss will change slightly.
it's so wired that every time rerun the scripts, the result always change even fix the seed.
whether the loss function is too complex and The model accuracy is not enough
The text was updated successfully, but these errors were encountered: