You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I ran the fcmae pretrainer on a custom dataset, but after using the checkpoints generated by the pretraining script in the fine tuning script, the network produces NaN as output for every image.
Using no weights or the provided imagenet self supervised pretrained weights does not produce this issue.
I inspected the weights I got from the pretraining and the values don't seem off. There are no NaNs or zeroes in there, nor are there any large outliers.
Has anybody had this issue as well, or know of something that might cause/solve this? Thanks in advance
The text was updated successfully, but these errors were encountered:
Hi @garg-aayush , in my case the problem turned out to be a double initialisation of the encoder weights. They were initialized both in the ConvNeXtV2 class, as wel as in the MAE class. I'd check if that is what happened for you too, otherwise their could be a range of other causes
@RienkF, in my case it was an autocast issue. Currently, I am running it in fp32 and it works fine. I might need to update some packages later to run it in fp16/bf16.
Hi all,
I ran the fcmae pretrainer on a custom dataset, but after using the checkpoints generated by the pretraining script in the fine tuning script, the network produces NaN as output for every image.
Using no weights or the provided imagenet self supervised pretrained weights does not produce this issue.
I inspected the weights I got from the pretraining and the values don't seem off. There are no NaNs or zeroes in there, nor are there any large outliers.
Has anybody had this issue as well, or know of something that might cause/solve this? Thanks in advance
The text was updated successfully, but these errors were encountered: