You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For multi-obj, the input frame size is [batch_size, color channels, H, W], and the input objects mask size is [batch_size, num_objects + BG, H, W], and the questions are:
When the STM module inputs the data whose batch size is greater than 1, it failed, so our work is based on batch_size =1;
The network-output logit's size is [batch_size, num_objects + BG, H, W], then resize it to [batch_size*H*W, num_objects + BG], and input the new size tensor into the CrossEntropyLoss, is it right?
Same as above, I calcutated loss after got the mulit-objct logit at every frame, didn't use softmax, because CrossEntropyLoss does softmax internally, , and then sum up losses from a sample of frames, and backward.
The text was updated successfully, but these errors were encountered:
For multi-obj, the input frame size is
[batch_size, color channels, H, W]
, and the input objects mask size is[batch_size, num_objects + BG, H, W]
, and the questions are:batch_size =1
;[batch_size, num_objects + BG, H, W]
, then resize it to[batch_size*H*W, num_objects + BG]
, and input the new size tensor into the CrossEntropyLoss, is it right?The text was updated successfully, but these errors were encountered: