Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Difficulty Reproducing Results #4

Open
pat-hanbury opened this issue Jun 25, 2019 · 1 comment
Open

Difficulty Reproducing Results #4

pat-hanbury opened this issue Jun 25, 2019 · 1 comment

Comments

@pat-hanbury
Copy link

Hello,

I am trying to recreate the results on UCF-101 using the provided pretrained models for PlayingViolin and IceDancing. I am using the test_refine.py script for testing, but I am getting odd results:

Seg2Vid2

seg2vid-playingviolin

Seg2Vid3

The only thing in the code I changed to make the code run is:

z_m = Vb(z_noise.repeat(frame1.size()[0] * 8, 1))

I had to change this line from
z_m = Vb(z_noise.repeat(frame1.size()[0] * 8, 1))
to
z_m = Vb(z_noise.repeat(frame1.size()[0] * 2 * 8, 1))

Without this change, I received an error complaining about the incorrect size of z_m.

Traceback (most recent call last):
  File "test_refine.py", line 147, in <module>
    a.test()
  File "test_refine.py", line 125, in test
    y_pred_before_refine, y_pred, mu, logvar, flow, flowback, mask_fw, mask_bw = vae(frame1, data, noise_bg, z_m)
  File "/home/hanburyp/env/lib/python3.6/site-packages/torch/nn/modules/module.py", line 489, in __call__
    result = self.forward(*input, **kwargs)
  File "/home/hanburyp/seg2vid/src/models/multiframe_genmask.py", line 262, in forward
    torch.cat([self.fc(z_m).view(-1, 64, int(opt.input_size[0] / 16), int(opt.input_size[1] / 16)), codex], 1))
RuntimeError: shape '[-1, 64, 16, 16]' is invalid for input of size 8192

Please advise on more details about UCF-101 inferences in order to reproduce your result. Thank you in advance!

@junting
Copy link
Owner

junting commented Jul 5, 2019

Hi,
Thanks for your interest! I checked your GIF files, it seems that the size of the input image is 256x256, however, the provided pre-trained models were trained by using images of size 128x128. So could please run the inference again with images of a smaller size?

Also, we have also fixed the error you reported.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants