-
Notifications
You must be signed in to change notification settings - Fork 230
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ConstantPad1d Deprecated #44
Comments
I've had the same problem. |
Here is another example of a custom autograd. I'm currently rewriting this function myself. https://pytorch.org/tutorials/beginner/examples_autograd/two_layer_net_custom_function.html |
Or maybe everyone encountering this problem should be using this branch instead https://github.com/Vichoko/pytorch-wavenet |
were you able to solve? I tried the other branch but I get the same error |
the same problem that we all meet has been solved by following steps:
by the way, i run the wavenet_demo successfully. |
see the lasted comment |
can refer my comment |
Hey,
Can you please help?
I am facing Runtime error for this function.
``
class ConstantPad1d(Function):
def init(self, target_size, dimension=0, value=0, pad_start=False):
super(ConstantPad1d, self).init()
self.target_size = target_size
self.dimension = dimension
self.value = value
self.pad_start = pad_start
``
RuntimeError: Legacy autograd function with non-static forward method is deprecated. Please use new-style autograd function with static forward method. (Example: https://pytorch.org/docs/stable/autograd.html#torch.autograd.Function)
Can yo please help?
The text was updated successfully, but these errors were encountered: