You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I think your implementation is fine. The reason for including W is due to the combination of variational dropout layer and fully connected layer. Despite the clip/mask operations, in the later computation, the W factor was canceled out. One minor difference though, is that your parameterization is the neuron-level dropout approach, where other implementations are the weight-level drop-connect approach equivalent.
Here:
variational-dropout/varout/layers.py
Lines 334 to 340 in 8d18ec2
Should depend on
w
as here: https://github.com/BayesWatch/tf-variational-dropoutThe text was updated successfully, but these errors were encountered: