Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Wrong logalpha parameterisation for sparsifying variational dropout #10

Open
gngdb opened this issue Oct 16, 2017 · 1 comment
Open

Comments

@gngdb
Copy link
Owner

gngdb commented Oct 16, 2017

Here:

logalpha = T.log(T.nnet.sigmoid(self.logitalpha)).eval()
# remove the old parameter
del self.params[self.logitalpha]
del self.logitalpha
self.logalpha = theano.shared(
value=logalpha,
name='logalpha')

Should depend on w as here: https://github.com/BayesWatch/tf-variational-dropout

@kzhai
Copy link

kzhai commented Nov 28, 2017

I think your implementation is fine. The reason for including W is due to the combination of variational dropout layer and fully connected layer. Despite the clip/mask operations, in the later computation, the W factor was canceled out. One minor difference though, is that your parameterization is the neuron-level dropout approach, where other implementations are the weight-level drop-connect approach equivalent.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants