You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Per @punkduckable , multihead attention should be implemented as a layer, not an activation function. However, the current implementation simply uses multihead attention as an activation function, which also disrupts overall structure within MultiLayerPerceptron.
multihead attention should be removed from activation, and probably implemented as a derived class of latent space.
The text was updated successfully, but these errors were encountered:
Per @punkduckable , multihead attention should be implemented as a layer, not an activation function. However, the current implementation simply uses multihead attention as an activation function, which also disrupts overall structure within
MultiLayerPerceptron
.multihead attention should be removed from activation, and probably implemented as a derived class of latent space.
The text was updated successfully, but these errors were encountered: