Releases: Rishit-dagli/Fast-Transformer
Releases · Rishit-dagli/Fast-Transformer
Fast Transformer v0.2.0
✅ Bug Fixes / Improvements
- Unit Tests for output rank and shape
- Looser dependency requirements (now supports all TensorFlow versions >= 2.5.0)
Fast Transformer v0.1.0
This is the initial release of Fast Transformer and implements Fast Transformer as a subclassed TensorFlow model.
Classes
- FastAttention: Implements additive attention as a TensorFlow Keras layer, and supports using relative positional encodings.
- PreNorm: Normalize the activations of the previous layer for each given example in a batch independently and apply some function to it, implemented as a TensorFlow Keras Layer.
- FeedForward: Create a FeedForward neural net with two
Dense
layers and GELU activation, implemented as a TensorFlow Keras Layer. - FastTransformer: Implements the FastTransformer model using all the other classes, allows using rotary embeddings, weight tie projections, and converts to logits. Implemented as a TensorFlow Keras Model.