Skip to content

Latest commit

 

History

History
15 lines (12 loc) · 679 Bytes

README.md

File metadata and controls

15 lines (12 loc) · 679 Bytes

Transformer

This is the implementation of the paper Attention Is All You Need. The code has been written in Python, so make the necessary changes to the scripts to run them. Thanks to Umar Jamilai and his videos on the topic, they were extremely helpful and valuable.

More comments will be added to make understanding the code easier.

I have used the opus_books dataset for this code.

Citation:

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2017).
Attention Is All You Need.
ArXiv.