Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Papers Voting #1

Open
hadyelsahar opened this issue May 1, 2020 · 2 comments
Open

Papers Voting #1

hadyelsahar opened this issue May 1, 2020 · 2 comments

Comments

@hadyelsahar
Copy link
Owner

In this issue you can either:

  • Add papers that you think are interesting to read and discuss (please stick to the format).
  • vote: should be done using 👍 on comments
@hadyelsahar
Copy link
Owner Author

Reformer: The Efficient Transformer
https://arxiv.org/abs/2001.04451

Summary:

Large Transformer models routinely achieve state-of-the-art results on a number of tasks but training these models can be prohibitively costly, especially on long sequences. We introduce two techniques to improve the efficiency of Transformers. For one, we replace dot-product attention by one that uses locality-sensitive hashing, changing its complexity from O(L2) to O(LlogL), where L is the length of the sequence. Furthermore, we use reversible residual layers instead of the standard residuals, which allows storing activations only once in the training process instead of N times, where N is the number of layers. The resulting model, the Reformer, performs on par with Transformer models while being much more memory-efficient and much faster on long sequences.

@hadyelsahar
Copy link
Owner Author

hadyelsahar commented May 1, 2020

Unsupervised Question Decomposition for Question Answering
https://arxiv.org/abs/2002.09758
Twitter thread: https://twitter.com/EthanJPerez/status/1232127027961942018

"We aim to improve question answering (QA) by decomposing hard questions into easier sub-questions that existing QA systems can answer. Since collecting labeled decompositions is cumbersome, we propose an unsupervised approach to produce sub-questions."

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant