-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ask about the approach #8
Comments
Of course, you can use other architectures such as HAN to process the news body, and the performance is usually slightly better. But it usually requires a larger GPU memory/smaller batch size. |
Yeah, I've tried to encode each sentence then encode article body on 2080Ti GPU. I can only train with batch size = 1 and it took 25s/step, maybe something went wrong. |
HAN stands for Hierarchical Attention Network, right? |
Yeah, HAN means Hierarchical Attention Network (Yang et al., 2016). You can replace the LSTM with CNN to boost the training speed. |
Yeup, thank you! |
In addition, it is highly recommended that you can use a smaller sentence length or fewer sentences. Although I believe that using the full news body is beneficial, it takes a large amount of GPU memory and the improvement is usually marginal. |
Yeah, thanks for your recommendation |
Thank you for sharing this great repository!
Can you share me the reason why you consider article body as a long sequence instead of sentences. If I want to encode each sentence then use sentences represent vector to encode article body, is it possible?
The text was updated successfully, but these errors were encountered: