Skip to content

mit0110/kerasAttention

Repository files navigation

Neural Attention in Keras

This repository contains a notebook with some simple prototypes to add an attention mechanism to an LSTM for sequence labeling. It is part of a presentation for the Tensorflow Meetup Buenos Aires, in June 2018.

We also add some scripts to visualize the attention obtained after training for the Named Entity Recognition task in a portion of the 2003 CONLL dataset.

Data

The original dataset was uploaded to Kaggle, along with a vanilla LSTM implementation. We have also hosted it into the UNC servers:

There are also some trained models you can download, as it takes some time to train in the whole dataset even using a GPU:

Requirements

To run the network, we recommend you to use python 3.5 and install

  • Keras 2.1.5
  • scikit-learn 0.19.1
  • pandas 0.23.0

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published