Skip to content
View universal-ner's full-sized avatar

Block or report universal-ner

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
universal-ner/README.md

UniversalNER 🚀

UniversalNER: Targeted Distillation from Large Language Models for Open Named Entity Recognition
Wenxuan Zhou*, Sheng Zhang*, Yu Gu, Muhao Chen, Hoifung Poon (*Equal Contribution)

[Project Page] [Demo] [Paper] [Data] [Model]

Release

  • [12/4] We add the evaluation code for NER datasets.
  • [9/14] We add our training code for finetuning the LLama base model with UniversalNER data.
  • [8/11] We release two more UniNER models, UniNER-7B-type-sup and UniNER-7B-all, which were finetuned on ChatGPT-generated data and 40 supervised datasets of various domains and offers better NER performance.
  • [8/10] We have released the inference code for running the model checkpoints. The code for pretraining and evaluation will be released soon.

Model License Data License Usage and License Notices: The data and model checkpoints are intended and licensed for research use only. They are also restricted to uses that follow the license agreement of LLaMA, Vicuna, and ChatGPT.

Install

This project relies on vllm. Ensure you have gcc version 5 or later, and CUDA versions between 11.0 and 11.8, as specified in the installation requirements for vllm.

  1. Clone this repository and navigate to the folder
git clone https://github.com/universal-ner/universal-ner.git
cd universal-ner
  1. Install the required packages
pip install -r requirements.txt

Demo

We use vllm for inference. The inference can be run with a single V100 16G GPU.

Gradio Web UI

To launch a Gradio demo locally, run the following command:

python -m src.serve.gradio_server \
    --model_path Universal-NER/UniNER-7B-type \
    --tensor_parallel_size 1 \
    --max_input_length 512

CLI Inference

Run the following command to use vllm for inference:

python -m src.serve.cli \
    --model_path Universal-NER/UniNER-7B-type \
    --tensor_parallel_size 1 \
    --max_input_length 512

Run the following command to use Huggingface Transformers for inference:

python -m src.serve.hf \
    --model_path Universal-NER/UniNER-7B-type \
    --tensor_parallel_size 1 \
    --max_input_length 512

Finetuning

Our training code is adapted from FastChat. See here for how to finetune the LLama base model with UniversalNER data.

Evaluation

To execute the evaluation process:

python -m src.eval.evaluate \
    --model_path Universal-NER/UniNER-7B-type \
    --data_path ./src/eval/test_data/CrossNER_AI.json \
    --tensor_parallel_size 1

Due to licensing restrictions associated with many NER datasets used in our evaluation, only the CrossNER and MIT datasets are included in this repository.

Citation

If you find UniversalNER helpful for your research and applications, please cite using this BibTeX:

@article{zhou2023universalner,
      title={UniversalNER: Targeted Distillation from Large Language Models for Open Named Entity Recognition}, 
      author={Wenxuan Zhou and Sheng Zhang and Yu Gu and Muhao Chen and Hoifung Poon},
      year={2023},
      eprint={2308.03279},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

Popular repositories Loading

  1. universal-ner universal-ner Public

    Python 345 28

  2. universal-ner.github.io universal-ner.github.io Public

    HTML