Instructions to use Universal-NER/UniNER-7B-definition with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use Universal-NER/UniNER-7B-definition with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="Universal-NER/UniNER-7B-definition")# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("Universal-NER/UniNER-7B-definition") model = AutoModelForCausalLM.from_pretrained("Universal-NER/UniNER-7B-definition") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use Universal-NER/UniNER-7B-definition with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "Universal-NER/UniNER-7B-definition" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "Universal-NER/UniNER-7B-definition", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/Universal-NER/UniNER-7B-definition
- SGLang
How to use Universal-NER/UniNER-7B-definition with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "Universal-NER/UniNER-7B-definition" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "Universal-NER/UniNER-7B-definition", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "Universal-NER/UniNER-7B-definition" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "Universal-NER/UniNER-7B-definition", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use Universal-NER/UniNER-7B-definition with Docker Model Runner:
docker model run hf.co/Universal-NER/UniNER-7B-definition
UniNER-7B-definition
Description: A UniNER-7B model trained from LLama-7B using the Pile-NER-definition data without human-labeled data. The data was collected by prompting gpt-3.5-turbo-0301 to label entities from passages and provide short-sentence definitions. The data collection prompt is as follows:
Given a paragraph, your task is to extract all entities and concepts, and define their type using a short sentence. The output should be in the following format: [("entity", "definition of entity type in a short sentence"), ... ]
Check our paper for more information. Check our repo about how to use the model.
Comparison with UniNER-7B-type
The UniNER-7B-type model, trained on Pile-NER-type, excels in recognizing common and short NER tags (e.g., person, location) and performs better on NER datasets. On the other hand, UniNER-7B-definition demonstrates superior capabilities in understanding short-sentence definitions of entity types. Additionally, it exhibits enhanced robustness against variations in type paraphrasing.
Inference
The template for inference instances is as follows:
A virtual assistant answers questions from a user based on the provided text.
USER: Text: {Fill the input text here}
ASSISTANT: I’ve read this text.
USER: What describes {Fill the entity type here} in the text?
ASSISTANT: (model's predictions in JSON format)
Note: Inferences are based on one entity type at a time. For multiple entity types, create separate instances for each type.
License
This model and its associated data are released under the CC BY-NC 4.0 license. They are primarily used for research purposes.
Citation
@article{zhou2023universalner,
title={UniversalNER: Targeted Distillation from Large Language Models for Open Named Entity Recognition},
author={Wenxuan Zhou and Sheng Zhang and Yu Gu and Muhao Chen and Hoifung Poon},
year={2023},
eprint={2308.03279},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
- Downloads last month
- 857