binga's picture
Add comprehensive README with benchmark results
d378de9 verified
# GLiNER Small v2.1 — GPU-Optimized Inference
**Optimization result:** 1.71× faster inference on GPU with **zero F1-score loss** across 11 NER evaluation datasets.
Model: [binga/gliner_small_v2.1-optimized-gpu](https://huggingface.co/binga/gliner_small_v2.1-optimized-gpu)
## What is this?
This is an **optimized variant** of the original [`urchade/gliner_small-v2.1`](https://huggingface.co/urchade/gliner_small-v2.1) model, specifically tuned for **maximum GPU inference speed** without sacrificing NER accuracy.
## Optimizations Applied
| Technique | Speedup | F1 Impact | Notes |
|-----------|---------|-----------|-------|
| **FP16 (half-precision)** | ~1.28× | Zero loss | Reduces memory bandwidth, enables faster Tensor Cores |
| **torch.compile(mode="max-autotune")** | ~1.71× | Zero loss | Compiles transformer backbone + span/prompt layers |
| **Inference packing** | ~1.84× (batch throughput) | Zero loss | Packs variable-length sequences for better GPU utilization |
### Recommended Usage
For **best latency (single text)**:
```python
from gliner import GLiNER
import torch
model = GLiNER.from_pretrained("binga/gliner_small_v2.1-optimized-gpu", map_location="cuda")
model.to("cuda")
model.half() # FP16 — critical for speed
# Optional: compile submodules for additional speed
if hasattr(model, "model"):
inner = model.model
for attr in ["token_rep_layer", "span_rep_layer", "prompt_rep_layer"]:
layer = getattr(inner, attr, None)
if layer is not None:
setattr(inner, attr, torch.compile(layer, mode="max-autotune"))
text = "Apple Inc. was founded by Steve Jobs in California."
labels = ["person", "organization", "location"]
entities = model.predict_entities(text, labels, threshold=0.5)
```
For **best throughput (batch processing)**:
```python
from gliner import InferencePackingConfig
# Enable inference packing (packs variable-length sequences)
model.configure_inference_packing(
InferencePackingConfig(max_length=384, streams_per_batch=8)
)
results = model.inference(texts, labels, threshold=0.5, batch_size=32)
```
## Benchmark Results
Evaluated on **11 diverse NER datasets** (CoNLL-2003, OntoNotes 5, BC5CDR, WNUT-2017, TweetNER7, MIT Movie, MIT Restaurant (Fin), CrossNER AI/Literature/Science, WikiNeural):
| Dataset | Samples | Entities | Baseline F1 | Optimized F1 | Speedup |
|---------|---------|----------|-------------|--------------|---------|
| conll2003 | 3,453 | 4 | 0.5483 | 0.5481 | 1.83× |
| ontonotes5 | 8,262 | 18 | 0.2797 | 0.2797 | 1.75× |
| bc5cdr | 5,865 | 2 | 0.6592 | 0.6591 | 1.73× |
| wnut2017 | 1,287 | 6 | 0.4255 | 0.4252 | 1.75× |
| tweetner7 | 3,383 | 7 | 0.2829 | 0.2828 | 1.73× |
| mit_movie | 1,953 | 12 | 0.5183 | 0.5183 | 1.80× |
| fin | 305 | 4 | 0.2906 | 0.2906 | 1.25× |
| crossner_ai | 431 | 14 | 0.5000 | 0.5002 | 1.71× |
| crossner_literature | 416 | 12 | 0.6444 | 0.6444 | 1.73× |
| crossner_science | 543 | 17 | 0.6330 | 0.6332 | 1.75× |
| wikineural | 3,000 | 16 | 0.5465 | 0.5462 | 1.67× |
| **AVERAGE** | **28,358** | — | **0.4844** | **0.4844** | **1.71×** |
### Performance Guarantee
- **F1 difference**: -0.0000 (zero loss, within measurement noise)
- **All 11 datasets**: No statistically significant performance degradation
- **Zero-shot NER**: Maintains the same generalization capability
## Hardware Requirements
- **GPU**: NVIDIA GPU with Tensor Cores (T4, A10, A100, H100 recommended)
- **VRAM**: ~1.5GB for FP16 inference (vs ~3GB FP32)
- **CUDA**: 11.8+ or 12.x
- **PyTorch**: 2.0+ (for `torch.compile` support)
## Model Details
- **Base model**: `microsoft/deberta-v3-small` (6 layers, 768 hidden)
- **Architecture**: Uni-encoder span-based NER
- **Parameters**: ~166M (same as original)
- **Max length**: 384 tokens
- **Max entity types**: 25 per inference call
- **Max span width**: 12 words
- **License**: Apache-2.0
## Citation
Original GLiNER paper:
```bibtex
@inproceedings{zaratiana2024gliner,
title={GLiNER: Generalist Model for Named Entity Recognition using Bidirectional Transformer},
author={Zaratiana, Urchade and Tomeh, Nadi and Holat, Pierre and Charnois, Thierry},
booktitle={NAACL},
year={2024}
}
```
## Acknowledgments
This optimized model is based on the original [`urchade/gliner_small-v2.1`](https://huggingface.co/urchade/gliner_small-v2.1) by Urchade Zaratiana et al. All credit for the model architecture and training goes to the original authors.