--- license: apache-2.0 tags: - ner - gliner - zero-shot - bootstrap - uv-script size_categories: - n<10K --- # davanstrien/eval-mentions-bootstrap Bootstrap NER dataset produced by [`urchade/gliner_multi-v2.1`](https://huggingface.co/urchade/gliner_multi-v2.1) over [`/input/cleaned-cards.parquet`](https://huggingface.co/datasets//input/cleaned-cards.parquet). Generated using [`uv-scripts/gliner/extract-entities.py`](https://huggingface.co/datasets/uv-scripts/gliner). ## Provenance | | | |---|---| | Source dataset | `/input/cleaned-cards.parquet` (split `train`) | | Text column | `card` | | Bootstrap model | `urchade/gliner_multi-v2.1` | | Entity types | `benchmark name, evaluation dataset, evaluation metric` | | Confidence threshold | 0.6 | | Samples processed | 10000 | | Total entities extracted | 15811 | | Inference device | `cuda` | | Wall clock | 951.7s (10.51 samples/s) | ## Schema Original `/input/cleaned-cards.parquet` columns plus an `entities` column: ```python entities: list of { "start": int, # character offset, inclusive "end": int, # character offset, exclusive "text": str, # the matched span "label": str, # one of ['benchmark name', 'evaluation dataset', 'evaluation metric'] "score": float, # GLiNER confidence in [0, 1] } ``` ## Caveats - These are **bootstrap labels**, not human-reviewed. Treat low-confidence (< 0.7) entities as candidates for review. - GLiNER is zero-shot: changing `--entity-types` changes what it extracts, but quality varies by entity type. - Long texts were truncated at 8000 characters before inference.