# NER Benchmark Results **Model:** Minibase-NER-Small **Dataset:** ner_benchmark_dataset.jsonl **Sample Size:** 100 **Date:** 2025-10-07T13:20:42.785262 ## Overall Performance | Metric | Score | Description | |--------|-------|-------------| | F1 Score | 0.435 | Overall NER performance (harmonic mean of precision and recall) | | Precision | 0.630 | Accuracy of entity predictions | | Recall | 0.343 | Ability to find all entities | | Average Latency | 76.6ms | Response time performance | ## Entity Type Performance | Entity Type | Accuracy | Correct/Total | |-------------|----------|---------------| | ENTITY | 0.936 | 103/110 | ## Key Improvements - **BIO Tagging**: Model outputs entities in BIO (Beginning-Inside-Outside) format - **Multiple Entity Types**: Supports PERSON, ORG, LOC, and MISC entities - **Entity-Level Evaluation**: Metrics calculated at entity level rather than token level - **Comprehensive Coverage**: Evaluates across different text domains ## Example Results ### Example 1 **Input:** John Smith works at Google in New York and uses Python programming language.... **Predicted:** PERGON, ORG... **F1 Score:** 0.000 ### Example 2 **Input:** Microsoft Corporation announced that Satya Nadella will visit London next week.... **Predicted:** 1. Microsoft Corporation... **F1 Score:** 0.500 ### Example 3 **Input:** The University of Cambridge is located in the United Kingdom and was founded by King Henry III.... **Predicted:** 1. The University of Cambridge 2. King Henry III... **F1 Score:** 0.800