ScrewCount / README.md
afshin-dini's picture
update teh header tags
f8e2d87
---
language:
- en
pretty_name: ScrewCount
license: cc-by-nc-nd-4.0
task_categories:
- object-detection
- other
tags:
- object-counting
- few-shot-counting
- exemplar-based-counting
- text-guided-counting
- industrial-vision
- manufacturing
- density-estimation
- dense-object-counting
- small-object-counting
- screws
- nuts
---
# ScrewCount
**ScrewCount** is a dataset for **dense small-object counting in industrial inspection settings**, designed to benchmark both **exemplar-based few-shot counting** and **text-guided object counting**. It focuses on challenging manufacturing scenarios with **small, overlapping, densely packed, and visually similar objects**, specifically **screws** and **nuts**.
The dataset was introduced in the paper: [**ScrewCount: A Dataset and Benchmark for Exemplar Efficiency and Text-Guided Few-Shot Object Counting**](https://www.scitepress.org/PublicationsDetail.aspx?ID=vsx2VblEuWQ=)
## Supported Tasks and Leaderboards
This dataset can be used for:
- **Few-shot object counting**
- **Exemplar-based counting**
- **Text-guided counting**
- **Dense small-object counting**
- **Point-supervised counting**
- **Density map regression**
- **Industrial object detection** (limited, depending on annotation usage)
### Evaluation Metrics
The benchmark primarily uses:
- **MAE (Mean Absolute Error)**
- **RMSE (Root Mean Squared Error)**
These are standard metrics for object counting tasks.
## Dataset Structure
Each sample consists of a high-resolution image of densely packed industrial objects, along with annotations that can support multiple counting paradigms.
A typical sample may include:
- `image`: RGB image
- `points`: point annotations for object instances
- `density_map`: generated density map corresponding to point annotations
- `exemplar_boxes`: up to **6 exemplar bounding boxes** per image for few-shot counting
- `split`: train / val
## Dataset Summary
### Curation Rationale
Most existing counting datasets focus on crowds, cells, or natural scenes. These benchmarks do not fully capture the challenges of **industrial inspection**, where objects are often:
- very small,
- densely packed,
- heavily overlapping,
- partially occluded,
- and visually similar.
ScrewCount was created to address this gap and provide a benchmark tailored to **real-world industrial counting scenarios** where annotation is expensive and few-shot methods are especially relevant.
### Source Data
The dataset contains **high-resolution 3024×4032 images** of:
- **screws**
- **nuts**
The objects vary in:
- size: **small, medium, large**
- shape
- color: **white and black**
Images represent dense manufacturing-like scenes with object counts often in the **hundreds per image**.
### Categories
| Category | Training Samples | Test Samples |
|----------|------------------|--------------|
| Screw | 500 | 100 |
| Nuts | 500 | 100 |
### Summary
- **2 object categories**
- **1000 training images total**
- **200 test images total**
- **Image resolution:** 3024×4032
- **Objects per image:** typically a few hundred
- **Exemplars per image:** 6
## Contributions
Contributions, issue reports, and benchmark reproductions are welcome through the associated repository.
## Citation
If you use this dataset, please cite:
```bibtex
@conference{visapp26,
author={Farnaz Delirie and Afshin Dini and Amirmasoud Molaei and Leila Sadeghi},
title={ScrewCount: A Dataset and Benchmark for Exemplar Efficiency and Text-Guided Few-Shot Object Counting},
booktitle={Proceedings of the 21st International Conference on Computer Vision Theory and Applications - Volume 1: VISAPP},
year={2026},
pages={306-313},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0014236300004084},
isbn={978-989-758-804-4},
issn={2184-4321},
}
```
## License
This model is licensed under the Attribution–NonCommercial 4.0 International License (CC BY-NC 4.0).