Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -207,18 +207,29 @@ pip install torch transformers accelerate pillow sentencepiece protobuf
|
|
| 207 |
- Black & white documents may receive lower scores (use Overall model instead)
|
| 208 |
- Requires GPU with sufficient VRAM for efficient inference
|
| 209 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 210 |
## Citation
|
| 211 |
|
|
|
|
|
|
|
| 212 |
```bibtex
|
| 213 |
-
@
|
| 214 |
-
title={DeQA-Doc
|
| 215 |
-
author={
|
| 216 |
-
|
| 217 |
-
|
| 218 |
-
url={https://huggingface.co/mapo80/DeQA-Doc-Color}
|
| 219 |
}
|
| 220 |
```
|
| 221 |
|
|
|
|
|
|
|
| 222 |
## License
|
| 223 |
|
| 224 |
Apache 2.0
|
|
|
|
| 207 |
- Black & white documents may receive lower scores (use Overall model instead)
|
| 208 |
- Requires GPU with sufficient VRAM for efficient inference
|
| 209 |
|
| 210 |
+
## Credits & Attribution
|
| 211 |
+
|
| 212 |
+
This model is based on the **DeQA-Doc** project by Junjie Gao et al., which won the **Championship** in the VQualA 2025 DIQA (Document Image Quality Assessment) Challenge.
|
| 213 |
+
|
| 214 |
+
**Original Repository**: [https://github.com/Junjie-Gao19/DeQA-Doc](https://github.com/Junjie-Gao19/DeQA-Doc)
|
| 215 |
+
|
| 216 |
+
All credit for the research, training methodology, and model architecture goes to the original authors.
|
| 217 |
+
|
| 218 |
## Citation
|
| 219 |
|
| 220 |
+
If you use this model in your research, please cite the original paper:
|
| 221 |
+
|
| 222 |
```bibtex
|
| 223 |
+
@inproceedings{deqadoc,
|
| 224 |
+
title={{DeQA-Doc}: Adapting {DeQA-Score} to Document Image Quality Assessment},
|
| 225 |
+
author={Gao, Junjie and Liu, Runze and Peng, Yingzhe and Yang, Shujian and Zhang, Jin and Yang, Kai and You, Zhiyuan},
|
| 226 |
+
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop},
|
| 227 |
+
year={2025},
|
|
|
|
| 228 |
}
|
| 229 |
```
|
| 230 |
|
| 231 |
+
**ArXiv**: [https://arxiv.org/abs/2507.12796](https://arxiv.org/abs/2507.12796)
|
| 232 |
+
|
| 233 |
## License
|
| 234 |
|
| 235 |
Apache 2.0
|