| base_model: | |
| - Qwen/Qwen2.5-VL-7B-Instruct | |
| license: mit | |
| pipeline_tag: image-text-to-text | |
| library_name: transformers | |
| # MagicAssessor-7B | |
| MagicAssessor-7B is a Vision-Language Model (VLM) developed for fine-grained artifact assessment in text-to-image generation. It is a core component of the comprehensive **MagicMirror** framework, which aims to systematically evaluate the perceptual quality and identify various anatomical and structural flaws in generated images. | |
| The model was introduced in the paper [MagicMirror: A Large-Scale Dataset and Benchmark for Fine-Grained Artifacts Assessment in Text-to-Image Generation](https://arxiv.org/abs/2509.10260). | |
| * **Paper**: [arXiv:2509.10260](https://arxiv.org/abs/2509.10260) | [Hugging Face Papers: 2509.10260](https://huggingface.co/papers/2509.10260) | |
| * **Project Page**: https://wj-inf.github.io/MagicMirror-page/ | |
| * **Code / GitHub Repository (MagicMirror Benchmark)**: https://github.com/wj-inf/MagicMirror | |
| * **Dataset (MagicData340K)**: https://huggingface.co/datasets/wj-inf/MagicData340k | |
| * **Model (MagicAssessor-7B - this repository)**: https://huggingface.co/wj-inf/MagicAssessor-7B |