Spaces:
Sleeping
Sleeping
Revert to Qwen2.5-VL-7B-Instruct
Browse files- app.py +3 -3
- evaluator.py +4 -4
app.py
CHANGED
|
@@ -7,7 +7,7 @@ AI image quality assessment using:
|
|
| 7 |
- Multi-image comparison
|
| 8 |
- Technical metrics (sharpness, colorfulness, contrast, CLIP)
|
| 9 |
|
| 10 |
-
Powered by
|
| 11 |
"""
|
| 12 |
|
| 13 |
import gradio as gr
|
|
@@ -892,7 +892,7 @@ with gr.Blocks(title="Image Evaluator", css=DARK_CSS, theme=gr.themes.Base()) as
|
|
| 892 |
<div style="text-align: center; padding: 20px 0 30px 0;">
|
| 893 |
<h1 style="color: #fafafa; font-size: 2.2em; font-weight: 700; margin: 0;">Image Evaluator</h1>
|
| 894 |
<p style="color: #71717a; font-size: 1em; margin-top: 8px;">
|
| 895 |
-
AI image quality assessment powered by <span style="color: #3b82f6;">
|
| 896 |
</p>
|
| 897 |
</div>
|
| 898 |
''')
|
|
@@ -1035,7 +1035,7 @@ with gr.Blocks(title="Image Evaluator", css=DARK_CSS, theme=gr.themes.Base()) as
|
|
| 1035 |
|
| 1036 |
gr.HTML('''
|
| 1037 |
<div style="text-align: center; padding: 30px 0 20px 0; color: #52525b; font-size: 0.85em;">
|
| 1038 |
-
Powered by
|
| 1039 |
</div>
|
| 1040 |
''')
|
| 1041 |
|
|
|
|
| 7 |
- Multi-image comparison
|
| 8 |
- Technical metrics (sharpness, colorfulness, contrast, CLIP)
|
| 9 |
|
| 10 |
+
Powered by Qwen2.5-VL-7B
|
| 11 |
"""
|
| 12 |
|
| 13 |
import gradio as gr
|
|
|
|
| 892 |
<div style="text-align: center; padding: 20px 0 30px 0;">
|
| 893 |
<h1 style="color: #fafafa; font-size: 2.2em; font-weight: 700; margin: 0;">Image Evaluator</h1>
|
| 894 |
<p style="color: #71717a; font-size: 1em; margin-top: 8px;">
|
| 895 |
+
AI image quality assessment powered by <span style="color: #3b82f6;">Qwen2.5-VL-7B</span>
|
| 896 |
</p>
|
| 897 |
</div>
|
| 898 |
''')
|
|
|
|
| 1035 |
|
| 1036 |
gr.HTML('''
|
| 1037 |
<div style="text-align: center; padding: 30px 0 20px 0; color: #52525b; font-size: 0.85em;">
|
| 1038 |
+
Powered by Qwen2.5-VL-7B | Soft-TIFA | CLIP | LPIPS
|
| 1039 |
</div>
|
| 1040 |
''')
|
| 1041 |
|
evaluator.py
CHANGED
|
@@ -204,8 +204,8 @@ class ImageEvaluator:
|
|
| 204 |
|
| 205 |
self.device = device if torch.cuda.is_available() else "cpu"
|
| 206 |
|
| 207 |
-
# Load
|
| 208 |
-
model_name = "Qwen/
|
| 209 |
|
| 210 |
self.vlm_model = AutoModelForImageTextToText.from_pretrained(
|
| 211 |
model_name,
|
|
@@ -877,8 +877,8 @@ class EditEvaluator:
|
|
| 877 |
|
| 878 |
self.device = device if torch.cuda.is_available() else "cpu"
|
| 879 |
|
| 880 |
-
# Load
|
| 881 |
-
model_name = "Qwen/
|
| 882 |
|
| 883 |
self.vlm_model = AutoModelForImageTextToText.from_pretrained(
|
| 884 |
model_name,
|
|
|
|
| 204 |
|
| 205 |
self.device = device if torch.cuda.is_available() else "cpu"
|
| 206 |
|
| 207 |
+
# Load Qwen2.5-VL-7B-Instruct
|
| 208 |
+
model_name = "Qwen/Qwen2.5-VL-7B-Instruct"
|
| 209 |
|
| 210 |
self.vlm_model = AutoModelForImageTextToText.from_pretrained(
|
| 211 |
model_name,
|
|
|
|
| 877 |
|
| 878 |
self.device = device if torch.cuda.is_available() else "cpu"
|
| 879 |
|
| 880 |
+
# Load Qwen2.5-VL-7B-Instruct
|
| 881 |
+
model_name = "Qwen/Qwen2.5-VL-7B-Instruct"
|
| 882 |
|
| 883 |
self.vlm_model = AutoModelForImageTextToText.from_pretrained(
|
| 884 |
model_name,
|