Update README.md
Browse files
README.md
CHANGED
|
@@ -333,7 +333,7 @@ We use the following _LLM-based metrics_:
|
|
| 333 |
# Results
|
| 334 |
In this section, we give the full experiment results, wherein the metrics of **Prec.**, **Rec.**, **F1.**, **R.L.**, **B.S.**, **Rel.**, **Eff.**, **Comp.**, **Pos.**, and **Avg.** represent image precision, image recall, image F1 score, rouge-l, BERTScore, image relevance, image effectiveness, comprehensive score, image position score, and average score, respectively. Specifically, the metric **Ord.** represents image ordering score.
|
| 335 |
## Comprehensive performance results on MRAMG-Wit(Web Dataset).
|
| 336 |
-
| Framework | Model | MRAMG-Wit
|
| 337 |
|------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
|
| 338 |
| | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
|
| 339 |
| Rule-Based | GPT-4o | 49.50 | 49.67 | 49.56 | 56.23 | 92.27 | 43.67 | 39.50 | 77.00 | 50.08 | 56.39 |
|
|
|
|
| 333 |
# Results
|
| 334 |
In this section, we give the full experiment results, wherein the metrics of **Prec.**, **Rec.**, **F1.**, **R.L.**, **B.S.**, **Rel.**, **Eff.**, **Comp.**, **Pos.**, and **Avg.** represent image precision, image recall, image F1 score, rouge-l, BERTScore, image relevance, image effectiveness, comprehensive score, image position score, and average score, respectively. Specifically, the metric **Ord.** represents image ordering score.
|
| 335 |
## Comprehensive performance results on MRAMG-Wit(Web Dataset).
|
| 336 |
+
| Framework | Model | MRAMG-Wit |
|
| 337 |
|------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
|
| 338 |
| | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
|
| 339 |
| Rule-Based | GPT-4o | 49.50 | 49.67 | 49.56 | 56.23 | 92.27 | 43.67 | 39.50 | 77.00 | 50.08 | 56.39 |
|