Update README.md
Browse files
README.md
CHANGED
|
@@ -332,7 +332,7 @@ We use the following _LLM-based metrics_:
|
|
| 332 |
|
| 333 |
# Results
|
| 334 |
In this section, we give the full experiment results, wherein the metrics of **Prec.**, **Rec.**, **F1.**, **R.L.**, **B.S.**, **Rel.**, **Eff.**, **Comp.**, **Pos.**, and **Avg.** represent image precision, image recall, image F1 score, rouge-l, BERTScore, image relevance, image effectiveness, comprehensive score, image position score, and average score, respectively. Specifically, the metric **Ord.** represents image ordering score.
|
| 335 |
-
## Comprehensive performance results on Wit(
|
| 336 |
| Framework | Model | MRAMG-Wit | | | | | | | | | |
|
| 337 |
|------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
|
| 338 |
| | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
|
|
@@ -396,7 +396,7 @@ In this section, we give the full experiment results, wherein the metrics of **P
|
|
| 396 |
| | Llama-3.3-70B-Instruct | 70.61 | 94.40 | 76.35 | 47.86 | 95.47 | 78.16 | 71.84 | 76.96 | 71.46 | 75.90 |
|
| 397 |
|
| 398 |
## Comprehensive performance results on MRAMG-Web(Web Dataset).
|
| 399 |
-
| Framework | Model | MRAMG-Web
|
| 400 |
|------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
|
| 401 |
| | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
|
| 402 |
| Rule-Based | GPT-4o | 32.47 | 16.93 | 22.11 | 39.17 | 90.56 | 29.47 | 27.81 | 73.87 | 32.80 | 40.58 |
|
|
|
|
| 332 |
|
| 333 |
# Results
|
| 334 |
In this section, we give the full experiment results, wherein the metrics of **Prec.**, **Rec.**, **F1.**, **R.L.**, **B.S.**, **Rel.**, **Eff.**, **Comp.**, **Pos.**, and **Avg.** represent image precision, image recall, image F1 score, rouge-l, BERTScore, image relevance, image effectiveness, comprehensive score, image position score, and average score, respectively. Specifically, the metric **Ord.** represents image ordering score.
|
| 335 |
+
## Comprehensive performance results on MRAMG-Wit(Web Dataset).
|
| 336 |
| Framework | Model | MRAMG-Wit | | | | | | | | | |
|
| 337 |
|------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
|
| 338 |
| | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
|
|
|
|
| 396 |
| | Llama-3.3-70B-Instruct | 70.61 | 94.40 | 76.35 | 47.86 | 95.47 | 78.16 | 71.84 | 76.96 | 71.46 | 75.90 |
|
| 397 |
|
| 398 |
## Comprehensive performance results on MRAMG-Web(Web Dataset).
|
| 399 |
+
| Framework | Model | MRAMG-Web | | | | | | | | | |
|
| 400 |
|------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
|
| 401 |
| | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
|
| 402 |
| Rule-Based | GPT-4o | 32.47 | 16.93 | 22.11 | 39.17 | 90.56 | 29.47 | 27.81 | 73.87 | 32.80 | 40.58 |
|