qinhany commited on
Commit
68627cb
·
verified ·
1 Parent(s): 6c1115d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -332,7 +332,7 @@ We use the following _LLM-based metrics_:
332
 
333
  # Results
334
  In this section, we give the full experiment results, wherein the metrics of **Prec.**, **Rec.**, **F1.**, **R.L.**, **B.S.**, **Rel.**, **Eff.**, **Comp.**, **Pos.**, and **Avg.** represent image precision, image recall, image F1 score, rouge-l, BERTScore, image relevance, image effectiveness, comprehensive score, image position score, and average score, respectively. Specifically, the metric **Ord.** represents image ordering score.
335
- ## Comprehensive performance results on Wit(Wit-MQA).
336
  | Framework | Model | MRAMG-Wit | | | | | | | | | |
337
  |------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
338
  | | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
@@ -396,7 +396,7 @@ In this section, we give the full experiment results, wherein the metrics of **P
396
  | | Llama-3.3-70B-Instruct | 70.61 | 94.40 | 76.35 | 47.86 | 95.47 | 78.16 | 71.84 | 76.96 | 71.46 | 75.90 |
397
 
398
  ## Comprehensive performance results on MRAMG-Web(Web Dataset).
399
- | Framework | Model | MRAMG-Web+ | | | | | | | | | |
400
  |------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
401
  | | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
402
  | Rule-Based | GPT-4o | 32.47 | 16.93 | 22.11 | 39.17 | 90.56 | 29.47 | 27.81 | 73.87 | 32.80 | 40.58 |
 
332
 
333
  # Results
334
  In this section, we give the full experiment results, wherein the metrics of **Prec.**, **Rec.**, **F1.**, **R.L.**, **B.S.**, **Rel.**, **Eff.**, **Comp.**, **Pos.**, and **Avg.** represent image precision, image recall, image F1 score, rouge-l, BERTScore, image relevance, image effectiveness, comprehensive score, image position score, and average score, respectively. Specifically, the metric **Ord.** represents image ordering score.
335
+ ## Comprehensive performance results on MRAMG-Wit(Web Dataset).
336
  | Framework | Model | MRAMG-Wit | | | | | | | | | |
337
  |------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
338
  | | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
 
396
  | | Llama-3.3-70B-Instruct | 70.61 | 94.40 | 76.35 | 47.86 | 95.47 | 78.16 | 71.84 | 76.96 | 71.46 | 75.90 |
397
 
398
  ## Comprehensive performance results on MRAMG-Web(Web Dataset).
399
+ | Framework | Model | MRAMG-Web | | | | | | | | | |
400
  |------------|------------------------|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|
401
  | | | Prec. | Rec. | F1 | R.L. | B.S. | Rel. | Eff. | Comp. | Pos. | Avg. |
402
  | Rule-Based | GPT-4o | 32.47 | 16.93 | 22.11 | 39.17 | 90.56 | 29.47 | 27.81 | 73.87 | 32.80 | 40.58 |