nielsr HF Staff commited on
Commit
5ec1520
·
verified ·
1 Parent(s): 7d5aa56

Improve dataset card for Zenodo10K: Add VLM-SlideEval paper context, task category, and license

Browse files

This pull request significantly improves the dataset card for `PPTAgent/Zenodo10K` by integrating information from the `VLM-SlideEval: Evaluating VLMs on Structured Comprehension and Perturbation Sensitivity in PPT` paper (https://huggingface.co/papers/2510.22045), which utilizes this dataset for evaluating Vision-Language Models.

Key changes include:
* Adding `task_categories: ['image-text-to-text']` to the top-level YAML metadata, reflecting the multimodal nature of the dataset's use in VLM evaluation.
* Adding `license: other` to the top-level YAML metadata, acknowledging the heterogeneous licenses of the individual PowerPoint files.
* Updating the main description to clearly link to both the originating `PPTAgent` paper and the `VLM-SlideEval` paper.
* Incorporating the abstract of the `VLM-SlideEval` paper to explain its relevance for VLM evaluation on presentation slides.
* Adding the BibTeX citation for the `VLM-SlideEval` paper, alongside the existing `PPTAgent` citation, to ensure full attribution.
* Retaining the existing GitHub link for `PPTAgent` and the associated sample usage, as they are part of the dataset's original context.

These updates enhance the discoverability and understanding of the dataset's utility across different research contexts.

Files changed (1) hide show
  1. README.md +19 -4
README.md CHANGED
@@ -1,4 +1,7 @@
1
  ---
 
 
 
2
  dataset_info:
3
  features:
4
  - name: filename
@@ -34,10 +37,15 @@ configs:
34
 
35
  # PPTAgent/Zenodo10K
36
 
37
- This is the dataset used in [PPTAgent](https://arxiv.org/abs/2501.03936), crawled from [zenodo](http://zenodo.org).
38
  To the best of our knowledge, it is the **largest presentation dataset** currently available, comprising over 10,000 **PowerPoint (.pptx)** files, all distributed under a clear and compliant license.
39
 
40
- For more information, please visit our [github repo](https://github.com/icip-cas/PPTAgent).
 
 
 
 
 
41
 
42
  ```python
43
  dirname = f"zenodo-pptx/pptx/{task['license']}/{task['created'][:4]}/"
@@ -62,5 +70,12 @@ If you find this project helpful, please use the following to cite it:
62
  url={https://arxiv.org/abs/2501.03936},
63
  }
64
  ```
65
-
66
-
 
 
 
 
 
 
 
 
1
  ---
2
+ task_categories:
3
+ - image-text-to-text
4
+ license: other
5
  dataset_info:
6
  features:
7
  - name: filename
 
37
 
38
  # PPTAgent/Zenodo10K
39
 
40
+ This is the dataset used in [PPTAgent: Generating and Evaluating Presentations Beyond Text-to-Slides](https://arxiv.org/abs/2501.03936) and also used for evaluation in [VLM-SlideEval: Evaluating VLMs on Structured Comprehension and Perturbation Sensitivity in PPT](https://huggingface.co/papers/2510.22045). It was crawled from [zenodo](http://zenodo.org).
41
  To the best of our knowledge, it is the **largest presentation dataset** currently available, comprising over 10,000 **PowerPoint (.pptx)** files, all distributed under a clear and compliant license.
42
 
43
+ For more information regarding the `PPTAgent` project, please visit its [github repo](https://github.com/icip-cas/PPTAgent).
44
+
45
+ ## VLM-SlideEval Abstract
46
+ Vision-language models (VLMs) are increasingly used to evaluate multimodal content, including presentation slides, yet their slide-specific understanding remains underexplored {despite their growing role as critics in agentic, model-forward pipelines}. We introduce VLM-SlideEval, an evaluation framework that probes VLMs along three axes: (1) element-level extraction from slide images aligned to ground truth; (2) robustness to controlled perturbations in geometry, style, and text; and (3) higher-level comprehension, such as recovering a deck's narrative order from shuffled slides. Using publicly available decks from Zenodo ( this https URL ), we standardize ground-truth element metadata from PowerPoint XML and live renderings into a unified, verifiable schema. Empirically, VLMs underperform on pixel-accurate extraction and show non-trivial agreement, fidelity, and consistency under controlled perturbations, while performing better on single-slide content understanding; however, they do not reliably capture narrative structure across slides. These results highlight the limits of current VLMs for slide evaluation and motivate calibrated, critic-in-the-loop evaluators that drive iterative refinement and selection in agentic pipelines.
47
+
48
+ ## Sample Usage
49
 
50
  ```python
51
  dirname = f"zenodo-pptx/pptx/{task['license']}/{task['created'][:4]}/"
 
70
  url={https://arxiv.org/abs/2501.03936},
71
  }
72
  ```
73
+ If you use this dataset for VLM evaluation, please also cite:
74
+ ```bibtex
75
+ @article{xu2025vlm,
76
+ title={VLM-SlideEval: Evaluating VLMs on Structured Comprehension and Perturbation Sensitivity in PPT},
77
+ author={Xu, Charles and Li, Qiyang and Luo, Jianlan and Levine, Sergey},
78
+ journal={arXiv preprint arXiv:2510.22045},
79
+ year={2025}
80
+ }
81
+ ```