Datasets:
Improve dataset card for Zenodo10K: Add VLM-SlideEval paper context, task category, and license
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,4 +1,7 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
dataset_info:
|
| 3 |
features:
|
| 4 |
- name: filename
|
|
@@ -34,10 +37,15 @@ configs:
|
|
| 34 |
|
| 35 |
# PPTAgent/Zenodo10K
|
| 36 |
|
| 37 |
-
This is the dataset used in [PPTAgent](https://arxiv.org/abs/2501.03936)
|
| 38 |
To the best of our knowledge, it is the **largest presentation dataset** currently available, comprising over 10,000 **PowerPoint (.pptx)** files, all distributed under a clear and compliant license.
|
| 39 |
|
| 40 |
-
For more information, please visit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 41 |
|
| 42 |
```python
|
| 43 |
dirname = f"zenodo-pptx/pptx/{task['license']}/{task['created'][:4]}/"
|
|
@@ -62,5 +70,12 @@ If you find this project helpful, please use the following to cite it:
|
|
| 62 |
url={https://arxiv.org/abs/2501.03936},
|
| 63 |
}
|
| 64 |
```
|
| 65 |
-
|
| 66 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
task_categories:
|
| 3 |
+
- image-text-to-text
|
| 4 |
+
license: other
|
| 5 |
dataset_info:
|
| 6 |
features:
|
| 7 |
- name: filename
|
|
|
|
| 37 |
|
| 38 |
# PPTAgent/Zenodo10K
|
| 39 |
|
| 40 |
+
This is the dataset used in [PPTAgent: Generating and Evaluating Presentations Beyond Text-to-Slides](https://arxiv.org/abs/2501.03936) and also used for evaluation in [VLM-SlideEval: Evaluating VLMs on Structured Comprehension and Perturbation Sensitivity in PPT](https://huggingface.co/papers/2510.22045). It was crawled from [zenodo](http://zenodo.org).
|
| 41 |
To the best of our knowledge, it is the **largest presentation dataset** currently available, comprising over 10,000 **PowerPoint (.pptx)** files, all distributed under a clear and compliant license.
|
| 42 |
|
| 43 |
+
For more information regarding the `PPTAgent` project, please visit its [github repo](https://github.com/icip-cas/PPTAgent).
|
| 44 |
+
|
| 45 |
+
## VLM-SlideEval Abstract
|
| 46 |
+
Vision-language models (VLMs) are increasingly used to evaluate multimodal content, including presentation slides, yet their slide-specific understanding remains underexplored {despite their growing role as critics in agentic, model-forward pipelines}. We introduce VLM-SlideEval, an evaluation framework that probes VLMs along three axes: (1) element-level extraction from slide images aligned to ground truth; (2) robustness to controlled perturbations in geometry, style, and text; and (3) higher-level comprehension, such as recovering a deck's narrative order from shuffled slides. Using publicly available decks from Zenodo ( this https URL ), we standardize ground-truth element metadata from PowerPoint XML and live renderings into a unified, verifiable schema. Empirically, VLMs underperform on pixel-accurate extraction and show non-trivial agreement, fidelity, and consistency under controlled perturbations, while performing better on single-slide content understanding; however, they do not reliably capture narrative structure across slides. These results highlight the limits of current VLMs for slide evaluation and motivate calibrated, critic-in-the-loop evaluators that drive iterative refinement and selection in agentic pipelines.
|
| 47 |
+
|
| 48 |
+
## Sample Usage
|
| 49 |
|
| 50 |
```python
|
| 51 |
dirname = f"zenodo-pptx/pptx/{task['license']}/{task['created'][:4]}/"
|
|
|
|
| 70 |
url={https://arxiv.org/abs/2501.03936},
|
| 71 |
}
|
| 72 |
```
|
| 73 |
+
If you use this dataset for VLM evaluation, please also cite:
|
| 74 |
+
```bibtex
|
| 75 |
+
@article{xu2025vlm,
|
| 76 |
+
title={VLM-SlideEval: Evaluating VLMs on Structured Comprehension and Perturbation Sensitivity in PPT},
|
| 77 |
+
author={Xu, Charles and Li, Qiyang and Luo, Jianlan and Levine, Sergey},
|
| 78 |
+
journal={arXiv preprint arXiv:2510.22045},
|
| 79 |
+
year={2025}
|
| 80 |
+
}
|
| 81 |
+
```
|