Commit ·
c2db44c
1
Parent(s): 1bb6656
Initial commit
Browse files
README.md
CHANGED
|
@@ -70,7 +70,7 @@ model-runs/
|
|
| 70 |
├── bertopic/
|
| 71 |
├── ctm/
|
| 72 |
└── mallet/
|
| 73 |
-
```
|
| 74 |
|
| 75 |
Each folder contains model-specific artifacts (see below).
|
| 76 |
|
|
@@ -119,14 +119,24 @@ If you use this dataset, please cite:
|
|
| 119 |
|
| 120 |
```bibtex
|
| 121 |
@inproceedings{hoyle-etal-2025-proxann,
|
| 122 |
-
|
| 123 |
-
|
| 124 |
-
|
| 125 |
-
|
| 126 |
-
|
| 127 |
-
|
| 128 |
-
|
| 129 |
-
|
| 130 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
}
|
| 132 |
```
|
|
|
|
| 70 |
├── bertopic/
|
| 71 |
├── ctm/
|
| 72 |
└── mallet/
|
| 73 |
+
```
|
| 74 |
|
| 75 |
Each folder contains model-specific artifacts (see below).
|
| 76 |
|
|
|
|
| 119 |
|
| 120 |
```bibtex
|
| 121 |
@inproceedings{hoyle-etal-2025-proxann,
|
| 122 |
+
title = "{P}rox{A}nn: Use-Oriented Evaluations of Topic Models and Document Clustering",
|
| 123 |
+
author = "Hoyle, Alexander Miserlis and
|
| 124 |
+
Calvo-Bartolom{\'e}, Lorena and
|
| 125 |
+
Boyd-Graber, Jordan Lee and
|
| 126 |
+
Resnik, Philip",
|
| 127 |
+
editor = "Che, Wanxiang and
|
| 128 |
+
Nabende, Joyce and
|
| 129 |
+
Shutova, Ekaterina and
|
| 130 |
+
Pilehvar, Mohammad Taher",
|
| 131 |
+
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
|
| 132 |
+
month = jul,
|
| 133 |
+
year = "2025",
|
| 134 |
+
address = "Vienna, Austria",
|
| 135 |
+
publisher = "Association for Computational Linguistics",
|
| 136 |
+
url = "https://aclanthology.org/2025.acl-long.772/",
|
| 137 |
+
doi = "10.18653/v1/2025.acl-long.772",
|
| 138 |
+
pages = "15872--15897",
|
| 139 |
+
ISBN = "979-8-89176-251-0",
|
| 140 |
+
abstract = "Topic models and document-clustering evaluations either use automated metrics that align poorly with human preferences, or require expert labels that are intractable to scale. We design a scalable human evaluation protocol and a corresponding automated approximation that reflect practitioners' real-world usage of models. Annotators{---}or an LLM-based proxy{---}review text items assigned to a topic or cluster, infer a category for the group, then apply that category to other documents. Using this protocol, we collect extensive crowdworker annotations of outputs from a diverse set of topic models on two datasets. We then use these annotations to validate automated proxies, finding that the best LLM proxy is statistically indistinguishable from a human annotator and can therefore serve as a reasonable substitute in automated evaluations."
|
| 141 |
}
|
| 142 |
```
|