Update README.md
Browse files
README.md
CHANGED
|
@@ -226,17 +226,22 @@ Curation is managed by our [data manager](https://tudatalib.ulb.tu-darmstadt.de/
|
|
| 226 |
Please cite this data using:
|
| 227 |
|
| 228 |
```
|
| 229 |
-
@inproceedings{rizvi-2024-sparc,
|
| 230 |
-
|
| 231 |
-
|
| 232 |
-
|
| 233 |
-
|
| 234 |
-
|
| 235 |
-
|
| 236 |
-
|
| 237 |
-
|
| 238 |
-
|
| 239 |
-
|
| 240 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 241 |
}
|
| 242 |
```
|
|
|
|
| 226 |
Please cite this data using:
|
| 227 |
|
| 228 |
```
|
| 229 |
+
@inproceedings{rizvi-etal-2024-sparc,
|
| 230 |
+
title = "{S}pa{RC} and {S}pa{RP}: Spatial Reasoning Characterization and Path Generation for Understanding Spatial Reasoning Capability of Large Language Models",
|
| 231 |
+
author = "Rizvi, Md Imbesat and
|
| 232 |
+
Zhu, Xiaodan and
|
| 233 |
+
Gurevych, Iryna",
|
| 234 |
+
editor = "Ku, Lun-Wei and
|
| 235 |
+
Martins, Andre and
|
| 236 |
+
Srikumar, Vivek",
|
| 237 |
+
booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
|
| 238 |
+
month = aug,
|
| 239 |
+
year = "2024",
|
| 240 |
+
address = "Bangkok, Thailand",
|
| 241 |
+
publisher = "Association for Computational Linguistics",
|
| 242 |
+
url = "https://aclanthology.org/2024.acl-long.261/",
|
| 243 |
+
doi = "10.18653/v1/2024.acl-long.261",
|
| 244 |
+
pages = "4750--4767",
|
| 245 |
+
abstract = "Spatial reasoning is a crucial component of both biological and artificial intelligence. In this work, we present a comprehensive study of the capability of current state-of-the-art large language models (LLMs) on spatial reasoning. To support our study, we created and contribute a novel Spatial Reasoning Characterization (SpaRC) framework and Spatial Reasoning Paths (SpaRP) datasets, to enable an in-depth understanding of the spatial relations and compositions as well as the usefulness of spatial reasoning chains. We found that all the state-of-the-art LLMs do not perform well on the datasets{---}their performances are consistently low across different setups. The spatial reasoning capability improves substantially as model sizes scale up. Finetuning both large language models (e.g., Llama-2-70B) and smaller ones (e.g., Llama-2-13B) can significantly improve their F1-scores by 7{--}32 absolute points. We also found that the top proprietary LLMs still significantly outperform their open-source counterparts in topological spatial understanding and reasoning."
|
| 246 |
}
|
| 247 |
```
|