Datasets:

Modalities:
Tabular
Text
Formats:
parquet
Languages:
German
ArXiv:
Libraries:
Datasets
Dask
License:
LukasHug commited on
Commit
09189eb
·
verified ·
1 Parent(s): 456988d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -8
README.md CHANGED
@@ -260,19 +260,17 @@ size_categories:
260
  - **Point of Contact:** [Lukas Helff](mailto:helff@cs.tu-darmstadt.de)
261
  - **License:** [CC BY](https://creativecommons.org/licenses/by/4.0/)
262
 
263
- # SLR-Bench: Scalable Logical Reasoning Benchmark for LLMs
264
  [![Eval & Reward Model](https://img.shields.io/badge/%F0%9F%A4%96%20Reward%20Model-HF-blueviolet)](https://huggingface.co/spaces/AIML-TUDA/VerifiableRewardsForScalableLogicalReasoning)
265
  [![GitHub](https://img.shields.io/badge/Code-GitHub-blue)](https://github.com/ml-research/ScalableLogicalReasoning)
266
  [![arXiv](https://img.shields.io/badge/arXiv-2506.15787-b31b1b.svg)](https://arxiv.org/abs/2506.15787)
 
 
267
 
 
 
268
 
269
-
270
- > **🆕 August 2025: Build your own Reasoning Problems with Verifiable Rewards. Source Code is now available!** 👉 [Generate your own Reasoning Task](https://github.com/ml-research/ScalableLogicalReasoning)
271
-
272
- > **🆕 June 2024: Evaluation & RLVR Reward Model Released!** 👉 [Demo on Hugging Face Spaces](https://huggingface.co/spaces/AIML-TUDA/VerifiableRewardsForScalableLogicalReasoning)
273
-
274
- **SLR-Bench** is a scalable, fully-automated benchmark designed to systematically evaluate and train Large Language Models (LLMs) in logical reasoning via inductive logic programming (ILP) tasks. Built with the [SLR framework](https://github.com/ml-research/ScalableLogicalReasoning), SLR-Bench presents LLMs with open-ended logic problems of progressively increasing difficulty, assesses their solutions via deterministic symbolic evaluation, and supports both curriculum learning and systematic measurement of reasoning performance.
275
-
276
 
277
  ## DS Overview
278
  - **Curriculum:** 20 complexity levels, grouped into 4 broad tiers (basic, easy, medium, hard)
 
260
  - **Point of Contact:** [Lukas Helff](mailto:helff@cs.tu-darmstadt.de)
261
  - **License:** [CC BY](https://creativecommons.org/licenses/by/4.0/)
262
 
263
+ # 🧠 SLR-Bench-German: Scalable Logical Reasoning Benchmark (German Edition)
264
  [![Eval & Reward Model](https://img.shields.io/badge/%F0%9F%A4%96%20Reward%20Model-HF-blueviolet)](https://huggingface.co/spaces/AIML-TUDA/VerifiableRewardsForScalableLogicalReasoning)
265
  [![GitHub](https://img.shields.io/badge/Code-GitHub-blue)](https://github.com/ml-research/ScalableLogicalReasoning)
266
  [![arXiv](https://img.shields.io/badge/arXiv-2506.15787-b31b1b.svg)](https://arxiv.org/abs/2506.15787)
267
+ [![SLR-Bench (English)](https://img.shields.io/badge/SLR--Bench-English-orange)](https://huggingface.co/datasets/AIML-TUDA/SLR-Bench)
268
+ [![SLR-Bench-German](https://img.shields.io/badge/SLR--Bench-German-red)](https://huggingface.co/datasets/AIML-TUDA/SLR-Bench-German)
269
 
270
+ **SLR-Bench-German** is the **German-language pendant** of the original [**SLR-Bench**](https://huggingface.co/datasets/AIML-TUDA/SLR-Bench) dataset.
271
+ It follows the same symbolic structure, evaluation framework, and curriculum as the English version but provides all **natural-language task prompts translated into German**.
272
 
273
+ This enables systematic evaluation and training of Large Language Models (LLMs) in logical reasoning in german, supporting both *multilingual reasoning* and *cross-lingual generalization* research.
 
 
 
 
 
 
274
 
275
  ## DS Overview
276
  - **Curriculum:** 20 complexity levels, grouped into 4 broad tiers (basic, easy, medium, hard)