--- dataset_info: features: - name: id dtype: int64 - name: latin dtype: string - name: german dtype: string - name: source dtype: string - name: tag dtype: string - name: score dtype: float64 splits: - name: train num_bytes: 154924781 num_examples: 406011 download_size: 91837604 dataset_size: 154924781 configs: - config_name: default data_files: - split: train path: data/train-* license: cc-by-4.0 task_categories: - translation language: - la - de size_categories: - 100K= 1** represents a valid score. ## 📚 Citation / References Falls Sie dieses Modell in Ihrer Forschung verwenden, bitten wir Sie, die zugrundeliegende Masterarbeit wie folgt zu zitieren: **Masterarbeit (Zenodo DOI):** Wenzel, M. (2025). Translatio ex Machina: Neuronale Maschinelle Übersetzung vom Lateinischen ins Deutsche [Zenodo]. Unveröffentlichte Masterarbeit, Fachhochschule Südwestfalen DOI: **[10.5281/zenodo.17940090](https://doi.org/10.5281/zenodo.17940090)** ----- ## 💻 Usage ### Understanding the `score` Feature The **`score`** column indicates the method and quality of the sentence alignment: * **`score` = -1:** This value signifies that the Latin and German sentences were **manually aligned** or by a special tooling. * **`score` \>= 1:** This value indicates that the alignment was **calculated by an automated alignment tool**. * **Interpretation:** A higher score suggests a **better alignment quality**. * **Recommendation:** For high-confidence automated alignments, we recommend using only entries where the score is **>= 1.2**. ### Loading and Filtering the Dataset You can easily filter the dataset to select only high-quality alignments (manual alignments OR high-scoring automated alignments) using the `filter()` method: ```python from datasets import load_dataset, DatasetDict # 1. Load the initial dataset (contains only the "train" split) dataset = load_dataset("fhswf/latin-german-parallel") train_dataset = dataset["train"] # 2. Filter the dataset to include only high-quality alignments: # - Entries with score == -1 (manual alignment) # - Entries with score >= 1.2 (high-confidence automated alignment) def filter_by_score(example): return example["score"] == -1 or example["score"] >= 1.2 high_quality_train = train_dataset.filter(filter_by_score) # Optional: Proceed with splitting the high-quality data temp_splits = high_quality_train.train_test_split(test_size=0.01, seed=42) test_validation_splits = temp_splits["test"].train_test_split(test_size=0.5, seed=42) dataset = DatasetDict({ "train": temp_splits["train"], "validation": test_validation_splits["train"], "test": test_validation_splits["test"], }) ```