File size: 3,867 Bytes
d862b72
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
---
license: apache-2.0
task_categories:
  - text-generation
  - text2text-generation
language:
  - en
  - ur
  - am
  - zh
tags:
  - code
  - multilingual
  - legesher
  - transpilation
  - tiny-aya-expedition
  - language-decoded
pretty_name: Language Decoded Data
size_categories:
  - 10K<n<100K
---

# Language Decoded | Multilingual Code Dataset

Multilingual Python code datasets for the **Language Decoded** project (part of Cohere's Tiny Aya Expedition), investigating whether code's reasoning benefit for language models is **language-dependent** or **structure-dependent**.

## Research Question

> Does fine-tuning on non-English code (Python with translated keywords) improve multilingual reasoning as much as English code does?

Prior work ([Aryabumi et al., 2024](https://arxiv.org/abs/2408.10914)) showed English code improves English reasoning by 8.2%, but never tested non-English code. This dataset enables that experiment.

## Dataset Structure

This repo contains multiple experimental conditions as subdirectories:

| Subdirectory | Condition | Description |
|---|---|---|
| `source-python/` | Source | Filtered Python files from The Stack (shared base) |
| `baseline/` | Condition 1 | No code augmentation (control) |
| `english-code/` | Condition 2 | Original English-keyword Python code |
| `multilingual-code-ur/` | Condition 3a | Python transpiled to Urdu keywords via Legesher |
| `multilingual-code-am/` | Condition 3b | Python transpiled to Amharic keywords via Legesher |
| `multilingual-code-zh/` | Condition 3c | Python transpiled to Chinese keywords via Legesher |
| `multilingual-text/` | Condition 4 | Non-code multilingual text (control) |

## Usage

```python
from datasets import load_dataset

# Load a specific condition
ds = load_dataset("Legesher/language-decoded-data", data_dir="multilingual-code-ur")
```

## Transpilation

Code translation is performed using [Legesher](https://github.com/Legesher/legesher), which translates Python reserved words (keywords, builtins, exceptions) into target languages while preserving code structure and semantics.

Example (English → Chinese):

```python
# English
for item in range(10):
    if item > 5:
        print(item)

# Chinese / 中文 (via Legesher)
循环 元素 在 范围(10):
    如果 元素 > 5:
        打印(元素)
```

## Source Data

- **Base**: [The Stack](https://huggingface.co/datasets/bigcode/the-stack-dedup) — permissively licensed Python subset
- **Filtering**: Quality-filtered to 50K-100K files
- **Transpilation tool**: [Legesher v0.6.0+](https://github.com/Legesher/legesher)

## Evaluation Benchmarks

Models fine-tuned on these conditions are evaluated on:

- **XNLI** — Cross-lingual natural language inference (15 languages)
- **XStoryCloze** — Story completion (11 languages)
- **TyDi QA** — Question answering (11 languages)
- **MMLU** — Multilingual knowledge

## Related Resources

- **Models**: [Legesher/language-decoded-lora](https://huggingface.co/Legesher/language-decoded-lora) — LoRA adapters trained on these conditions
- **Community code**: [Legesher/language-decoded-community](https://huggingface.co/datasets/Legesher/language-decoded-community) — Human-written native language code
- **Experiments**: [Legesher/language-decoded-experiments](https://huggingface.co/datasets/Legesher/language-decoded-experiments) — Training logs and eval results
- **Paper**: Coming soon

## Citation

```bibtex
@misc{language-decoded-2026,
  title={Language Decoded: Investigating Language-Dependent vs. Structure-Dependent Reasoning Benefits of Code},
  author={Madison Edgar and Saad Bazaz and Rafay Mustafa and Sarah Jawaid and Rashik Shahjahan and Khojasteh Mirza and Sohaib Bazaz},
  year={2026},
  publisher={Hugging Face},
  url={https://huggingface.co/datasets/Legesher/language-decoded-data}
}
```

## License

Apache 2.0