Dataset Viewer
Auto-converted to Parquet Duplicate
translation_status
stringclasses
1 value
source_code
stringlengths
250
288k
plaincode_en
stringlengths
129
518k
plaincode_es
stringlengths
144
640k
plaincode_fr
stringlengths
138
597k
roundtrip_python_from_en
stringlengths
90
285k
roundtrip_ast_ok
bool
1 class
source_compile_ok
bool
1 class
roundtrip_compile_ok
bool
1 class
source_metadata
dict
hashes
dict
row_authority
stringclasses
1 value
mandarin_source
stringclasses
2 values
plaincode_zh
stringlengths
92
373k
roundtrip_python_from_zh
stringlengths
250
288k
roundtrip_exact_ok_zh
bool
1 class
roundtrip_ast_ok_zh
bool
1 class
roundtrip_compile_ok_zh
bool
1 class
roundtrip_error_zh
null
roundtrip_sha256_from_zh
stringlengths
64
64
roundtrip_char_count_from_zh
int64
250
288k
roundtrip_byte_count_from_zh
int64
250
288k
v2_row_index
int64
0
99.7k
v2_engine_zip_name
stringclasses
3 values
v2_engine_mode
stringclasses
2 values
v2_processed_at_utc
stringlengths
20
32
v2_required_exact_languages
stringclasses
1 value
v2_failed_required_exact_languages
stringclasses
1 value
v2_all_required_exact_ok
bool
1 class
v2_source_sha256
stringlengths
64
64
_v2_match_debug
dict
ok
import env_check from configparser import ConfigParser from func import * import warnings import sys import os import re warnings.filterwarnings('ignore') def sys_path(): path = './phantomjs/bin/' if sys.platform.startswith('win'): return path + 'phantomjs.exe' elif sys.platform.startswith('linux'...
Load env_check. Load ConfigParser from configparser. Load everything from func. Load warnings. Load sys. Load os. Load re. Call warnings dot filterwarnings with "ignore". Define function sys_path: Set path to "./phantomjs/bin/". If sys dot platform dot startswith with "win": Return path plus "phantomjs....
Importar env_check. Importar ConfigParser desde configparser. Importar todo desde func. Importar warnings. Importar sys. Importar os. Importar re. Llamar warnings punto filterwarnings con "ignore". Definir función sys_path: Establecer path como "./phantomjs/bin/". Si sys punto platform punto startswith con "win...
Charger env_check. Charger ConfigParser depuis configparser. Charger tout depuis func. Charger warnings. Charger sys. Charger os. Charger re. Appeler warnings point de filterwarnings avec "ignore". Définir fonction sys_path: Affecter path à "./phantomjs/bin/". Si sys point de platform point de startswith avec "...
import env_check from configparser import ConfigParser from func import * import warnings import sys import os import re warnings.filterwarnings('ignore') def sys_path(): path = './phantomjs/bin/' if sys.platform.startswith('win'): return path + 'phantomjs.exe' elif sys.platform.startswith('linux')...
true
true
true
{ "artifact_hash": "5e48d96c6899566a7e65fc87c124518628a85b347ae9db2d0d11eb96919080b4", "id": "1737185", "max_stars_count": 1, "max_stars_repo_name": "DavidClarence/PKUAutoSubmit", "max_stars_repo_path": "main.py", "normalized_source_hash": "8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac", ...
{ "artifact_hash": "5e48d96c6899566a7e65fc87c124518628a85b347ae9db2d0d11eb96919080b4", "normalized_source_hash": "8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac", "raw_source_hash": "d2a39c7b9d6c391194dc5b07c3a3e94ed05b12256f00cabc179ac1b6eb741d9e", "runtime_signature_pair_hash": null, "sour...
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
导入 env_check. 导入 ConfigParser 从 configparser. 从以下内容导入全部 func. 导入 warnings. 导入 sys. 导入 os. 导入 re. 调用 warnings 点 filterwarnings 用 "ignore". 定义函数 sys_path: 设 path 为 "./phantomjs/bin/". 如果 sys 点 platform 点 startswith 用 "win": 返回 path 加上 "phantomjs.exe". 否则如果 sys 点 platform 点 startswith 用 "linux": ...
import env_check from configparser import ConfigParser from func import * import warnings import sys import os import re warnings.filterwarnings('ignore') def sys_path(): path = './phantomjs/bin/' if sys.platform.startswith('win'): return path + 'phantomjs.exe' elif sys.platform.startswith('linux'...
true
true
true
null
8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac
1,787
1,889
0
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.227723+00:00
en,es,fr,zh
true
8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac
{ "v2_row_index": 0, "v2_source_sha256": "8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac" }
ok
"""Manifold test invariants.""" import tensorflow as tf import numpy as np def random_constant(shape, dtype): return tf.constant( np.random.uniform(size=shape, high=1e-1), dtype=dtype.as_numpy_dtype, ) class TestInvariants(tf.test.TestCase): def check_random(self, manifold, shape, dtype)...
Evaluate "Manifold test invariants.". Load tensorflow, referred to as tf. Load numpy, referred to as np. Define function random_constant with parameters shape, dtype: Return tf dot constant with (np dot random dot uniform with size set to shape and high set to 0.1) and dtype set to dtype dot as_numpy_dtype. Define ...
Evaluar "Manifold test invariants.". Importar tensorflow, referido como tf. Importar numpy, referido como np. Definir función random_constant con parámetros shape, dtype: Devolver tf punto constant con (np punto random punto uniform con size establecido a shape y también high establecido a 0.1) y también dtype esta...
Évaluer "Manifold test invariants.". Charger tensorflow, référé comme tf. Charger numpy, référé comme np. Définir fonction random_constant avec paramètres shape, dtype: Retourner tf point de constant avec (np point de random point de uniform avec size défini à shape et high défini à 0.1) et dtype défini à dtype poi...
"""Manifold test invariants.""" import tensorflow as tf import numpy as np def random_constant(shape, dtype): return tf.constant(np.random.uniform(size=shape, high=0.1), dtype=dtype.as_numpy_dtype) class TestInvariants(tf.test.TestCase): def check_random(self, manifold, shape, dtype): """Check random...
true
true
true
{ "artifact_hash": "d28d8be1fae8c3fdb4f55fcf5a5362581b45a8e8b9c4a8735ef6d5a99b1a991a", "id": "1737186", "max_stars_count": 33, "max_stars_repo_name": "vishalbelsare/tensorflow-riemopt", "max_stars_repo_path": "tensorflow_riemopt/manifolds/test_invariants.py", "normalized_source_hash": "72bd7b31f2a751b3e0ca7...
{ "artifact_hash": "d28d8be1fae8c3fdb4f55fcf5a5362581b45a8e8b9c4a8735ef6d5a99b1a991a", "normalized_source_hash": "72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad", "raw_source_hash": "959338f7e029fb97faba7e3548bed444306cf907a534649f9c9347d45ce345e4", "runtime_signature_pair_hash": null, "sour...
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
求值 "Manifold test invariants.". 导入 tensorflow, 别名为 tf. 导入 numpy, 别名为 np. 定义函数 random_constant 参数为 shape, dtype: 返回 tf 点 constant 用 (np 点 random 点 uniform 用 size 设为 shape 并且 high 设为 0.1) 并且 dtype 设为 dtype 点 as_numpy_dtype. 定义类 TestInvariants 继承自 tf 点 test 点 TestCase: 定义方法 check_random 参数为 self, manifold, shape, ...
"""Manifold test invariants.""" import tensorflow as tf import numpy as np def random_constant(shape, dtype): return tf.constant( np.random.uniform(size=shape, high=1e-1), dtype=dtype.as_numpy_dtype, ) class TestInvariants(tf.test.TestCase): def check_random(self, manifold, shape, dtype)...
true
true
true
null
72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad
8,672
8,672
1
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.242638+00:00
en,es,fr,zh
true
72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad
{ "v2_row_index": 1, "v2_source_sha256": "72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad" }
ok
import pkgutil import InputDataProcessing from pcapng import FileScanner def get_pcap_description(input_file, default="<not provided>"): """ Extracts the error description from the PCAPng file (comment section). :param input_file: Path to the PCAP file. :param default: When no comment is found, this v...
Load pkgutil. Load InputDataProcessing. Load FileScanner from pcapng. Define function get_pcap_description with parameters input_file, default (default: "<not provided>"): Text block: "" " Extracts the error description from the PCAPng file (comment section)." " :param input_file: Path to the PCAP...
Importar pkgutil. Importar InputDataProcessing. Importar FileScanner desde pcapng. Definir función get_pcap_description con parámetros input_file, default (predeterminado: "<not provided>"): Texto literal: "" " Extracts the error description from the PCAPng file (comment section)." " :param input_...
Charger pkgutil. Charger InputDataProcessing. Charger FileScanner depuis pcapng. Définir fonction get_pcap_description avec paramètres input_file, default (par défaut: "<not provided>"): Texte littéral: "" " Extracts the error description from the PCAPng file (comment section)." " :param input_fil...
import pkgutil import InputDataProcessing from pcapng import FileScanner def get_pcap_description(input_file, default='<not provided>'): """ Extracts the error description from the PCAPng file (comment section). :param input_file: Path to the PCAP file. :param default: When no comment is found, this va...
true
true
true
{ "artifact_hash": "38d2f2a11a9ea328d06b762d0454fb6d1ffb479d30c032f12cc1032dd1262b03", "id": "1737187", "max_stars_count": 0, "max_stars_repo_name": "marhoSVK/semiauto-diagnostics", "max_stars_repo_path": "Action_extend.py", "normalized_source_hash": "3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312...
{ "artifact_hash": "38d2f2a11a9ea328d06b762d0454fb6d1ffb479d30c032f12cc1032dd1262b03", "normalized_source_hash": "3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3", "raw_source_hash": "3a2498c2d144053ee28a3d1aabd43ac023ccc0acf346922016c4afe5365b3360", "runtime_signature_pair_hash": null, "sour...
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
导入 pkgutil. 导入 InputDataProcessing. 导入 FileScanner 从 pcapng. 定义函数 get_pcap_description 参数为 input_file, default (默认: "<not provided>"): 文本块: "" " Extracts the error description from the PCAPng file (comment section)." " :param input_file: Path to the PCAP file." " :param default: When no com...
import pkgutil import InputDataProcessing from pcapng import FileScanner def get_pcap_description(input_file, default="<not provided>"): """ Extracts the error description from the PCAPng file (comment section). :param input_file: Path to the PCAP file. :param default: When no comment is found, this v...
true
true
true
null
3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3
1,853
1,853
2
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.242943+00:00
en,es,fr,zh
true
3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3
{ "v2_row_index": 2, "v2_source_sha256": "3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3" }
ok
# Generated by Django 2.1.1 on 2018-10-04 04:56 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='PlaceWeather', fields=[ ('id', models.AutoF...
# Generated by Django 2.1.1 on 2018-10-04 04:56 Load migrations, models from django.db. Define class Migration inheriting from migrations.Migration: Set initial to True. Set dependencies to an empty list. Set operations to the list [ (migrations dot CreateModel with name set to "PlaceWeather" and fields...
# Generated por Django 2.1.1 on 2018-10-04 04:56 Importar migrations, models desde django.db. Definir clase Migration heredando de migrations.Migration: Establecer initial como True. Establecer dependencies como una lista vacía. Establecer operations como la lista [ (migrations punto CreateModel con nam...
# Generated par Django 2.1.1 on 2018-10-04 04:56 Charger migrations, models depuis django.db. Définir classe Migration héritant de migrations.Migration: Affecter initial à True. Affecter dependencies à une liste vide. Affecter operations à la liste [ (migrations point de CreateModel avec name défini à "...
# Generated by Django 2.1.1 on 2018-10-04 04:56 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [] operations = [migrations.CreateModel(name='PlaceWeather', fields=[('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, ve...
true
true
true
{ "artifact_hash": "64ab55b0cab1d78f880906d127d8234816557e33ae62a4309c9dcdad188b8615", "id": "1737188", "max_stars_count": 2, "max_stars_repo_name": "Escapist-007/Dockerized-Pub-Sub-Distributed-System", "max_stars_repo_path": "phase_01/MyWayPoints_v1/MyWayPoints/WayPoints/migrations/0001_initial.py", "norma...
{ "artifact_hash": "64ab55b0cab1d78f880906d127d8234816557e33ae62a4309c9dcdad188b8615", "normalized_source_hash": "295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa", "raw_source_hash": "93c0dea6f5372adf083d2ef522f2fac461441d7a7968dd90570c119f70996a24", "runtime_signature_pair_hash": null, "sour...
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
# Generated 按 Django 2.1.1 on 2018-10-04 04:56 导入 migrations, models 从 django 点 db. 定义类 Migration 继承自 migrations 点 Migration: 设 initial 为 True. 设 dependencies 为 空列表. 设 operations 为 列表 [ (migrations 点 CreateModel 用 name 设为 "PlaceWeather" 并且 fields 设为 列表 [ 元组 ("id", (models 点 AutoField 用 auto_crea...
# Generated by Django 2.1.1 on 2018-10-04 04:56 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='PlaceWeather', fields=[ ('id', models.AutoF...
true
true
true
null
295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa
1,077
1,077
3
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.249990+00:00
en,es,fr,zh
true
295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa
{ "v2_row_index": 3, "v2_source_sha256": "295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa" }
ok
"import os\nimport psutil\nimport traceback\n\nimport h5py\nimport numpy as np\n\nfrom buffalo.data (...TRUNCATED)
"Load os.\nLoad psutil.\nLoad traceback.\nLoad h5py.\nLoad numpy, referred to as np.\nLoad prepro fr(...TRUNCATED)
"Importar os.\nImportar psutil.\nImportar traceback.\nImportar h5py.\nImportar numpy, referido como (...TRUNCATED)
"Charger os.\nCharger psutil.\nCharger traceback.\nCharger h5py.\nCharger numpy, référé comme np.(...TRUNCATED)
"import os\nimport psutil\nimport traceback\nimport h5py\nimport numpy as np\nfrom buffalo.data impo(...TRUNCATED)
true
true
true
{"artifact_hash":"01ba4cd5e9a264b6e85cb6b55349843e9c07a0714fe6dbbb9d5b0c822779dab6","id":"1737191","(...TRUNCATED)
{"artifact_hash":"01ba4cd5e9a264b6e85cb6b55349843e9c07a0714fe6dbbb9d5b0c822779dab6","normalized_sour(...TRUNCATED)
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
"导入 os.\n导入 psutil.\n导入 traceback.\n导入 h5py.\n导入 numpy, 别名为 np.\n导入 pr(...TRUNCATED)
"import os\nimport psutil\nimport traceback\n\nimport h5py\nimport numpy as np\n\nfrom buffalo.data (...TRUNCATED)
true
true
true
null
9cc856fe97656efa950426a1ff2351f0adc075353cf35a5c128fd1b2bae57a72
9,783
9,783
4
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.262630+00:00
en,es,fr,zh
true
9cc856fe97656efa950426a1ff2351f0adc075353cf35a5c128fd1b2bae57a72
{"v2_row_index":4,"v2_source_sha256":"9cc856fe97656efa950426a1ff2351f0adc075353cf35a5c128fd1b2bae57a(...TRUNCATED)
ok
"import pandas\n\ndef add_full_name(path_to_csv, path_to_new_csv):\n #Assume you will be reading (...TRUNCATED)
"Load pandas.\nDefine function add_full_name with parameters path_to_csv, path_to_new_csv:\n # As(...TRUNCATED)
"Importar pandas.\nDefinir función add_full_name con parámetros path_to_csv, path_to_new_csv:\n (...TRUNCATED)
"Charger pandas.\nDéfinir fonction add_full_name avec paramètres path_to_csv, path_to_new_csv:\n (...TRUNCATED)
"import pandas\n\ndef add_full_name(path_to_csv, path_to_new_csv):\n df = pandas.read_csv(path_to(...TRUNCATED)
true
true
true
{"artifact_hash":"41bb7ad785ab2359e746b190a2599bcc0844387bad71f17bc8616e18c8fc3713","id":"1737192","(...TRUNCATED)
{"artifact_hash":"41bb7ad785ab2359e746b190a2599bcc0844387bad71f17bc8616e18c8fc3713","normalized_sour(...TRUNCATED)
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
"导入 pandas.\n定义函数 add_full_name 参数为 path_to_csv, path_to_new_csv:\n # Assume yo(...TRUNCATED)
"import pandas\n\ndef add_full_name(path_to_csv, path_to_new_csv):\n #Assume you will be reading (...TRUNCATED)
true
true
true
null
0124385b1fb972601378d7774926eba679cfad99642aab1ad9b367d1e7963e85
1,164
1,164
5
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.269981+00:00
en,es,fr,zh
true
0124385b1fb972601378d7774926eba679cfad99642aab1ad9b367d1e7963e85
{"v2_row_index":5,"v2_source_sha256":"0124385b1fb972601378d7774926eba679cfad99642aab1ad9b367d1e7963e(...TRUNCATED)
ok
"import asyncio\nfrom timeit import default_timer\n\nimport aiohttp\n\nimport settings\n\n\nasync de(...TRUNCATED)
"Load asyncio.\nLoad default_timer from timeit.\nLoad aiohttp.\nLoad settings.\nDefine async functio(...TRUNCATED)
"Importar asyncio.\nImportar default_timer desde timeit.\nImportar aiohttp.\nImportar settings.\nDef(...TRUNCATED)
"Charger asyncio.\nCharger default_timer depuis timeit.\nCharger aiohttp.\nCharger settings.\nDéfin(...TRUNCATED)
"import asyncio\nfrom timeit import default_timer\nimport aiohttp\nimport settings\n\nasync def fetc(...TRUNCATED)
true
true
true
{"artifact_hash":"6daa8d4e8cc2efc0e5d24dfba163cea60806e96504aee6b317a7c83beda15305","id":"1737193","(...TRUNCATED)
{"artifact_hash":"6daa8d4e8cc2efc0e5d24dfba163cea60806e96504aee6b317a7c83beda15305","normalized_sour(...TRUNCATED)
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
"导入 asyncio.\n导入 default_timer 从 timeit.\n导入 aiohttp.\n导入 settings.\n定义异步(...TRUNCATED)
"import asyncio\nfrom timeit import default_timer\n\nimport aiohttp\n\nimport settings\n\n\nasync de(...TRUNCATED)
true
true
true
null
21bdcfe8905af1779309ffb67760203d14f6708b31fa8e76b24cc022510a110c
860
860
6
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.291498+00:00
en,es,fr,zh
true
21bdcfe8905af1779309ffb67760203d14f6708b31fa8e76b24cc022510a110c
{"v2_row_index":6,"v2_source_sha256":"21bdcfe8905af1779309ffb67760203d14f6708b31fa8e76b24cc022510a11(...TRUNCATED)
ok
"import asyncio\nimport asyncssh\nimport sys\nimport os\nimport crypt\nfrom importlib.util import fi(...TRUNCATED)
"Load asyncio.\nLoad asyncssh.\nLoad sys.\nLoad os.\nLoad crypt.\nLoad find_spec from importlib.util(...TRUNCATED)
"Importar asyncio.\nImportar asyncssh.\nImportar sys.\nImportar os.\nImportar crypt.\nImportar find_(...TRUNCATED)
"Charger asyncio.\nCharger asyncssh.\nCharger sys.\nCharger os.\nCharger crypt.\nCharger find_spec d(...TRUNCATED)
"import asyncio\nimport asyncssh\nimport sys\nimport os\nimport crypt\nfrom importlib.util import fi(...TRUNCATED)
true
true
true
{"artifact_hash":"b2da5200e43bdd1db4d50cad36628e48ec81bbab93e64bbbada9c60a77acfb77","id":"1737194","(...TRUNCATED)
{"artifact_hash":"b2da5200e43bdd1db4d50cad36628e48ec81bbab93e64bbbada9c60a77acfb77","normalized_sour(...TRUNCATED)
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
"导入 asyncio.\n导入 asyncssh.\n导入 sys.\n导入 os.\n导入 crypt.\n导入 find_spec 从 imp(...TRUNCATED)
"import asyncio\nimport asyncssh\nimport sys\nimport os\nimport crypt\nfrom importlib.util import fi(...TRUNCATED)
true
true
true
null
77f05f17795d4c289aa24fe3adebbe90335078b286eb5c6c371543becb4cb01e
7,230
7,230
7
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.292076+00:00
en,es,fr,zh
true
77f05f17795d4c289aa24fe3adebbe90335078b286eb5c6c371543becb4cb01e
{"v2_row_index":7,"v2_source_sha256":"77f05f17795d4c289aa24fe3adebbe90335078b286eb5c6c371543becb4cb0(...TRUNCATED)
ok
"# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED)
"# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED)
"# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED)
"# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED)
"# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED)
true
true
true
{"artifact_hash":"d10bb428d0640745d72f8ac0a508b93a9167732eee9ba5006754e93fb2f7f083","id":"1737196","(...TRUNCATED)
{"artifact_hash":"d10bb428d0640745d72f8ac0a508b93a9167732eee9ba5006754e93fb2f7f083","normalized_sour(...TRUNCATED)
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
"# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED)
"# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED)
true
true
true
null
7edf844d2d0e2f6a3bc93e3e4f18c59c4c1e245a810f771da91da140a7705e81
1,149
1,691
8
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.497595+00:00
en,es,fr,zh
true
7edf844d2d0e2f6a3bc93e3e4f18c59c4c1e245a810f771da91da140a7705e81
{"v2_row_index":8,"v2_source_sha256":"7edf844d2d0e2f6a3bc93e3e4f18c59c4c1e245a810f771da91da140a7705e(...TRUNCATED)
ok
"# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED)
"# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED)
"# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED)
"# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED)
"# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED)
true
true
true
{"artifact_hash":"e5a8cc0051f26778bcc126c2932aef102e63895a68dcd25e8ef7279cb3a6729b","id":"1737198","(...TRUNCATED)
{"artifact_hash":"e5a8cc0051f26778bcc126c2932aef102e63895a68dcd25e8ef7279cb3a6729b","normalized_sour(...TRUNCATED)
cuarzo-100k-v1
cuarzo-100k-v2-cleaned
"# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED)
"# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED)
true
true
true
null
acf91aabb7629720da6b5758cbc8e7f001bad1d3b9828e9f812975784d8b8db2
2,425
2,425
9
PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip
sharded_spawn
2026-05-07T09:45:52.524888+00:00
en,es,fr,zh
true
acf91aabb7629720da6b5758cbc8e7f001bad1d3b9828e9f812975784d8b8db2
{"v2_row_index":9,"v2_source_sha256":"acf91aabb7629720da6b5758cbc8e7f001bad1d3b9828e9f812975784d8b8d(...TRUNCATED)
End of preview. Expand in Data Studio

Cuarzo-100K v2

99,683 bidirectional deterministically verified Python to human language pairs across English, Spanish, French, and Mandarin Chinese. Released by Cuarzo AI.

This is the second public release from Aether, our proprietary engine for generating deterministic paired data between code and human language. v2 delivers Mandarin Chinese as a full fourth language surface, expands the verification schema to per-language roundtrip checks across all four surfaces, and adds complete provenance instrumentation at the record level.

The Aether engine is under active development. More source languages, more human languages, larger volumes, and tighter verification are all in progress.

What's new in v2

  • Mandarin Chinese added. Every record now includes a plaincode_zh surface generated directly from the AST — not translated from English. The engine handles advanced Python constructs including async def, await, context managers, and decorator patterns.
  • Per-language verification. v2 expands roundtrip validation to all four language surfaces independently. Every record carries separate AST equality, compilation, and exact-match checks for EN, ES, FR, and ZH.
  • Improved human language phrasing for Spanish and French. The ES and FR representations in v2 reflect refinements to Aether's structured grammar for both languages.
  • Full provenance instrumentation. Each record includes processing timestamps, engine build identifiers, per-surface SHA256 hashes, and byte/character counts — enabling end-to-end independent auditability.
  • 100% acceptance across all checks. All 99,683 records pass every verification gate across all four language surfaces.

What makes this different

Most code-language datasets fall into one of three buckets:

  • Scraped from the web (docstrings, READMEs) — noisy, inconsistent, frequently wrong.
  • Generated by LLMs — fast and cheap, but probabilistic and unverifiable.
  • Hand-annotated by humans — accurate but slow, expensive, and structurally English-dominant.

Cuarzo takes a different path. Feed in a supported source language, and our Aether engine returns the functionally equivalent intent in another supported language — not a loose summary, not an approximate rewriting, but a verifiable, bidirectional, deterministically controlled alignment across surfaces. The same input always produces the same output. Every pair has to meet strict acceptance standards and is checked structurally before acceptance.

The pipeline

Cuarzo-100K was curated from StarCoderData's Python stream using a multi-stage acceptance pipeline:

  1. Hard filters on source quality, structure, and licensing.
  2. Multi-stage deduplication — both raw and structural — within the corpus, our own internal datasets, and all future releases.
  3. Compile checks on source to ensure parseability.
  4. Aether translation to structured human language representations in EN, ES, FR, and ZH.
  5. Roundtrip validation per language — regenerate Python from each human language representation, then verify functional equivalence to the source independently for each surface.
  6. Strict acceptance gates — only records that pass every check across all four surfaces make it into the dataset.

What's in each record

Core content

Field Description
source_code Original Python from StarCoderData
plaincode_en Aether's structured English representation
plaincode_es Aether's structured Spanish representation
plaincode_fr Aether's structured French representation
plaincode_zh Aether's structured Mandarin Chinese representation
translation_status Pipeline status flag (ok for all accepted records)

Verification — per language

Field Description
roundtrip_python_from_en/es/fr/zh Python regenerated from each language representation
roundtrip_ast_ok Boolean: source AST matches roundtrip AST (aggregate across all surfaces)
roundtrip_ast_ok_en/es/fr/zh Boolean: AST equality check per language surface
roundtrip_compile_ok Boolean: all roundtrips compile (aggregate)
roundtrip_compile_ok_en/es/fr/zh Boolean: compilation check per language roundtrip
roundtrip_exact_ok_en/es/fr/zh Boolean: exact structural match per language roundtrip
roundtrip_error_en/es/fr/zh Error message if roundtrip failed; null for all accepted records
source_compile_ok Boolean: original source compiles
source_compile_error Compile error if source failed; null for all accepted records
all_required_exact_ok Boolean: all four language surfaces passed all verification gates
required_exact_languages Languages required to pass for acceptance (value: en,es,fr,zh for all records)
failed_required_exact_languages Languages that failed, if any; null for all accepted records
failure_reasons Failure description if any gate failed; null for all accepted records

Provenance & hashes

Field Description
source_sha256 SHA256 of the original source file
roundtrip_sha256_from_en/es/fr/zh SHA256 of each roundtrip Python output
hashes Full provenance hash bundle (artifact hash, normalized source hash)
artifact_hash Hash of the Aether engine configuration used to generate this record
source_metadata Source repo name, file path, GitHub stars

Size metrics

Field Description
source_byte_count Byte length of source
source_char_count Character length of source
roundtrip_byte_count_from_en/es/fr/zh Byte length of each roundtrip Python output
roundtrip_char_count_from_en/es/fr/zh Character length of each roundtrip Python output

Composition

Verification

All 99,683 records pass every check across all four language surfaces.

Check Records passing
source_compile_ok 99,683 / 99,683 (100.00%)
roundtrip_compile_ok (aggregate) 99,683 / 99,683 (100.00%)
roundtrip_compile_ok_en 99,683 / 99,683 (100.00%)
roundtrip_compile_ok_es 99,683 / 99,683 (100.00%)
roundtrip_compile_ok_fr 99,683 / 99,683 (100.00%)
roundtrip_compile_ok_zh 99,683 / 99,683 (100.00%)
roundtrip_ast_ok (aggregate) 99,683 / 99,683 (100.00%)
roundtrip_ast_ok_en 99,683 / 99,683 (100.00%)
roundtrip_ast_ok_es 99,683 / 99,683 (100.00%)
roundtrip_ast_ok_fr 99,683 / 99,683 (100.00%)
roundtrip_ast_ok_zh 99,683 / 99,683 (100.00%)
all_required_exact_ok 99,683 / 99,683 (100.00%)

Multilingual coverage

Language Records with non-empty representation
English (plaincode_en) 99,683 / 99,683 (100.00%)
Spanish (plaincode_es) 99,683 / 99,683 (100.00%)
French (plaincode_fr) 99,683 / 99,683 (100.00%)
Mandarin Chinese (plaincode_zh) 99,683 / 99,683 (100.00%)

Every record contains all five surfaces: Python source plus four natural-language representations, each independently verified.

Source diversity

Property Value
Source corpus StarCoderData Python stream
Unique source repositories 83,489
Average records per repository 1.2

The dataset is drawn from the long tail of public Python code rather than concentrated in a small number of large repositories. The average of 1.2 records per repository means most source files come from distinct projects; concentration is minimal.

Source code size distribution

Size (characters) Records Share
< 500 13,217 13.26%
500 – 2,000 45,648 45.79%
2,000 – 5,000 26,208 26.29%
5,000 – 10,000 9,644 9.67%
10,000 – 20,000 3,628 3.64%
> 20,000 1,347 1.35%

Source code statistics: minimum 250 chars, maximum 288,385 chars, mean 3,067 chars, median 1,534 chars.

ML and Deep Learning composition

Category Records Share
Records using any ML library 15,832 15.88%
Records using a deep learning framework 4,887 4.90%

Top libraries by record count:

Library Records
numpy 10,238
pandas 3,514
torch 2,811
tensorflow 1,570
sklearn 1,192
cv2 1,188
keras 588
nltk 276
transformers 125
spacy 104

Records may use multiple libraries; library counts can sum to more than the total number of records using any ML library.

Multilingual parity

Global enterprises do not operate in one language, but most public code-language datasets do. Cuarzo-100K v2 gives AI teams verified Python-to-language training records across English, Spanish, French, and Mandarin Chinese — with additional languages planned in future releases.

Each language surface is generated from the same verified semantic foundation — the AST — not translated from another language surface. Teams can train multilingual code understanding without sacrificing consistency, auditability, or behavioral grounding. The Mandarin surface is generated from the same verified semantic foundation as the other language surfaces, with independent per-surface roundtrip verification applied.

Example (excerpt):

Python source:

def connection(host, port, username, password):
    try:
        connection = get_connection(host=host, port=port, username=username, password=password, use_tls=True)
        return connection
    except Exception as e:
        print(e)

English representation:

Define method connection with parameters host, port, username, password:
    Attempt the following:
        Set connection to get_connection with host set to host, port set to port, ...
        Return connection.
    If an error of type Exception occurs, bound as e:
        Print e.

Spanish representation:

Definir método connection con parámetros host, port, username, password:
    Intentar lo siguiente:
        Establecer connection como get_connection con host establecido a host, port establecido a port, ...
        Devolver connection.
    Si ocurre un error de tipo Exception, como e:
        Imprimir e.

French representation:

Définir méthode connection avec paramètres host, port, username, password:
    Essayer ce qui suit:
        Affecter connection à get_connection avec host défini à host, port défini à port, ...
        Retourner connection.
    Si une erreur du type Exception survient, lié comme e:
        Afficher e.

Mandarin Chinese representation:

定义方法 connection,参数为 host、port、username、password:
    尝试以下操作:
        将 connection 设置为 get_connection,其中 host 设为 host,port 设为 port,...
        返回 connection。
    若发生 Exception 类型的错误,绑定为 e:
        打印 e。

Why we're releasing this

Curating high quality code-language data is a huge bottleneck for the AI industry today, and every known existing source forces the same tradeoff: cheap and unreliable, or accurate and unscalable without heavy expenses. The teams pushing the frontier of code generation, code understanding, and multilingual AI are all hitting the same ceiling at once.

The team at Cuarzo AI takes a fundamentally different approach: verifiable alignment between code and natural language, with functional equivalence guaranteed and verified by construction rather than estimated by sampling. The same input always produces the same output. Every pair is checked before release — and in v2, checked independently across all four language surfaces.

We're releasing this dataset so the field can build on it. Better training data, better evaluation, and better cross-lingual alignment all start with pairs that are actually consistent — and the only way to prove that is to put the data in researchers' hands and invite them to verify it. Use it. Stress-test it. Tell us what you find.

Intended uses

Cuarzo-100K v2 is designed for:

  • Training and fine-tuning code-language models with verified, multilingual pairs
  • Multilingual code generation, explanation, and translation research
  • Evaluation and benchmarking of code-language models, especially in non-English settings
  • Cross-lingual alignment research between Latin-script and CJK-script representations of code semantics
  • Provenance and reproducibility research, given the full per-record instrumentation

Permissively licensed under Apache 2.0 for commercial and non-commercial use.

Known limitations

We declare these openly so users can plan around them.

  • AST equivalence, not full runtime semantic equivalence. Verification confirms that source and roundtrip produce identical normalized ASTs. This is a strong correctness guarantee for nearly all practical purposes, but does not cover all runtime scenarios involving evaluation order of side effects or implementation-defined behavior.

  • Comments are preserved verbatim. Source code comments (#-prefixed) are treated as literal annotations and preserved across language representations rather than translated. This is intentional — comment translation introduces ambiguity around author intent and is out of scope for this release.

  • StarCoderData distribution. Source files inherit the composition of StarCoderData's Python stream. Long-tail repository content is well represented; highly popular code patterns may be relatively underrepresented compared to popularity-weighted scraped datasets.

  • ML/DL composition is minoritarian. 15.88% of records use general ML libraries; 4.90% use deep learning frameworks. Users with ML-specific needs may want to filter accordingly.

  • Python only in this release. Aether currently supports Python paired with EN/ES/FR/ZH. Additional source languages (C++, Go, Rust) and additional human languages (Hindi, Japanese, German, and beyond) are on the roadmap.

How to load

from datasets import load_dataset

ds = load_dataset("Cuarzo-AI/cuarzo-100k-v2")

# Each record contains five language surfaces plus full per-language verification
sample = ds["train"][0]
print(sample["source_code"])
print(sample["plaincode_zh"])
print(sample["roundtrip_ast_ok_zh"])  # True for all accepted records

Try it. Tell us what you find.

We invite AI labs, research teams, and code-tooling companies to fine-tune or evaluate models on Cuarzo-100K and share results. Standard code benchmarks (HumanEval, MBPP, MultiPL-E) and multilingual evaluations are particularly relevant given the dataset's structural properties. We're especially interested in head-to-head comparisons against fine-tunes on web-scraped or LLM-generated paired data, and in evaluations that specifically test non-English and CJK code understanding.

The per-record provenance fields enable a class of evaluation that isn't possible with other datasets: you can filter by engine build, by processing time, by roundtrip character count, and verify any individual record's equivalence chain independently. We'd like to see that capability used.

If you find that models trained on Cuarzo-100K produce stronger results — or weaker ones — we want to hear about it.

The Aether engine

Cuarzo-100K v2 is a snapshot of what the Aether engine produces at public-release scale. The engine itself is significantly more capable than this dataset reflects.

Volume. Aether is designed for industrial-scale generation. The pipeline that produced this dataset operates throughputs that make hundreds of millions of records commercially viable. The 99,683 records in this release represent a controlled public sample, not a ceiling. AI labs and enterprises requiring training corpora in the tens or hundreds of millions of verified pairs should contact us directly.

Language coverage. The current public release covers Python paired with EN, ES, FR, and ZH. Aether's architecture is language-agnostic at the source and target levels. Additional programming languages — including C++, Go, and Rust — and additional human languages — including Hindi, Portuguese, German — are in active development. Labs with specific language-pair requirements can work with us to prioritize and schedule production runs for their exact needs.

Domain and codebase customization. Aether can be targeted at specific domains, libraries, or internal codebases. A lab building a code model for biomedical Python, for financial analytics tooling, or for a proprietary internal framework can receive verified paired data scoped precisely to that domain — not a general-purpose corpus filtered after the fact. The engine generates from your source material with the same deterministic verification guarantees as the public dataset. For labs with proprietary codebases, Aether is available as an API: the engine runs against your code on your infrastructure, and no source code is transmitted to or stored by Cuarzo AI. You receive verified paired data without exposing your codebase.

Evaluation and benchmark data. Beyond training data, Aether can generate held-out evaluation sets with known provenance and verified correctness properties. For labs that need benchmark data with formal equivalence guarantees rather than crowd-sourced annotations, this is a meaningful structural improvement.

Custom verification thresholds. Acceptance criteria are configurable. Labs with stricter requirements — tighter AST matching, higher exact-match thresholds, or additional compile-time checks — can request production runs with those parameters applied.

Reach out at hello@cuarzoai.com to discuss volume, language coverage, and customization. We respond to every legitimate inquiry within two business days.

Citation

If you use this dataset in research or product work, please cite:

@dataset{cuarzo_100k_v2_2026,
  title  = {Cuarzo-100K: Deterministically Paired Python and Multilingual Human-Language Code Representations},
  author = {Cuarzo AI},
  year   = {2026},
  url    = {https://huggingface.co/datasets/Cuarzo-AI/cuarzo-100k-v2}
}

License

Released under the Apache License 2.0.

Source code is sourced from StarCoderData (Python stream) under its respective license. The source_metadata field preserves the originating repository name, file path, and stars count for each record. Users should review individual source repository licenses for any downstream use of specific source files.


Cuarzo-100K-v2 is a public release from the Cuarzo AI team. We work with AI labs, code-tooling companies, and multilingual AI programs that need training and evaluation data with verifiable and provable consistency.

Downloads last month
50