Datasets:
translation_status stringclasses 1
value | source_code stringlengths 250 288k | plaincode_en stringlengths 129 518k | plaincode_es stringlengths 144 640k | plaincode_fr stringlengths 138 597k | roundtrip_python_from_en stringlengths 90 285k | roundtrip_ast_ok bool 1
class | source_compile_ok bool 1
class | roundtrip_compile_ok bool 1
class | source_metadata dict | hashes dict | row_authority stringclasses 1
value | mandarin_source stringclasses 2
values | plaincode_zh stringlengths 92 373k ⌀ | roundtrip_python_from_zh stringlengths 250 288k ⌀ | roundtrip_exact_ok_zh bool 1
class | roundtrip_ast_ok_zh bool 1
class | roundtrip_compile_ok_zh bool 1
class | roundtrip_error_zh null | roundtrip_sha256_from_zh stringlengths 64 64 ⌀ | roundtrip_char_count_from_zh int64 250 288k ⌀ | roundtrip_byte_count_from_zh int64 250 288k ⌀ | v2_row_index int64 0 99.7k ⌀ | v2_engine_zip_name stringclasses 3
values | v2_engine_mode stringclasses 2
values | v2_processed_at_utc stringlengths 20 32 ⌀ | v2_required_exact_languages stringclasses 1
value | v2_failed_required_exact_languages stringclasses 1
value | v2_all_required_exact_ok bool 1
class | v2_source_sha256 stringlengths 64 64 ⌀ | _v2_match_debug dict |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ok | import env_check
from configparser import ConfigParser
from func import *
import warnings
import sys
import os
import re
warnings.filterwarnings('ignore')
def sys_path():
path = './phantomjs/bin/'
if sys.platform.startswith('win'):
return path + 'phantomjs.exe'
elif sys.platform.startswith('linux'... | Load env_check.
Load ConfigParser from configparser.
Load everything from func.
Load warnings.
Load sys.
Load os.
Load re.
Call warnings dot filterwarnings with "ignore".
Define function sys_path:
Set path to "./phantomjs/bin/".
If sys dot platform dot startswith with "win":
Return path plus "phantomjs.... | Importar env_check.
Importar ConfigParser desde configparser.
Importar todo desde func.
Importar warnings.
Importar sys.
Importar os.
Importar re.
Llamar warnings punto filterwarnings con "ignore".
Definir función sys_path:
Establecer path como "./phantomjs/bin/".
Si sys punto platform punto startswith con "win... | Charger env_check.
Charger ConfigParser depuis configparser.
Charger tout depuis func.
Charger warnings.
Charger sys.
Charger os.
Charger re.
Appeler warnings point de filterwarnings avec "ignore".
Définir fonction sys_path:
Affecter path à "./phantomjs/bin/".
Si sys point de platform point de startswith avec "... | import env_check
from configparser import ConfigParser
from func import *
import warnings
import sys
import os
import re
warnings.filterwarnings('ignore')
def sys_path():
path = './phantomjs/bin/'
if sys.platform.startswith('win'):
return path + 'phantomjs.exe'
elif sys.platform.startswith('linux')... | true | true | true | {
"artifact_hash": "5e48d96c6899566a7e65fc87c124518628a85b347ae9db2d0d11eb96919080b4",
"id": "1737185",
"max_stars_count": 1,
"max_stars_repo_name": "DavidClarence/PKUAutoSubmit",
"max_stars_repo_path": "main.py",
"normalized_source_hash": "8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac",
... | {
"artifact_hash": "5e48d96c6899566a7e65fc87c124518628a85b347ae9db2d0d11eb96919080b4",
"normalized_source_hash": "8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac",
"raw_source_hash": "d2a39c7b9d6c391194dc5b07c3a3e94ed05b12256f00cabc179ac1b6eb741d9e",
"runtime_signature_pair_hash": null,
"sour... | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | 导入 env_check.
导入 ConfigParser 从 configparser.
从以下内容导入全部 func.
导入 warnings.
导入 sys.
导入 os.
导入 re.
调用 warnings 点 filterwarnings 用 "ignore".
定义函数 sys_path:
设 path 为 "./phantomjs/bin/".
如果 sys 点 platform 点 startswith 用 "win":
返回 path 加上 "phantomjs.exe".
否则如果 sys 点 platform 点 startswith 用 "linux":
... | import env_check
from configparser import ConfigParser
from func import *
import warnings
import sys
import os
import re
warnings.filterwarnings('ignore')
def sys_path():
path = './phantomjs/bin/'
if sys.platform.startswith('win'):
return path + 'phantomjs.exe'
elif sys.platform.startswith('linux'... | true | true | true | null | 8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac | 1,787 | 1,889 | 0 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.227723+00:00 | en,es,fr,zh | true | 8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac | {
"v2_row_index": 0,
"v2_source_sha256": "8c5158d27997a5edf266a8b9b8725495a58a6506d74d582a9051863d1a17f1ac"
} | |
ok | """Manifold test invariants."""
import tensorflow as tf
import numpy as np
def random_constant(shape, dtype):
return tf.constant(
np.random.uniform(size=shape, high=1e-1),
dtype=dtype.as_numpy_dtype,
)
class TestInvariants(tf.test.TestCase):
def check_random(self, manifold, shape, dtype)... | Evaluate "Manifold test invariants.".
Load tensorflow, referred to as tf.
Load numpy, referred to as np.
Define function random_constant with parameters shape, dtype:
Return tf dot constant with (np dot random dot uniform with size set to shape and high set to 0.1) and dtype set to dtype dot as_numpy_dtype.
Define ... | Evaluar "Manifold test invariants.".
Importar tensorflow, referido como tf.
Importar numpy, referido como np.
Definir función random_constant con parámetros shape, dtype:
Devolver tf punto constant con (np punto random punto uniform con size establecido a shape y también high establecido a 0.1) y también dtype esta... | Évaluer "Manifold test invariants.".
Charger tensorflow, référé comme tf.
Charger numpy, référé comme np.
Définir fonction random_constant avec paramètres shape, dtype:
Retourner tf point de constant avec (np point de random point de uniform avec size défini à shape et high défini à 0.1) et dtype défini à dtype poi... | """Manifold test invariants."""
import tensorflow as tf
import numpy as np
def random_constant(shape, dtype):
return tf.constant(np.random.uniform(size=shape, high=0.1), dtype=dtype.as_numpy_dtype)
class TestInvariants(tf.test.TestCase):
def check_random(self, manifold, shape, dtype):
"""Check random... | true | true | true | {
"artifact_hash": "d28d8be1fae8c3fdb4f55fcf5a5362581b45a8e8b9c4a8735ef6d5a99b1a991a",
"id": "1737186",
"max_stars_count": 33,
"max_stars_repo_name": "vishalbelsare/tensorflow-riemopt",
"max_stars_repo_path": "tensorflow_riemopt/manifolds/test_invariants.py",
"normalized_source_hash": "72bd7b31f2a751b3e0ca7... | {
"artifact_hash": "d28d8be1fae8c3fdb4f55fcf5a5362581b45a8e8b9c4a8735ef6d5a99b1a991a",
"normalized_source_hash": "72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad",
"raw_source_hash": "959338f7e029fb97faba7e3548bed444306cf907a534649f9c9347d45ce345e4",
"runtime_signature_pair_hash": null,
"sour... | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | 求值 "Manifold test invariants.".
导入 tensorflow, 别名为 tf.
导入 numpy, 别名为 np.
定义函数 random_constant 参数为 shape, dtype:
返回 tf 点 constant 用 (np 点 random 点 uniform 用 size 设为 shape 并且 high 设为 0.1) 并且 dtype 设为 dtype 点 as_numpy_dtype.
定义类 TestInvariants 继承自 tf 点 test 点 TestCase:
定义方法 check_random 参数为 self, manifold, shape, ... | """Manifold test invariants."""
import tensorflow as tf
import numpy as np
def random_constant(shape, dtype):
return tf.constant(
np.random.uniform(size=shape, high=1e-1),
dtype=dtype.as_numpy_dtype,
)
class TestInvariants(tf.test.TestCase):
def check_random(self, manifold, shape, dtype)... | true | true | true | null | 72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad | 8,672 | 8,672 | 1 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.242638+00:00 | en,es,fr,zh | true | 72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad | {
"v2_row_index": 1,
"v2_source_sha256": "72bd7b31f2a751b3e0ca75ddcfd8873c8cb8826acd5d68598b3c32d48f37d3ad"
} | |
ok | import pkgutil
import InputDataProcessing
from pcapng import FileScanner
def get_pcap_description(input_file, default="<not provided>"):
"""
Extracts the error description from the PCAPng file (comment section).
:param input_file: Path to the PCAP file.
:param default: When no comment is found, this v... | Load pkgutil.
Load InputDataProcessing.
Load FileScanner from pcapng.
Define function get_pcap_description with parameters input_file, default (default: "<not provided>"):
Text block:
""
" Extracts the error description from the PCAPng file (comment section)."
" :param input_file: Path to the PCAP... | Importar pkgutil.
Importar InputDataProcessing.
Importar FileScanner desde pcapng.
Definir función get_pcap_description con parámetros input_file, default (predeterminado: "<not provided>"):
Texto literal:
""
" Extracts the error description from the PCAPng file (comment section)."
" :param input_... | Charger pkgutil.
Charger InputDataProcessing.
Charger FileScanner depuis pcapng.
Définir fonction get_pcap_description avec paramètres input_file, default (par défaut: "<not provided>"):
Texte littéral:
""
" Extracts the error description from the PCAPng file (comment section)."
" :param input_fil... | import pkgutil
import InputDataProcessing
from pcapng import FileScanner
def get_pcap_description(input_file, default='<not provided>'):
"""
Extracts the error description from the PCAPng file (comment section).
:param input_file: Path to the PCAP file.
:param default: When no comment is found, this va... | true | true | true | {
"artifact_hash": "38d2f2a11a9ea328d06b762d0454fb6d1ffb479d30c032f12cc1032dd1262b03",
"id": "1737187",
"max_stars_count": 0,
"max_stars_repo_name": "marhoSVK/semiauto-diagnostics",
"max_stars_repo_path": "Action_extend.py",
"normalized_source_hash": "3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312... | {
"artifact_hash": "38d2f2a11a9ea328d06b762d0454fb6d1ffb479d30c032f12cc1032dd1262b03",
"normalized_source_hash": "3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3",
"raw_source_hash": "3a2498c2d144053ee28a3d1aabd43ac023ccc0acf346922016c4afe5365b3360",
"runtime_signature_pair_hash": null,
"sour... | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | 导入 pkgutil.
导入 InputDataProcessing.
导入 FileScanner 从 pcapng.
定义函数 get_pcap_description 参数为 input_file, default (默认: "<not provided>"):
文本块:
""
" Extracts the error description from the PCAPng file (comment section)."
" :param input_file: Path to the PCAP file."
" :param default: When no com... | import pkgutil
import InputDataProcessing
from pcapng import FileScanner
def get_pcap_description(input_file, default="<not provided>"):
"""
Extracts the error description from the PCAPng file (comment section).
:param input_file: Path to the PCAP file.
:param default: When no comment is found, this v... | true | true | true | null | 3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3 | 1,853 | 1,853 | 2 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.242943+00:00 | en,es,fr,zh | true | 3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3 | {
"v2_row_index": 2,
"v2_source_sha256": "3b2ce846c05e925ea0468ac8416593d7f651765a0ba8483aab685312b97deab3"
} | |
ok | # Generated by Django 2.1.1 on 2018-10-04 04:56
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='PlaceWeather',
fields=[
('id', models.AutoF... | # Generated by Django 2.1.1 on 2018-10-04 04:56
Load migrations, models from django.db.
Define class Migration inheriting from migrations.Migration:
Set initial to True.
Set dependencies to an empty list.
Set operations to the list [
(migrations dot CreateModel with name set to "PlaceWeather" and fields... | # Generated por Django 2.1.1 on 2018-10-04 04:56
Importar migrations, models desde django.db.
Definir clase Migration heredando de migrations.Migration:
Establecer initial como True.
Establecer dependencies como una lista vacía.
Establecer operations como la lista [
(migrations punto CreateModel con nam... | # Generated par Django 2.1.1 on 2018-10-04 04:56
Charger migrations, models depuis django.db.
Définir classe Migration héritant de migrations.Migration:
Affecter initial à True.
Affecter dependencies à une liste vide.
Affecter operations à la liste [
(migrations point de CreateModel avec name défini à "... | # Generated by Django 2.1.1 on 2018-10-04 04:56
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = []
operations = [migrations.CreateModel(name='PlaceWeather', fields=[('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, ve... | true | true | true | {
"artifact_hash": "64ab55b0cab1d78f880906d127d8234816557e33ae62a4309c9dcdad188b8615",
"id": "1737188",
"max_stars_count": 2,
"max_stars_repo_name": "Escapist-007/Dockerized-Pub-Sub-Distributed-System",
"max_stars_repo_path": "phase_01/MyWayPoints_v1/MyWayPoints/WayPoints/migrations/0001_initial.py",
"norma... | {
"artifact_hash": "64ab55b0cab1d78f880906d127d8234816557e33ae62a4309c9dcdad188b8615",
"normalized_source_hash": "295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa",
"raw_source_hash": "93c0dea6f5372adf083d2ef522f2fac461441d7a7968dd90570c119f70996a24",
"runtime_signature_pair_hash": null,
"sour... | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | # Generated 按 Django 2.1.1 on 2018-10-04 04:56
导入 migrations, models 从 django 点 db.
定义类 Migration 继承自 migrations 点 Migration:
设 initial 为 True.
设 dependencies 为 空列表.
设 operations 为 列表 [
(migrations 点 CreateModel 用 name 设为 "PlaceWeather" 并且 fields 设为 列表 [
元组 ("id", (models 点 AutoField 用 auto_crea... | # Generated by Django 2.1.1 on 2018-10-04 04:56
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='PlaceWeather',
fields=[
('id', models.AutoF... | true | true | true | null | 295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa | 1,077 | 1,077 | 3 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.249990+00:00 | en,es,fr,zh | true | 295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa | {
"v2_row_index": 3,
"v2_source_sha256": "295cd003b7428c35abd126f70bc6a9955f359a4e20bba3aa3c6f8252a2b14dfa"
} | |
ok | "import os\nimport psutil\nimport traceback\n\nimport h5py\nimport numpy as np\n\nfrom buffalo.data (...TRUNCATED) | "Load os.\nLoad psutil.\nLoad traceback.\nLoad h5py.\nLoad numpy, referred to as np.\nLoad prepro fr(...TRUNCATED) | "Importar os.\nImportar psutil.\nImportar traceback.\nImportar h5py.\nImportar numpy, referido como (...TRUNCATED) | "Charger os.\nCharger psutil.\nCharger traceback.\nCharger h5py.\nCharger numpy, référé comme np.(...TRUNCATED) | "import os\nimport psutil\nimport traceback\nimport h5py\nimport numpy as np\nfrom buffalo.data impo(...TRUNCATED) | true | true | true | {"artifact_hash":"01ba4cd5e9a264b6e85cb6b55349843e9c07a0714fe6dbbb9d5b0c822779dab6","id":"1737191","(...TRUNCATED) | {"artifact_hash":"01ba4cd5e9a264b6e85cb6b55349843e9c07a0714fe6dbbb9d5b0c822779dab6","normalized_sour(...TRUNCATED) | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | "导入 os.\n导入 psutil.\n导入 traceback.\n导入 h5py.\n导入 numpy, 别名为 np.\n导入 pr(...TRUNCATED) | "import os\nimport psutil\nimport traceback\n\nimport h5py\nimport numpy as np\n\nfrom buffalo.data (...TRUNCATED) | true | true | true | null | 9cc856fe97656efa950426a1ff2351f0adc075353cf35a5c128fd1b2bae57a72 | 9,783 | 9,783 | 4 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.262630+00:00 | en,es,fr,zh | true | 9cc856fe97656efa950426a1ff2351f0adc075353cf35a5c128fd1b2bae57a72 | {"v2_row_index":4,"v2_source_sha256":"9cc856fe97656efa950426a1ff2351f0adc075353cf35a5c128fd1b2bae57a(...TRUNCATED) | |
ok | "import pandas\n\ndef add_full_name(path_to_csv, path_to_new_csv):\n #Assume you will be reading (...TRUNCATED) | "Load pandas.\nDefine function add_full_name with parameters path_to_csv, path_to_new_csv:\n # As(...TRUNCATED) | "Importar pandas.\nDefinir función add_full_name con parámetros path_to_csv, path_to_new_csv:\n (...TRUNCATED) | "Charger pandas.\nDéfinir fonction add_full_name avec paramètres path_to_csv, path_to_new_csv:\n (...TRUNCATED) | "import pandas\n\ndef add_full_name(path_to_csv, path_to_new_csv):\n df = pandas.read_csv(path_to(...TRUNCATED) | true | true | true | {"artifact_hash":"41bb7ad785ab2359e746b190a2599bcc0844387bad71f17bc8616e18c8fc3713","id":"1737192","(...TRUNCATED) | {"artifact_hash":"41bb7ad785ab2359e746b190a2599bcc0844387bad71f17bc8616e18c8fc3713","normalized_sour(...TRUNCATED) | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | "导入 pandas.\n定义函数 add_full_name 参数为 path_to_csv, path_to_new_csv:\n # Assume yo(...TRUNCATED) | "import pandas\n\ndef add_full_name(path_to_csv, path_to_new_csv):\n #Assume you will be reading (...TRUNCATED) | true | true | true | null | 0124385b1fb972601378d7774926eba679cfad99642aab1ad9b367d1e7963e85 | 1,164 | 1,164 | 5 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.269981+00:00 | en,es,fr,zh | true | 0124385b1fb972601378d7774926eba679cfad99642aab1ad9b367d1e7963e85 | {"v2_row_index":5,"v2_source_sha256":"0124385b1fb972601378d7774926eba679cfad99642aab1ad9b367d1e7963e(...TRUNCATED) | |
ok | "import asyncio\nfrom timeit import default_timer\n\nimport aiohttp\n\nimport settings\n\n\nasync de(...TRUNCATED) | "Load asyncio.\nLoad default_timer from timeit.\nLoad aiohttp.\nLoad settings.\nDefine async functio(...TRUNCATED) | "Importar asyncio.\nImportar default_timer desde timeit.\nImportar aiohttp.\nImportar settings.\nDef(...TRUNCATED) | "Charger asyncio.\nCharger default_timer depuis timeit.\nCharger aiohttp.\nCharger settings.\nDéfin(...TRUNCATED) | "import asyncio\nfrom timeit import default_timer\nimport aiohttp\nimport settings\n\nasync def fetc(...TRUNCATED) | true | true | true | {"artifact_hash":"6daa8d4e8cc2efc0e5d24dfba163cea60806e96504aee6b317a7c83beda15305","id":"1737193","(...TRUNCATED) | {"artifact_hash":"6daa8d4e8cc2efc0e5d24dfba163cea60806e96504aee6b317a7c83beda15305","normalized_sour(...TRUNCATED) | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | "导入 asyncio.\n导入 default_timer 从 timeit.\n导入 aiohttp.\n导入 settings.\n定义异步(...TRUNCATED) | "import asyncio\nfrom timeit import default_timer\n\nimport aiohttp\n\nimport settings\n\n\nasync de(...TRUNCATED) | true | true | true | null | 21bdcfe8905af1779309ffb67760203d14f6708b31fa8e76b24cc022510a110c | 860 | 860 | 6 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.291498+00:00 | en,es,fr,zh | true | 21bdcfe8905af1779309ffb67760203d14f6708b31fa8e76b24cc022510a110c | {"v2_row_index":6,"v2_source_sha256":"21bdcfe8905af1779309ffb67760203d14f6708b31fa8e76b24cc022510a11(...TRUNCATED) | |
ok | "import asyncio\nimport asyncssh\nimport sys\nimport os\nimport crypt\nfrom importlib.util import fi(...TRUNCATED) | "Load asyncio.\nLoad asyncssh.\nLoad sys.\nLoad os.\nLoad crypt.\nLoad find_spec from importlib.util(...TRUNCATED) | "Importar asyncio.\nImportar asyncssh.\nImportar sys.\nImportar os.\nImportar crypt.\nImportar find_(...TRUNCATED) | "Charger asyncio.\nCharger asyncssh.\nCharger sys.\nCharger os.\nCharger crypt.\nCharger find_spec d(...TRUNCATED) | "import asyncio\nimport asyncssh\nimport sys\nimport os\nimport crypt\nfrom importlib.util import fi(...TRUNCATED) | true | true | true | {"artifact_hash":"b2da5200e43bdd1db4d50cad36628e48ec81bbab93e64bbbada9c60a77acfb77","id":"1737194","(...TRUNCATED) | {"artifact_hash":"b2da5200e43bdd1db4d50cad36628e48ec81bbab93e64bbbada9c60a77acfb77","normalized_sour(...TRUNCATED) | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | "导入 asyncio.\n导入 asyncssh.\n导入 sys.\n导入 os.\n导入 crypt.\n导入 find_spec 从 imp(...TRUNCATED) | "import asyncio\nimport asyncssh\nimport sys\nimport os\nimport crypt\nfrom importlib.util import fi(...TRUNCATED) | true | true | true | null | 77f05f17795d4c289aa24fe3adebbe90335078b286eb5c6c371543becb4cb01e | 7,230 | 7,230 | 7 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.292076+00:00 | en,es,fr,zh | true | 77f05f17795d4c289aa24fe3adebbe90335078b286eb5c6c371543becb4cb01e | {"v2_row_index":7,"v2_source_sha256":"77f05f17795d4c289aa24fe3adebbe90335078b286eb5c6c371543becb4cb0(...TRUNCATED) | |
ok | "# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED) | "# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED) | "# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED) | "# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED) | "# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED) | true | true | true | {"artifact_hash":"d10bb428d0640745d72f8ac0a508b93a9167732eee9ba5006754e93fb2f7f083","id":"1737196","(...TRUNCATED) | {"artifact_hash":"d10bb428d0640745d72f8ac0a508b93a9167732eee9ba5006754e93fb2f7f083","normalized_sour(...TRUNCATED) | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | "# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED) | "# лица на фигури\n# Да се напише програма, която въвежда (...TRUNCATED) | true | true | true | null | 7edf844d2d0e2f6a3bc93e3e4f18c59c4c1e245a810f771da91da140a7705e81 | 1,149 | 1,691 | 8 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.497595+00:00 | en,es,fr,zh | true | 7edf844d2d0e2f6a3bc93e3e4f18c59c4c1e245a810f771da91da140a7705e81 | {"v2_row_index":8,"v2_source_sha256":"7edf844d2d0e2f6a3bc93e3e4f18c59c4c1e245a810f771da91da140a7705e(...TRUNCATED) | |
ok | "# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED) | "# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED) | "# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED) | "# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED) | "# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED) | true | true | true | {"artifact_hash":"e5a8cc0051f26778bcc126c2932aef102e63895a68dcd25e8ef7279cb3a6729b","id":"1737198","(...TRUNCATED) | {"artifact_hash":"e5a8cc0051f26778bcc126c2932aef102e63895a68dcd25e8ef7279cb3a6729b","normalized_sour(...TRUNCATED) | cuarzo-100k-v1 | cuarzo-100k-v2-cleaned | "# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED) | "# Copyright 2020 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apach(...TRUNCATED) | true | true | true | null | acf91aabb7629720da6b5758cbc8e7f001bad1d3b9828e9f812975784d8b8db2 | 2,425 | 2,425 | 9 | PcDemo2-cpp-runtime-cleanup-round1048-full copy 3.zip | sharded_spawn | 2026-05-07T09:45:52.524888+00:00 | en,es,fr,zh | true | acf91aabb7629720da6b5758cbc8e7f001bad1d3b9828e9f812975784d8b8db2 | {"v2_row_index":9,"v2_source_sha256":"acf91aabb7629720da6b5758cbc8e7f001bad1d3b9828e9f812975784d8b8d(...TRUNCATED) |
Cuarzo-100K v2
99,683 bidirectional deterministically verified Python to human language pairs across English, Spanish, French, and Mandarin Chinese. Released by Cuarzo AI.
This is the second public release from Aether, our proprietary engine for generating deterministic paired data between code and human language. v2 delivers Mandarin Chinese as a full fourth language surface, expands the verification schema to per-language roundtrip checks across all four surfaces, and adds complete provenance instrumentation at the record level.
The Aether engine is under active development. More source languages, more human languages, larger volumes, and tighter verification are all in progress.
What's new in v2
- Mandarin Chinese added. Every record now includes a
plaincode_zhsurface generated directly from the AST — not translated from English. The engine handles advanced Python constructs includingasync def,await, context managers, and decorator patterns. - Per-language verification. v2 expands roundtrip validation to all four language surfaces independently. Every record carries separate AST equality, compilation, and exact-match checks for EN, ES, FR, and ZH.
- Improved human language phrasing for Spanish and French. The ES and FR representations in v2 reflect refinements to Aether's structured grammar for both languages.
- Full provenance instrumentation. Each record includes processing timestamps, engine build identifiers, per-surface SHA256 hashes, and byte/character counts — enabling end-to-end independent auditability.
- 100% acceptance across all checks. All 99,683 records pass every verification gate across all four language surfaces.
What makes this different
Most code-language datasets fall into one of three buckets:
- Scraped from the web (docstrings, READMEs) — noisy, inconsistent, frequently wrong.
- Generated by LLMs — fast and cheap, but probabilistic and unverifiable.
- Hand-annotated by humans — accurate but slow, expensive, and structurally English-dominant.
Cuarzo takes a different path. Feed in a supported source language, and our Aether engine returns the functionally equivalent intent in another supported language — not a loose summary, not an approximate rewriting, but a verifiable, bidirectional, deterministically controlled alignment across surfaces. The same input always produces the same output. Every pair has to meet strict acceptance standards and is checked structurally before acceptance.
The pipeline
Cuarzo-100K was curated from StarCoderData's Python stream using a multi-stage acceptance pipeline:
- Hard filters on source quality, structure, and licensing.
- Multi-stage deduplication — both raw and structural — within the corpus, our own internal datasets, and all future releases.
- Compile checks on source to ensure parseability.
- Aether translation to structured human language representations in EN, ES, FR, and ZH.
- Roundtrip validation per language — regenerate Python from each human language representation, then verify functional equivalence to the source independently for each surface.
- Strict acceptance gates — only records that pass every check across all four surfaces make it into the dataset.
What's in each record
Core content
| Field | Description |
|---|---|
source_code |
Original Python from StarCoderData |
plaincode_en |
Aether's structured English representation |
plaincode_es |
Aether's structured Spanish representation |
plaincode_fr |
Aether's structured French representation |
plaincode_zh |
Aether's structured Mandarin Chinese representation |
translation_status |
Pipeline status flag (ok for all accepted records) |
Verification — per language
| Field | Description |
|---|---|
roundtrip_python_from_en/es/fr/zh |
Python regenerated from each language representation |
roundtrip_ast_ok |
Boolean: source AST matches roundtrip AST (aggregate across all surfaces) |
roundtrip_ast_ok_en/es/fr/zh |
Boolean: AST equality check per language surface |
roundtrip_compile_ok |
Boolean: all roundtrips compile (aggregate) |
roundtrip_compile_ok_en/es/fr/zh |
Boolean: compilation check per language roundtrip |
roundtrip_exact_ok_en/es/fr/zh |
Boolean: exact structural match per language roundtrip |
roundtrip_error_en/es/fr/zh |
Error message if roundtrip failed; null for all accepted records |
source_compile_ok |
Boolean: original source compiles |
source_compile_error |
Compile error if source failed; null for all accepted records |
all_required_exact_ok |
Boolean: all four language surfaces passed all verification gates |
required_exact_languages |
Languages required to pass for acceptance (value: en,es,fr,zh for all records) |
failed_required_exact_languages |
Languages that failed, if any; null for all accepted records |
failure_reasons |
Failure description if any gate failed; null for all accepted records |
Provenance & hashes
| Field | Description |
|---|---|
source_sha256 |
SHA256 of the original source file |
roundtrip_sha256_from_en/es/fr/zh |
SHA256 of each roundtrip Python output |
hashes |
Full provenance hash bundle (artifact hash, normalized source hash) |
artifact_hash |
Hash of the Aether engine configuration used to generate this record |
source_metadata |
Source repo name, file path, GitHub stars |
Size metrics
| Field | Description |
|---|---|
source_byte_count |
Byte length of source |
source_char_count |
Character length of source |
roundtrip_byte_count_from_en/es/fr/zh |
Byte length of each roundtrip Python output |
roundtrip_char_count_from_en/es/fr/zh |
Character length of each roundtrip Python output |
Composition
Verification
All 99,683 records pass every check across all four language surfaces.
| Check | Records passing |
|---|---|
source_compile_ok |
99,683 / 99,683 (100.00%) |
roundtrip_compile_ok (aggregate) |
99,683 / 99,683 (100.00%) |
roundtrip_compile_ok_en |
99,683 / 99,683 (100.00%) |
roundtrip_compile_ok_es |
99,683 / 99,683 (100.00%) |
roundtrip_compile_ok_fr |
99,683 / 99,683 (100.00%) |
roundtrip_compile_ok_zh |
99,683 / 99,683 (100.00%) |
roundtrip_ast_ok (aggregate) |
99,683 / 99,683 (100.00%) |
roundtrip_ast_ok_en |
99,683 / 99,683 (100.00%) |
roundtrip_ast_ok_es |
99,683 / 99,683 (100.00%) |
roundtrip_ast_ok_fr |
99,683 / 99,683 (100.00%) |
roundtrip_ast_ok_zh |
99,683 / 99,683 (100.00%) |
all_required_exact_ok |
99,683 / 99,683 (100.00%) |
Multilingual coverage
| Language | Records with non-empty representation |
|---|---|
English (plaincode_en) |
99,683 / 99,683 (100.00%) |
Spanish (plaincode_es) |
99,683 / 99,683 (100.00%) |
French (plaincode_fr) |
99,683 / 99,683 (100.00%) |
Mandarin Chinese (plaincode_zh) |
99,683 / 99,683 (100.00%) |
Every record contains all five surfaces: Python source plus four natural-language representations, each independently verified.
Source diversity
| Property | Value |
|---|---|
| Source corpus | StarCoderData Python stream |
| Unique source repositories | 83,489 |
| Average records per repository | 1.2 |
The dataset is drawn from the long tail of public Python code rather than concentrated in a small number of large repositories. The average of 1.2 records per repository means most source files come from distinct projects; concentration is minimal.
Source code size distribution
| Size (characters) | Records | Share |
|---|---|---|
| < 500 | 13,217 | 13.26% |
| 500 – 2,000 | 45,648 | 45.79% |
| 2,000 – 5,000 | 26,208 | 26.29% |
| 5,000 – 10,000 | 9,644 | 9.67% |
| 10,000 – 20,000 | 3,628 | 3.64% |
| > 20,000 | 1,347 | 1.35% |
Source code statistics: minimum 250 chars, maximum 288,385 chars, mean 3,067 chars, median 1,534 chars.
ML and Deep Learning composition
| Category | Records | Share |
|---|---|---|
| Records using any ML library | 15,832 | 15.88% |
| Records using a deep learning framework | 4,887 | 4.90% |
Top libraries by record count:
| Library | Records |
|---|---|
| numpy | 10,238 |
| pandas | 3,514 |
| torch | 2,811 |
| tensorflow | 1,570 |
| sklearn | 1,192 |
| cv2 | 1,188 |
| keras | 588 |
| nltk | 276 |
| transformers | 125 |
| spacy | 104 |
Records may use multiple libraries; library counts can sum to more than the total number of records using any ML library.
Multilingual parity
Global enterprises do not operate in one language, but most public code-language datasets do. Cuarzo-100K v2 gives AI teams verified Python-to-language training records across English, Spanish, French, and Mandarin Chinese — with additional languages planned in future releases.
Each language surface is generated from the same verified semantic foundation — the AST — not translated from another language surface. Teams can train multilingual code understanding without sacrificing consistency, auditability, or behavioral grounding. The Mandarin surface is generated from the same verified semantic foundation as the other language surfaces, with independent per-surface roundtrip verification applied.
Example (excerpt):
Python source:
def connection(host, port, username, password):
try:
connection = get_connection(host=host, port=port, username=username, password=password, use_tls=True)
return connection
except Exception as e:
print(e)
English representation:
Define method connection with parameters host, port, username, password:
Attempt the following:
Set connection to get_connection with host set to host, port set to port, ...
Return connection.
If an error of type Exception occurs, bound as e:
Print e.
Spanish representation:
Definir método connection con parámetros host, port, username, password:
Intentar lo siguiente:
Establecer connection como get_connection con host establecido a host, port establecido a port, ...
Devolver connection.
Si ocurre un error de tipo Exception, como e:
Imprimir e.
French representation:
Définir méthode connection avec paramètres host, port, username, password:
Essayer ce qui suit:
Affecter connection à get_connection avec host défini à host, port défini à port, ...
Retourner connection.
Si une erreur du type Exception survient, lié comme e:
Afficher e.
Mandarin Chinese representation:
定义方法 connection,参数为 host、port、username、password:
尝试以下操作:
将 connection 设置为 get_connection,其中 host 设为 host,port 设为 port,...
返回 connection。
若发生 Exception 类型的错误,绑定为 e:
打印 e。
Why we're releasing this
Curating high quality code-language data is a huge bottleneck for the AI industry today, and every known existing source forces the same tradeoff: cheap and unreliable, or accurate and unscalable without heavy expenses. The teams pushing the frontier of code generation, code understanding, and multilingual AI are all hitting the same ceiling at once.
The team at Cuarzo AI takes a fundamentally different approach: verifiable alignment between code and natural language, with functional equivalence guaranteed and verified by construction rather than estimated by sampling. The same input always produces the same output. Every pair is checked before release — and in v2, checked independently across all four language surfaces.
We're releasing this dataset so the field can build on it. Better training data, better evaluation, and better cross-lingual alignment all start with pairs that are actually consistent — and the only way to prove that is to put the data in researchers' hands and invite them to verify it. Use it. Stress-test it. Tell us what you find.
Intended uses
Cuarzo-100K v2 is designed for:
- Training and fine-tuning code-language models with verified, multilingual pairs
- Multilingual code generation, explanation, and translation research
- Evaluation and benchmarking of code-language models, especially in non-English settings
- Cross-lingual alignment research between Latin-script and CJK-script representations of code semantics
- Provenance and reproducibility research, given the full per-record instrumentation
Permissively licensed under Apache 2.0 for commercial and non-commercial use.
Known limitations
We declare these openly so users can plan around them.
AST equivalence, not full runtime semantic equivalence. Verification confirms that source and roundtrip produce identical normalized ASTs. This is a strong correctness guarantee for nearly all practical purposes, but does not cover all runtime scenarios involving evaluation order of side effects or implementation-defined behavior.
Comments are preserved verbatim. Source code comments (
#-prefixed) are treated as literal annotations and preserved across language representations rather than translated. This is intentional — comment translation introduces ambiguity around author intent and is out of scope for this release.StarCoderData distribution. Source files inherit the composition of StarCoderData's Python stream. Long-tail repository content is well represented; highly popular code patterns may be relatively underrepresented compared to popularity-weighted scraped datasets.
ML/DL composition is minoritarian. 15.88% of records use general ML libraries; 4.90% use deep learning frameworks. Users with ML-specific needs may want to filter accordingly.
Python only in this release. Aether currently supports Python paired with EN/ES/FR/ZH. Additional source languages (C++, Go, Rust) and additional human languages (Hindi, Japanese, German, and beyond) are on the roadmap.
How to load
from datasets import load_dataset
ds = load_dataset("Cuarzo-AI/cuarzo-100k-v2")
# Each record contains five language surfaces plus full per-language verification
sample = ds["train"][0]
print(sample["source_code"])
print(sample["plaincode_zh"])
print(sample["roundtrip_ast_ok_zh"]) # True for all accepted records
Try it. Tell us what you find.
We invite AI labs, research teams, and code-tooling companies to fine-tune or evaluate models on Cuarzo-100K and share results. Standard code benchmarks (HumanEval, MBPP, MultiPL-E) and multilingual evaluations are particularly relevant given the dataset's structural properties. We're especially interested in head-to-head comparisons against fine-tunes on web-scraped or LLM-generated paired data, and in evaluations that specifically test non-English and CJK code understanding.
The per-record provenance fields enable a class of evaluation that isn't possible with other datasets: you can filter by engine build, by processing time, by roundtrip character count, and verify any individual record's equivalence chain independently. We'd like to see that capability used.
If you find that models trained on Cuarzo-100K produce stronger results — or weaker ones — we want to hear about it.
The Aether engine
Cuarzo-100K v2 is a snapshot of what the Aether engine produces at public-release scale. The engine itself is significantly more capable than this dataset reflects.
Volume. Aether is designed for industrial-scale generation. The pipeline that produced this dataset operates throughputs that make hundreds of millions of records commercially viable. The 99,683 records in this release represent a controlled public sample, not a ceiling. AI labs and enterprises requiring training corpora in the tens or hundreds of millions of verified pairs should contact us directly.
Language coverage. The current public release covers Python paired with EN, ES, FR, and ZH. Aether's architecture is language-agnostic at the source and target levels. Additional programming languages — including C++, Go, and Rust — and additional human languages — including Hindi, Portuguese, German — are in active development. Labs with specific language-pair requirements can work with us to prioritize and schedule production runs for their exact needs.
Domain and codebase customization. Aether can be targeted at specific domains, libraries, or internal codebases. A lab building a code model for biomedical Python, for financial analytics tooling, or for a proprietary internal framework can receive verified paired data scoped precisely to that domain — not a general-purpose corpus filtered after the fact. The engine generates from your source material with the same deterministic verification guarantees as the public dataset. For labs with proprietary codebases, Aether is available as an API: the engine runs against your code on your infrastructure, and no source code is transmitted to or stored by Cuarzo AI. You receive verified paired data without exposing your codebase.
Evaluation and benchmark data. Beyond training data, Aether can generate held-out evaluation sets with known provenance and verified correctness properties. For labs that need benchmark data with formal equivalence guarantees rather than crowd-sourced annotations, this is a meaningful structural improvement.
Custom verification thresholds. Acceptance criteria are configurable. Labs with stricter requirements — tighter AST matching, higher exact-match thresholds, or additional compile-time checks — can request production runs with those parameters applied.
Reach out at hello@cuarzoai.com to discuss volume, language coverage, and customization. We respond to every legitimate inquiry within two business days.
Citation
If you use this dataset in research or product work, please cite:
@dataset{cuarzo_100k_v2_2026,
title = {Cuarzo-100K: Deterministically Paired Python and Multilingual Human-Language Code Representations},
author = {Cuarzo AI},
year = {2026},
url = {https://huggingface.co/datasets/Cuarzo-AI/cuarzo-100k-v2}
}
License
Released under the Apache License 2.0.
Source code is sourced from StarCoderData (Python stream) under its respective license. The source_metadata field preserves the originating repository name, file path, and stars count for each record. Users should review individual source repository licenses for any downstream use of specific source files.
Cuarzo-100K-v2 is a public release from the Cuarzo AI team. We work with AI labs, code-tooling companies, and multilingual AI programs that need training and evaluation data with verifiable and provable consistency.
- Downloads last month
- 50