File size: 6,906 Bytes
d58b3fb d5eafa3 d58b3fb 1145702 d58b3fb bc2c223 1145702 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 |
---
dataset_name: WikiCulture
language:
- en
license: mit
task_categories:
- text-generation
size_categories:
- 100K<n<1M
source_datasets:
- wikipedia
configs:
- config_name: AF
data_files:
- split: train
path: AF.parquet
- config_name: AS
data_files:
- split: train
path: AS.parquet
- config_name: AU
data_files:
- split: train
path: AU.parquet
- config_name: CH
data_files:
- split: train
path: CH.parquet
- config_name: EU
data_files:
- split: train
path: EU.parquet
- config_name: LA
data_files:
- split: train
path: LA.parquet
- config_name: ME
data_files:
- split: train
path: ME.parquet
- config_name: NA
data_files:
- split: train
path: NA.parquet
---
# Wikipedia Culture Dataset (`nDNA/WikiCulture`)
## Overview
This dataset provides coarse-grained geographic culture labels for English Wikipedia articles, mapped into eight buckets:
* **NA**: North America
* **EU**: Europe
* **AU**: Oceania (UN M49 “Oceania”)
* **AS**: Asia (excluding Western Asia and Greater China)
* **CH**: Greater China (CN, HK, MO, TW)
* **AF**: Africa
* **LA**: Latin America & Caribbean (Americas excluding Northern America)
* **ME**: Middle East (UN M49 “Western Asia”)
The labels are intended for controlled sampling and stratified analysis of Wikipedia content by broad region. They are not intended as fine-grained cultural or ethnographic ground truth.
## Data Sources
The dataset is constructed from three public components:
1. **UN M49** country/area regional classification (UN Statistics Division), used to generate a deterministic mapping from ISO3166-1 alpha-2 codes to the eight buckets.
2. **Wikipedia Cultural Diversity** dataset, used as the article-level signal source (including per-article ISO country code and Wikidata-derived fields) [https://doi.org/10.6084/m9.figshare.7039514].
3. **`wikimedia/wikipedia` (Hugging Face dataset)**, used to attach the full article text via a join.
## Labeling Methodology
### 1) Deterministic ISO2 → bucket mapping (UN M49)
An ISO3166-1 alpha-2 code is mapped to one of the eight buckets using UN M49 region/subregion fields:
* **EU** if UN M49 Region = *Europe*
* **AF** if Region = *Africa*
* **AU** if Region = *Oceania*
* **ME** if Sub-region = *Western Asia*
* **NA** if Region = *Americas* and Sub-region = *Northern America*
* **LA** if Region = *Americas* and not Northern America
* **CH** if ISO2 ∈ {CN, HK, MO, TW} (Taiwan forced to CH)
* **AS** if Region = *Asia* and not Western Asia and not CH
This mapping is deterministic and versioned.
### 2) Geo label assignment from CCC
Each article in the CCC dump provides `iso3166` (ISO3166-1 alpha-2). The primary label is:
**culture_geo = bucket(iso3166)**
Rows with missing `iso3166` or unmapped ISO2 are excluded.
### 3) High-precision consistency filtering
To increase label precision and remove cross-regional or ambiguous items, an additional consistency check is applied using CCC’s Wikidata-derived columns:
* `country_wd`
* `location_wd`
**Offline QID → ISO2 bootstrap.**
Because no online Wikidata queries are used, we construct a bootstrapped mapping from QID to ISO2 by taking the most frequent `iso3166` observed for each `qitem` in the CCC dump.
**Evidence extraction.**
For each article, QIDs are parsed from `country_wd` and `location_wd`, mapped to ISO2 using the bootstrapped table, then converted to bucket sets via the UN M49 mapping.
**Hard-strong inclusion rule.**
An article is retained as *high-precision* if:
* It has at least one non-empty evidence set from `country_wd` or `location_wd`, and
* If `location_wd` evidence is present, it must equal `{culture_geo}` exactly; otherwise,
* `country_wd` evidence must equal `{culture_geo}` exactly.
This rule is deliberately discard-heavy.
### 4) Middle East retention exception
In the CCC snapshot used, many ME-labeled rows have empty evidence after offline QID resolution (coverage limitations of the bootstrapped mapping). To avoid eliminating nearly all ME-labeled examples, a restricted exception is applied:
* If `culture_geo == "ME"` and both evidence sets are empty, the row is retained (geo-only).
## Attaching Wikipedia Text (`wikimedia/wikipedia` join)
The final dataset includes the full article text by merging the filtered labeled table with the Hugging Face dataset **`wikimedia/wikipedia`**.
* A **left inner join** is performed on the article identifier:
* CCC / labeled table: `page_id`
* `wikimedia/wikipedia`: `id`
Only rows that match across both sources are included in the released dataset, ensuring every labeled example has associated article text.
## Dataset Fields
Each row contains:
* `page_title` (string): Wikipedia article title (from CCC / Wikipedia metadata)
* `Qid` (string): Wikidata QID (`qitem`)
* `culture_geo` (string): one of `{NA, EU, AU, AS, CH, AF, LA, ME}`
* `text` (string): The article text (from `wikimedia/wikipedia`)
## Intended Use
This dataset is suitable for:
* constructing geographically stratified Wikipedia subsets,
* controlled pretraining mixtures or evaluation subsets by region,
* studying representation and performance disparities across broad regions.
It is not suitable for:
* fine-grained cultural identity claims,
* country-level ground truth without additional validation,
* resolving inherently transnational or multi-regional topics (which are preferentially excluded by design).
## Limitations
* Labels are coarse and operationalize “culture” as broad geography.
* The strict filter excludes many global and transnational pages.
* The offline QID→ISO2 bootstrap is a heuristic and may miss QIDs not well-represented in the CCC rows.
* The Middle East geo-only exception yields a subset with weaker evidence than the hard-strong subset, but is retained to mitigate coverage loss under offline constraints.
* The final dataset includes only articles present in both the labeled CCC-derived table and `wikimedia/wikipedia` after the `page_id` ↔ `id` join.
## Reproducibility
The pipeline is deterministic given:
* the UN M49 table snapshot used,
* the CCC dump snapshot used,
* the fixed mapping rules and filtering rule described above,
* the `wikimedia/wikipedia` snapshot/config used for the text join.
## Licensing and Attribution
This dataset is derived from Wikipedia-related resources and UN statistical classifications. Users should comply with the licensing and attribution requirements of:
* Wikipedia content (CC BY-SA and related terms),
* the CCC dataset’s licensing/terms,
* UN M49 documentation and terms where applicable.
## Citation
If you use this dataset, cite:
* the CCC dataset (Wikipedia Cultural Diversity / Cultural Context Content) and associated publication,
* UN M49 (UN Statistics Division),
* the `wikimedia/wikipedia` Hugging Face dataset,
* this dataset repository (`nDNA/WikiCulture`). |