docfreemo commited on
Commit
f9b2820
·
verified ·
1 Parent(s): ff2ed16

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +11 -37
README.md CHANGED
@@ -12,31 +12,6 @@ tags:
12
  - triples
13
  size_categories:
14
  - 1K<n<10K
15
- configs:
16
- - config_name: default
17
- data_files:
18
- - split: data
19
- path: data/data-*
20
- dataset_info:
21
- features:
22
- - name: subject
23
- dtype: string
24
- - name: predicate
25
- dtype: string
26
- - name: object
27
- dtype: string
28
- - name: object_type
29
- dtype: string
30
- - name: object_datatype
31
- dtype: string
32
- - name: object_language
33
- dtype: string
34
- splits:
35
- - name: data
36
- num_bytes: 622920511
37
- num_examples: 3854624
38
- download_size: 106525981
39
- dataset_size: 622920511
40
  ---
41
 
42
  # WordNet RDF
@@ -111,12 +86,11 @@ from datasets import load_dataset
111
  # Load the dataset
112
  dataset = load_dataset("CleverThis/wordnet")
113
 
114
- # Access train and test splits
115
- train_data = dataset["train"]
116
- test_data = dataset["test"]
117
 
118
  # Iterate over triples
119
- for row in train_data:
120
  subject = row["subject"]
121
  predicate = row["predicate"]
122
  obj = row["object"]
@@ -138,7 +112,7 @@ The dataset can be converted back to any RDF format (Turtle, N-Triples, RDF/XML,
138
  from datasets import load_dataset
139
  from rdflib import Graph, URIRef, Literal, BNode
140
 
141
- def convert_to_rdf(dataset_name, output_file="output.ttl", split="train"):
142
  """Convert HuggingFace dataset back to RDF Turtle format."""
143
  # Load dataset
144
  dataset = load_dataset(dataset_name)
@@ -203,28 +177,28 @@ from datasets import load_dataset
203
  dataset = load_dataset("CleverThis/wordnet")
204
 
205
  # Find all triples with English literals
206
- english_literals = dataset["train"].filter(
207
  lambda x: x["object_type"] == "literal" and x["object_language"] == "en"
208
  )
209
  print(f"Found {len(english_literals)} English literals")
210
 
211
  # Find all rdf:type statements
212
- type_statements = dataset["train"].filter(
213
  lambda x: "rdf-syntax-ns#type" in x["predicate"]
214
  )
215
  print(f"Found {len(type_statements)} type statements")
216
 
217
  # Convert to Pandas for analysis
218
  import pandas as pd
219
- df = dataset["train"].to_pandas()
220
 
221
  # Analyze predicate distribution
222
  print(df["predicate"].value_counts())
223
  ```
224
 
225
- ### Dataset Splits
226
 
227
- The dataset is split into **train** (95%) and **test** (5%) sets for machine learning tasks such as:
228
 
229
  - Knowledge graph completion
230
  - Link prediction
@@ -251,7 +225,7 @@ The specification includes:
251
  - **Original Size**: 0.21 GB
252
  - **Conversion Tool**: [CleverErnie RDF Pipeline](https://github.com/CleverThis/cleverernie)
253
  - **Format Version**: 1.0
254
- - **Conversion Date**: 2025-11-02
255
 
256
 
257
  ## Citation
@@ -290,7 +264,7 @@ https://en-word.net/static/english-wordnet-2024.ttl.gz
290
 
291
  - Converted using: [CleverErnie GISM](https://github.com/cleverthis/cleverernie)
292
  - Conversion script: `scripts/convert_rdf_to_hf_dataset.py`
293
- - Train/test split: 95%/5%
294
 
295
  ### Maintenance
296
 
 
12
  - triples
13
  size_categories:
14
  - 1K<n<10K
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ---
16
 
17
  # WordNet RDF
 
86
  # Load the dataset
87
  dataset = load_dataset("CleverThis/wordnet")
88
 
89
+ # Access the data
90
+ data = dataset["data"]
 
91
 
92
  # Iterate over triples
93
+ for row in data:
94
  subject = row["subject"]
95
  predicate = row["predicate"]
96
  obj = row["object"]
 
112
  from datasets import load_dataset
113
  from rdflib import Graph, URIRef, Literal, BNode
114
 
115
+ def convert_to_rdf(dataset_name, output_file="output.ttl", split="data"):
116
  """Convert HuggingFace dataset back to RDF Turtle format."""
117
  # Load dataset
118
  dataset = load_dataset(dataset_name)
 
177
  dataset = load_dataset("CleverThis/wordnet")
178
 
179
  # Find all triples with English literals
180
+ english_literals = dataset["data"].filter(
181
  lambda x: x["object_type"] == "literal" and x["object_language"] == "en"
182
  )
183
  print(f"Found {len(english_literals)} English literals")
184
 
185
  # Find all rdf:type statements
186
+ type_statements = dataset["data"].filter(
187
  lambda x: "rdf-syntax-ns#type" in x["predicate"]
188
  )
189
  print(f"Found {len(type_statements)} type statements")
190
 
191
  # Convert to Pandas for analysis
192
  import pandas as pd
193
+ df = dataset["data"].to_pandas()
194
 
195
  # Analyze predicate distribution
196
  print(df["predicate"].value_counts())
197
  ```
198
 
199
+ ### Dataset Format
200
 
201
+ The dataset contains all triples in a single **data** split, suitable for machine learning tasks such as:
202
 
203
  - Knowledge graph completion
204
  - Link prediction
 
225
  - **Original Size**: 0.21 GB
226
  - **Conversion Tool**: [CleverErnie RDF Pipeline](https://github.com/CleverThis/cleverernie)
227
  - **Format Version**: 1.0
228
+ - **Conversion Date**: 2025-11-05
229
 
230
 
231
  ## Citation
 
264
 
265
  - Converted using: [CleverErnie GISM](https://github.com/cleverthis/cleverernie)
266
  - Conversion script: `scripts/convert_rdf_to_hf_dataset.py`
267
+ - Dataset format: Single 'data' split with all triples
268
 
269
  ### Maintenance
270