Upload 6 files
Browse files- .gitattributes +1 -0
- cn_relations_clean.json +3 -0
- src/cn_analyze.py +23 -0
- src/cn_clean.py +35 -0
- src/cn_inspect.py +24 -0
- src/cn_parser.py +54 -0
- src/parse_cn.sh +24 -0
.gitattributes
CHANGED
|
@@ -53,3 +53,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 53 |
*.jpg filter=lfs diff=lfs merge=lfs -text
|
| 54 |
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
| 55 |
*.webp filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 53 |
*.jpg filter=lfs diff=lfs merge=lfs -text
|
| 54 |
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
| 55 |
*.webp filter=lfs diff=lfs merge=lfs -text
|
| 56 |
+
cn_relations_clean.json filter=lfs diff=lfs merge=lfs -text
|
cn_relations_clean.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0efc7e1ab6d75dc7392be33765b5cbba9d8f388d321a8e0dbf51d33840800d37
|
| 3 |
+
size 961527622
|
src/cn_analyze.py
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import json
|
| 2 |
+
|
| 3 |
+
# Load the relations dictionary from the JSON file
|
| 4 |
+
input_file_path = "cn_relations_clean.json"
|
| 5 |
+
with open(input_file_path, "r") as json_file:
|
| 6 |
+
relations_dict = json.load(json_file)
|
| 7 |
+
|
| 8 |
+
# Prepare the results
|
| 9 |
+
results = []
|
| 10 |
+
num_languages = len(relations_dict)
|
| 11 |
+
results.append(f"Number of languages: {num_languages}\n")
|
| 12 |
+
|
| 13 |
+
for language_code, start_nodes in relations_dict.items():
|
| 14 |
+
num_start_edges = len(start_nodes.keys())
|
| 15 |
+
num_end_edges = len([end_node for edges in start_nodes.values() for end_node in edges])
|
| 16 |
+
results.append(f"Language: {language_code} - Number of start edges: {num_start_edges} - Number of end edges: {num_end_edges}\n")
|
| 17 |
+
|
| 18 |
+
# Write the results to a text file
|
| 19 |
+
output_file_path = "cn_relations_summary.txt"
|
| 20 |
+
with open(output_file_path, "w") as txt_file:
|
| 21 |
+
txt_file.writelines(results)
|
| 22 |
+
|
| 23 |
+
print(f"Summary saved to {output_file_path}")
|
src/cn_clean.py
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import json
|
| 2 |
+
import re
|
| 3 |
+
|
| 4 |
+
def clean_prefixes(data):
|
| 5 |
+
cleaned_data = {}
|
| 6 |
+
prefix_pattern = re.compile(r'^/c/[^/]+/') # Matches the prefix part '/c/<lang_code>/'
|
| 7 |
+
|
| 8 |
+
for lang, words in data.items():
|
| 9 |
+
cleaned_words = {}
|
| 10 |
+
for key, values in words.items():
|
| 11 |
+
# Clean the key
|
| 12 |
+
cleaned_key = prefix_pattern.sub('', key).replace('_', ' ')
|
| 13 |
+
|
| 14 |
+
# Clean the values
|
| 15 |
+
cleaned_values = [prefix_pattern.sub('', value).replace('_', ' ') for value in values]
|
| 16 |
+
|
| 17 |
+
cleaned_words[cleaned_key] = cleaned_values
|
| 18 |
+
|
| 19 |
+
cleaned_data[lang] = cleaned_words
|
| 20 |
+
|
| 21 |
+
return cleaned_data
|
| 22 |
+
|
| 23 |
+
# Load the relations dictionary from the JSON file
|
| 24 |
+
input_file_path = "cn_relations.json"
|
| 25 |
+
with open(input_file_path, "r") as json_file:
|
| 26 |
+
data = json.load(json_file)
|
| 27 |
+
|
| 28 |
+
cleaned_data = clean_prefixes(data)
|
| 29 |
+
|
| 30 |
+
print(cleaned_data['zdj'])
|
| 31 |
+
|
| 32 |
+
# Save the cleaned data to a new JSON file
|
| 33 |
+
output_file_path = "cn_relations_clean.json"
|
| 34 |
+
with open(output_file_path, "w") as json_file:
|
| 35 |
+
json.dump(cleaned_data, json_file, ensure_ascii=False, indent=4)
|
src/cn_inspect.py
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import json
|
| 2 |
+
from itertools import chain
|
| 3 |
+
|
| 4 |
+
# Load the relations dictionary from the JSON file
|
| 5 |
+
input_file_path = "cn_relations.json"
|
| 6 |
+
with open(input_file_path, "r") as json_file:
|
| 7 |
+
relations_dict = json.load(json_file)
|
| 8 |
+
|
| 9 |
+
# Define the specific language to look for
|
| 10 |
+
specific_language = "ii"
|
| 11 |
+
|
| 12 |
+
# Check if the specific language exists in the dictionary
|
| 13 |
+
if specific_language in relations_dict:
|
| 14 |
+
start_nodes = relations_dict[specific_language]
|
| 15 |
+
print(start_nodes)
|
| 16 |
+
num_start_edges = sum(len(edges) for edges in start_nodes.values())
|
| 17 |
+
all_end_edges = list(chain.from_iterable(start_nodes.values()))
|
| 18 |
+
num_start_edges = len(start_nodes.keys())
|
| 19 |
+
num_end_edges = len([end_node for edges in start_nodes.values() for end_node in edges])
|
| 20 |
+
print(f"Language: {specific_language} - Number of start edges: {num_start_edges} - Number of end edges: {num_end_edges}\n")
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
else:
|
| 24 |
+
print(f"Language {specific_language} not found in the dataset.")
|
src/cn_parser.py
ADDED
|
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import gzip
|
| 2 |
+
import json
|
| 3 |
+
|
| 4 |
+
# Define the dictionary to store the relations
|
| 5 |
+
print("--------> Sequential processing setup!")
|
| 6 |
+
relations_dict = {}
|
| 7 |
+
|
| 8 |
+
# Path to the gzipped file
|
| 9 |
+
file_path = "conceptnet-assertions-5.7.0.csv.gz"
|
| 10 |
+
|
| 11 |
+
# Function to parse each line and extract words and relations
|
| 12 |
+
def parse_line(line):
|
| 13 |
+
parts = line.strip().split("\t")
|
| 14 |
+
if len(parts) >= 3:
|
| 15 |
+
# Extract the start and end nodes
|
| 16 |
+
start_node_uri = parts[2]
|
| 17 |
+
end_node_uri = parts[3]
|
| 18 |
+
|
| 19 |
+
# Extract the language code
|
| 20 |
+
start_lang_code = start_node_uri.split("/")[2]
|
| 21 |
+
language_code = start_lang_code
|
| 22 |
+
|
| 23 |
+
# Extract the node names
|
| 24 |
+
start_node = start_node_uri.split("/")[3]
|
| 25 |
+
end_node = end_node_uri.split("/")[3]
|
| 26 |
+
|
| 27 |
+
# Filter out end nodes that are http:// links
|
| 28 |
+
if end_node_uri.startswith("http://"):
|
| 29 |
+
return
|
| 30 |
+
|
| 31 |
+
# Initialize the language dictionary if it doesn't exist
|
| 32 |
+
if language_code not in relations_dict:
|
| 33 |
+
relations_dict[language_code] = {}
|
| 34 |
+
|
| 35 |
+
# Initialize the start node list if it doesn't exist
|
| 36 |
+
if start_node not in relations_dict[language_code]:
|
| 37 |
+
relations_dict[language_code][start_node] = []
|
| 38 |
+
# Add the end node to the start node's list
|
| 39 |
+
relations_dict[language_code][start_node].append(end_node)
|
| 40 |
+
|
| 41 |
+
# Open the gzipped file and parse each line sequentially
|
| 42 |
+
with gzip.open(file_path, "rt") as f:
|
| 43 |
+
print("--------> File opened!")
|
| 44 |
+
for line in f:
|
| 45 |
+
parse_line(line)
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
# Save the relations dictionary to a JSON file
|
| 49 |
+
print("--------> About to save the file!")
|
| 50 |
+
output_file_path = "cn_relations.json"
|
| 51 |
+
with open(output_file_path, "w") as json_file:
|
| 52 |
+
json.dump(relations_dict, json_file, ensure_ascii=False)
|
| 53 |
+
|
| 54 |
+
print("Data saved to", output_file_path)
|
src/parse_cn.sh
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/bin/bash
|
| 2 |
+
|
| 3 |
+
# download ConceptNet data from the official website
|
| 4 |
+
# file name - conceptnet-assertions-5.7.0.csv.gz
|
| 5 |
+
# link - https://s3.amazonaws.com/conceptnet/downloads/2019/edges/conceptnet-assertions-5.7.0.csv.gz
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
echo "Running cn_parser.py..."
|
| 9 |
+
python cn_parser.py
|
| 10 |
+
|
| 11 |
+
echo "Running cn_inspect.py..."
|
| 12 |
+
python cn_inspect.py
|
| 13 |
+
|
| 14 |
+
echo "Running cn_clean.py..."
|
| 15 |
+
python cn_clean.py
|
| 16 |
+
|
| 17 |
+
echo "Running cn_analyze.py..."
|
| 18 |
+
python cn_analyze.py
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
# output files:
|
| 22 |
+
# 1. complete conceptnet data without sources and relation types
|
| 23 |
+
# 2. clean concpenete data without prefixes and underscores between words
|
| 24 |
+
# 3. analysis file with the number of languages and edges for each language
|