File size: 3,948 Bytes
260a565 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 |
import pandas as pd
import torch
from torch_geometric.data import HeteroData
def load_ohada_graph(graph_dir='ohada_graph'):
"""
Load the OHADA-CCJA Legal Knowledge Graph as a PyG HeteroData object.
Node types: case, domain, state, acte, article, party
Edge types: cites, classified_as, originates_from, references, cites_article, involves
"""
data = HeteroData()
# Load nodes
cases = pd.read_csv(f'{graph_dir}/nodes/cases.csv')
domains = pd.read_csv(f'{graph_dir}/nodes/legal_domains.csv')
states = pd.read_csv(f'{graph_dir}/nodes/member_states.csv')
actes = pd.read_csv(f'{graph_dir}/nodes/actes_uniformes.csv')
articles = pd.read_csv(f'{graph_dir}/nodes/articles.csv')
parties = pd.read_csv(f'{graph_dir}/nodes/parties.csv')
# Create ID mappings
case_id_map = {cid: i for i, cid in enumerate(cases['case_id'])}
domain_id_map = {did: i for i, did in enumerate(domains['domain_id'])}
state_id_map = {sid: i for i, sid in enumerate(states['state_id'])}
acte_id_map = {aid: i for i, aid in enumerate(actes['acte_id'])}
article_id_map = {int(a): i for i, a in enumerate(articles['article_number'])}
party_id_map = {pid: i for i, pid in enumerate(parties['party_id'])}
party_name_map = dict(zip(parties['name'], parties['party_id']))
# Node counts
data['case'].num_nodes = len(cases)
data['domain'].num_nodes = len(domains)
data['state'].num_nodes = len(states)
data['acte'].num_nodes = len(actes)
data['article'].num_nodes = len(articles)
data['party'].num_nodes = len(parties)
# Node features: year as a basic feature for cases
years = cases['year'].fillna(0).values.astype(float)
data['case'].x = torch.tensor(years, dtype=torch.float).unsqueeze(1)
# Load edges
def load_edges(file, src_col, tgt_col, src_map, tgt_map):
df = pd.read_csv(f'{graph_dir}/edges/{file}')
valid = df[src_col].map(src_map).notna() & df[tgt_col].map(tgt_map).notna()
df = df[valid]
src = torch.tensor(df[src_col].map(src_map).values.astype(int))
tgt = torch.tensor(df[tgt_col].map(tgt_map).values.astype(int))
return torch.stack([src, tgt], dim=0)
# Case β cites β Case
cite_df = pd.read_csv(f'{graph_dir}/edges/case_cites_case.csv')
valid = cite_df['source_case_id'].isin(case_id_map) & cite_df['cited_case_id'].isin(case_id_map)
cite_valid = cite_df[valid]
if len(cite_valid) > 0:
src = torch.tensor([case_id_map[x] for x in cite_valid['source_case_id']])
tgt = torch.tensor([case_id_map[x] for x in cite_valid['cited_case_id']])
data['case', 'cites', 'case'].edge_index = torch.stack([src, tgt])
# Case β classified_as β Domain
data['case', 'classified_as', 'domain'].edge_index = load_edges(
'case_classified_as_domain.csv', 'case_id', 'domain_id', case_id_map, domain_id_map)
# Case β originates_from β State
data['case', 'originates_from', 'state'].edge_index = load_edges(
'case_originates_from_state.csv', 'case_id', 'state_id', case_id_map, state_id_map)
# Case β references β Acte
data['case', 'references', 'acte'].edge_index = load_edges(
'case_references_acte.csv', 'case_id', 'acte_id', case_id_map, acte_id_map)
# Case β cites_article β Article
art_df = pd.read_csv(f'{graph_dir}/edges/case_cites_article.csv')
valid = art_df['case_id'].isin(case_id_map) & art_df['article_number'].isin(article_id_map)
art_valid = art_df[valid]
if len(art_valid) > 0:
src = torch.tensor([case_id_map[x] for x in art_valid['case_id']])
tgt = torch.tensor([article_id_map[int(x)] for x in art_valid['article_number']])
data['case', 'cites_article', 'article'].edge_index = torch.stack([src, tgt])
return data
# Usage:
# data = load_ohada_graph('ohada_graph')
# print(data)
|