File size: 3,948 Bytes
260a565
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89

import pandas as pd
import torch
from torch_geometric.data import HeteroData

def load_ohada_graph(graph_dir='ohada_graph'):
    """
    Load the OHADA-CCJA Legal Knowledge Graph as a PyG HeteroData object.
    
    Node types: case, domain, state, acte, article, party
    Edge types: cites, classified_as, originates_from, references, cites_article, involves
    """
    data = HeteroData()
    
    # Load nodes
    cases = pd.read_csv(f'{graph_dir}/nodes/cases.csv')
    domains = pd.read_csv(f'{graph_dir}/nodes/legal_domains.csv')
    states = pd.read_csv(f'{graph_dir}/nodes/member_states.csv')
    actes = pd.read_csv(f'{graph_dir}/nodes/actes_uniformes.csv')
    articles = pd.read_csv(f'{graph_dir}/nodes/articles.csv')
    parties = pd.read_csv(f'{graph_dir}/nodes/parties.csv')
    
    # Create ID mappings
    case_id_map = {cid: i for i, cid in enumerate(cases['case_id'])}
    domain_id_map = {did: i for i, did in enumerate(domains['domain_id'])}
    state_id_map = {sid: i for i, sid in enumerate(states['state_id'])}
    acte_id_map = {aid: i for i, aid in enumerate(actes['acte_id'])}
    article_id_map = {int(a): i for i, a in enumerate(articles['article_number'])}
    party_id_map = {pid: i for i, pid in enumerate(parties['party_id'])}
    party_name_map = dict(zip(parties['name'], parties['party_id']))
    
    # Node counts
    data['case'].num_nodes = len(cases)
    data['domain'].num_nodes = len(domains)
    data['state'].num_nodes = len(states)
    data['acte'].num_nodes = len(actes)
    data['article'].num_nodes = len(articles)
    data['party'].num_nodes = len(parties)
    
    # Node features: year as a basic feature for cases
    years = cases['year'].fillna(0).values.astype(float)
    data['case'].x = torch.tensor(years, dtype=torch.float).unsqueeze(1)
    
    # Load edges
    def load_edges(file, src_col, tgt_col, src_map, tgt_map):
        df = pd.read_csv(f'{graph_dir}/edges/{file}')
        valid = df[src_col].map(src_map).notna() & df[tgt_col].map(tgt_map).notna()
        df = df[valid]
        src = torch.tensor(df[src_col].map(src_map).values.astype(int))
        tgt = torch.tensor(df[tgt_col].map(tgt_map).values.astype(int))
        return torch.stack([src, tgt], dim=0)
    
    # Case β†’ cites β†’ Case
    cite_df = pd.read_csv(f'{graph_dir}/edges/case_cites_case.csv')
    valid = cite_df['source_case_id'].isin(case_id_map) & cite_df['cited_case_id'].isin(case_id_map)
    cite_valid = cite_df[valid]
    if len(cite_valid) > 0:
        src = torch.tensor([case_id_map[x] for x in cite_valid['source_case_id']])
        tgt = torch.tensor([case_id_map[x] for x in cite_valid['cited_case_id']])
        data['case', 'cites', 'case'].edge_index = torch.stack([src, tgt])
    
    # Case β†’ classified_as β†’ Domain
    data['case', 'classified_as', 'domain'].edge_index = load_edges(
        'case_classified_as_domain.csv', 'case_id', 'domain_id', case_id_map, domain_id_map)
    
    # Case β†’ originates_from β†’ State
    data['case', 'originates_from', 'state'].edge_index = load_edges(
        'case_originates_from_state.csv', 'case_id', 'state_id', case_id_map, state_id_map)
    
    # Case β†’ references β†’ Acte
    data['case', 'references', 'acte'].edge_index = load_edges(
        'case_references_acte.csv', 'case_id', 'acte_id', case_id_map, acte_id_map)
    
    # Case β†’ cites_article β†’ Article
    art_df = pd.read_csv(f'{graph_dir}/edges/case_cites_article.csv')
    valid = art_df['case_id'].isin(case_id_map) & art_df['article_number'].isin(article_id_map)
    art_valid = art_df[valid]
    if len(art_valid) > 0:
        src = torch.tensor([case_id_map[x] for x in art_valid['case_id']])
        tgt = torch.tensor([article_id_map[int(x)] for x in art_valid['article_number']])
        data['case', 'cites_article', 'article'].edge_index = torch.stack([src, tgt])
    
    return data


# Usage:
# data = load_ohada_graph('ohada_graph')
# print(data)