File size: 2,844 Bytes
17a25ed
376276d
 
17a25ed
 
376276d
 
17a25ed
0eb7efd
 
fa6b88d
cf9b14b
f8bd0b4
 
77f5d1b
 
 
8d043df
77f5d1b
 
 
 
 
cb1e3cc
359f586
 
 
8a2d776
b31f884
 
 
359f586
 
cb1e3cc
8a2d776
b31f884
 
 
cb1e3cc
8a2d776
b31f884
 
 
cb1e3cc
 
 
 
 
 
77f5d1b
 
f8bd0b4
6b00de9
 
f8bd0b4
1565343
 
6b00de9
b0309d4
77f5d1b
 
6b00de9
cf9b14b
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
---
language:
- en
tags:
- hypergraph
task_categories:
- graph-ml
---
# email-Enron

[Zenodo](https://zenodo.org/records/10155819) | [Cornell](https://www.cs.cornell.edu/~arb/data/email-Enron/) | [Source Paper](https://arxiv.org/abs/1802.06916)

email-Enron is an undirected hypergraph built from the Enron email corpus, designed for higher-order network / hypergraph machine learning. In email communication, a single message can involve more than two people; this dataset captures that group interaction by modeling each email as a hyperedge connecting the sender and all recipients, while nodes represent Enron email addresses (restricted to a core set of employees).

## Usage

```python
dataset = load_dataset("daqh/email-Enron", split="full")
hypergraphs = [xgi.from_hif_dict(d, nodetype=int, edgetype=int) for d in dataset]
```

## Statistics

<div align="center">
  <table>
    <tbody>
      <tr>
        <td colspan="2" align="center">
          <figure>
            <img src="assets/hypergraph.png">
          </figure>
        </td>
      </tr>
      <tr>
        <td align="center">
          <figure>
            <img src="assets/node-degree-distribution.png">
          </figure>
        </td>
        <td align="center">
          <figure>
            <img src="assets/hyperedge-size-distribution.png">
          </figure>
        </td>
      </tr>
    </tbody>
  </table>
</div>

## Content

The hypergraph is stored in HIF (Hypergraph Interchange Format) as a JSON object, following the schema used to exchange higher-order network data across tools. Concretely, the dataset provides the canonical HIF fields-network-type, metadata, nodes, edges, and incidences-so you can reconstruct the full incidence structure without additional processing.

In addition to the raw hypergraph topology, vector features are provided for both nodes and hyperedges (in their attribute dictionaries), enabling out-of-the-box experimentation with representation learning and downstream tasks:
- Spectral features: eigenvectors of the (hypergraph) Laplacian (computed via sparse eigensolvers).
- [Node2Vec](https://arxiv.org/abs/1607.00653) embeddings: random-walk–based structural embeddings.
- [VilLain](https://dl.acm.org/doi/10.1145/3589334.3645454) embeddings: self-supervised hypergraph representation learning via virtual label propagation.

Basic statistics (as packaged here): 143 nodes, 1512 hyperedges, 1 connected component.

## Citation

```
@article{Benson-2018-simplicial,
 author = {Benson, Austin R. and Abebe, Rediet and Schaub, Michael T. and Jadbabaie, Ali and Kleinberg, Jon},
 title = {Simplicial closure and higher-order link prediction},
 year = {2018},
 doi = {10.1073/pnas.1800683115},
 publisher = {National Academy of Sciences},
 issn = {0027-8424},
 journal = {Proceedings of the National Academy of Sciences}
}
```