Datasets:
datetime listlengths 114 2.19k | type_event listlengths 114 2.19k | time_since_last_event listlengths 114 2.19k | time_since_start listlengths 114 2.19k | seq_idx int64 0 1.35k | dim_process int64 3 3 | seq_len int64 114 2.19k |
|---|---|---|---|---|---|---|
[
"2015-12-10T14:39:21",
"2015-12-10T15:39:04",
"2015-12-10T16:17:05",
"2015-12-10T16:32:10",
"2015-12-10T16:46:48",
"2015-12-10T16:47:36",
"2015-12-10T16:51:42",
"2015-12-10T16:56:52",
"2015-12-10T17:09:49",
"2015-12-10T17:10:52",
"2015-12-10T17:12:47",
"2015-12-10T17:15:49",
"2015-12-10T17:2... | [
1,
0,
0,
2,
0,
0,
0,
0,
2,
0,
2,
2,
0,
2,
2,
2,
0,
0,
0,
2,
2,
2,
2,
1,
2,
0,
0,
1,
2,
2,
0,
2,
0,
2,
2,
2,
2,
0,
2,
1,
2,
0,
0,
2,
1,
0,
2,
2,
1,
0,
2,
2,
0,
2,
1,
0,
0,
0,
0,
2,
0,
0,
0,
0... | [
0,
0.9952777777777778,
0.6336111111111111,
0.2513888888888889,
0.24388888888888888,
0.013333333333333334,
0.06833333333333333,
0.08611111111111111,
0.21583333333333332,
0.0175,
0.03194444444444444,
0.050555555555555555,
0.22555555555555556,
0.0275,
0.04722222222222222,
0.03083333333333... | [
0,
0.9952777777777778,
1.6288888888888888,
1.8802777777777777,
2.1241666666666665,
2.1375,
2.2058333333333335,
2.2919444444444443,
2.5077777777777777,
2.5252777777777777,
2.5572222222222223,
2.6077777777777778,
2.8333333333333335,
2.8608333333333333,
2.9080555555555554,
2.9388888888888... | 0 | 3 | 114 |
[
"2015-12-11T00:01:29",
"2015-12-11T00:36:02",
"2015-12-11T00:55:01",
"2015-12-11T01:05:27",
"2015-12-11T01:14:20",
"2015-12-11T01:15:42",
"2015-12-11T01:29:52",
"2015-12-11T02:00:40",
"2015-12-11T02:10:29",
"2015-12-11T02:26:33",
"2015-12-11T02:34:58",
"2015-12-11T03:02:34",
"2015-12-11T03:0... | [
0,
0,
0,
0,
0,
2,
0,
0,
0,
1,
2,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
1,
1,
1,
1,
2,
2,
2,
1,
1,
0,
0,
0,
2,
0,
0,
0,
2,
0,
0,
2,
2,
0,
2,
1,
2,
2,
0,
2,
0,
1,
0,
2,
2,
2,
0,
1,
2,
2,
2,
0,
1,
2... | [
0,
0.5758333333333333,
0.3163888888888889,
0.1738888888888889,
0.14805555555555555,
0.02277777777777778,
0.2361111111111111,
0.5133333333333333,
0.16361111111111112,
0.2677777777777778,
0.14027777777777778,
0.46,
0.10944444444444444,
0.10777777777777778,
0.014444444444444444,
0.1775,
... | [
0,
0.5758333333333333,
0.8922222222222222,
1.066111111111111,
1.2141666666666666,
1.2369444444444444,
1.4730555555555556,
1.986388888888889,
2.15,
2.417777777777778,
2.5580555555555557,
3.0180555555555557,
3.1275,
3.2352777777777777,
3.249722222222222,
3.4272222222222224,
3.765555555... | 1 | 3 | 391 |
[
"2015-12-12T00:15:12",
"2015-12-12T00:25:14",
"2015-12-12T00:29:36",
"2015-12-12T00:36:21",
"2015-12-12T00:46:59",
"2015-12-12T00:51:53",
"2015-12-12T00:59:34",
"2015-12-12T01:04:03",
"2015-12-12T01:06:34",
"2015-12-12T01:09:57",
"2015-12-12T01:33:27",
"2015-12-12T01:46:20",
"2015-12-12T02:0... | [
0,
0,
1,
2,
0,
1,
2,
0,
0,
0,
0,
0,
2,
0,
0,
0,
0,
1,
0,
1,
2,
1,
0,
2,
0,
2,
1,
0,
2,
0,
0,
0,
1,
0,
2,
0,
2,
2,
2,
0,
0,
2,
0,
1,
0,
2,
2,
0,
2,
1,
0,
0,
0,
2,
0,
0,
0,
1,
1,
0,
2,
1,
0,
1... | [
0,
0.16722222222222222,
0.07277777777777777,
0.1125,
0.17722222222222223,
0.08166666666666667,
0.12805555555555556,
0.07472222222222222,
0.041944444444444444,
0.05638888888888889,
0.39166666666666666,
0.21472222222222223,
0.2722222222222222,
0.08805555555555555,
0.44055555555555553,
0.... | [
0,
0.16722222222222222,
0.24,
0.3525,
0.5297222222222222,
0.6113888888888889,
0.7394444444444445,
0.8141666666666667,
0.8561111111111112,
0.9125,
1.3041666666666667,
1.518888888888889,
1.791111111111111,
1.8791666666666667,
2.319722222222222,
2.3294444444444444,
2.539722222222222,
... | 2 | 3 | 402 |
[
"2015-12-13T00:06:54",
"2015-12-13T00:17:25",
"2015-12-13T00:17:59",
"2015-12-13T00:22:03",
"2015-12-13T00:22:03",
"2015-12-13T00:23:41",
"2015-12-13T00:35:23",
"2015-12-13T00:51:07",
"2015-12-13T00:51:36",
"2015-12-13T01:09:52",
"2015-12-13T01:27:38",
"2015-12-13T01:38:40",
"2015-12-13T01:5... | [
0,
0,
0,
0,
0,
0,
0,
0,
2,
1,
0,
0,
0,
0,
0,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
1,
1,
0,
0,
2,
0,
0,
0,
0,
0,
0,
2,
2,
2,
2,
0,
1,
0,
0,
1,
1,
0,
0,
2,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
0,
0.17527777777777778,
0.009444444444444445,
0.06777777777777778,
0,
0.02722222222222222,
0.195,
0.26222222222222225,
0.008055555555555555,
0.30444444444444446,
0.2961111111111111,
0.18388888888888888,
0.2238888888888889,
0.010277777777777778,
0.19583333333333333,
0.0811111111111111,
... | [
0,
0.17527777777777778,
0.18472222222222223,
0.2525,
0.2525,
0.2797222222222222,
0.4747222222222222,
0.7369444444444444,
0.745,
1.0494444444444444,
1.3455555555555556,
1.5294444444444444,
1.7533333333333334,
1.763611111111111,
1.9594444444444445,
2.0405555555555557,
2.451111111111111... | 3 | 3 | 316 |
[
"2015-12-14T00:12:49",
"2015-12-14T00:22:14",
"2015-12-14T00:38:13",
"2015-12-14T00:43:45",
"2015-12-14T01:04:35",
"2015-12-14T01:06:34",
"2015-12-14T01:07:42",
"2015-12-14T01:17:28",
"2015-12-14T01:28:24",
"2015-12-14T01:33:31",
"2015-12-14T01:35:22",
"2015-12-14T01:35:22",
"2015-12-14T01:4... | [
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
2,
2,
2,
2,
2,
0,
0,
2,
2,
2,
0,
0,
1,
2,
2,
2,
2,
2,
0,
0,
0,
1,
0,
0,
1,
2,
2,
1,
1,
0,
2,
2,
0,
2,
2... | [
0,
0.15694444444444444,
0.2663888888888889,
0.09222222222222222,
0.3472222222222222,
0.03305555555555555,
0.01888888888888889,
0.16277777777777777,
0.18222222222222223,
0.08527777777777777,
0.030833333333333334,
0,
0.09111111111111111,
0.013333333333333334,
0.10555555555555556,
0.45833... | [
0,
0.15694444444444444,
0.42333333333333334,
0.5155555555555555,
0.8627777777777778,
0.8958333333333334,
0.9147222222222222,
1.0775,
1.2597222222222222,
1.345,
1.3758333333333332,
1.3758333333333332,
1.4669444444444444,
1.4802777777777778,
1.5858333333333334,
2.0441666666666665,
2.20... | 4 | 3 | 444 |
[
"2015-12-15T00:08:18",
"2015-12-15T00:15:28",
"2015-12-15T00:45:21",
"2015-12-15T01:07:16",
"2015-12-15T01:08:49",
"2015-12-15T01:10:45",
"2015-12-15T01:17:58",
"2015-12-15T01:30:27",
"2015-12-15T01:40:11",
"2015-12-15T01:42:27",
"2015-12-15T01:44:06",
"2015-12-15T01:45:48",
"2015-12-15T01:4... | [
0,
0,
0,
0,
0,
0,
0,
2,
0,
0,
0,
2,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
2,
0,
0,
2,
0,
0,
0,
0,
0,
0,
0,
0,
2,
0,
2,
0,
2,
0,
2,
2,
2,
2,
1,
0,
0,
2,
1,
2,
0,
2,
2,
0,
2,
2,
2,
0,
2,
2... | [
0,
0.11944444444444445,
0.49805555555555553,
0.36527777777777776,
0.025833333333333333,
0.03222222222222222,
0.12027777777777778,
0.20805555555555555,
0.1622222222222222,
0.03777777777777778,
0.0275,
0.028333333333333332,
0.059444444444444446,
0.17833333333333334,
0.11833333333333333,
... | [
0,
0.11944444444444445,
0.6175,
0.9827777777777778,
1.0086111111111111,
1.0408333333333333,
1.1611111111111112,
1.3691666666666666,
1.531388888888889,
1.5691666666666666,
1.5966666666666667,
1.625,
1.6844444444444444,
1.8627777777777779,
1.981111111111111,
2.2119444444444443,
2.9375,... | 5 | 3 | 419 |
[
"2015-12-16T00:05:47",
"2015-12-16T00:06:03",
"2015-12-16T00:14:25",
"2015-12-16T00:50:05",
"2015-12-16T01:07:12",
"2015-12-16T01:19:10",
"2015-12-16T01:41:36",
"2015-12-16T01:42:42",
"2015-12-16T01:50:37",
"2015-12-16T01:54:29",
"2015-12-16T02:16:11",
"2015-12-16T02:24:06",
"2015-12-16T02:3... | [
1,
2,
2,
2,
2,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
2,
0,
2,
0,
0,
0,
0,
0,
1,
0,
2,
2,
2,
0,
0,
0,
1,
0,
1,
0,
0,
0,
0,
0,
0,
2,
1,
0,
2,
0,
2,
0,
1,
2,
2,
0,
2,
2,
0,
0,
0,
2,
0,
0,
0,
1,
2,
2,
0... | [
0,
0.0044444444444444444,
0.13944444444444445,
0.5944444444444444,
0.2852777777777778,
0.19944444444444445,
0.3738888888888889,
0.018333333333333333,
0.13194444444444445,
0.06444444444444444,
0.3616666666666667,
0.13194444444444445,
0.1175,
0.06888888888888889,
0.12222222222222222,
0.0... | [
0,
0.0044444444444444444,
0.1438888888888889,
0.7383333333333333,
1.023611111111111,
1.2230555555555556,
1.5969444444444445,
1.6152777777777778,
1.7472222222222222,
1.8116666666666668,
2.1733333333333333,
2.305277777777778,
2.4227777777777777,
2.4916666666666667,
2.613888888888889,
2.6... | 6 | 3 | 376 |
[
"2015-12-17T00:12:04",
"2015-12-17T00:15:16",
"2015-12-17T00:20:57",
"2015-12-17T00:23:13",
"2015-12-17T00:24:10",
"2015-12-17T00:26:23",
"2015-12-17T00:27:00",
"2015-12-17T00:38:34",
"2015-12-17T00:45:43",
"2015-12-17T01:04:16",
"2015-12-17T01:16:57",
"2015-12-17T01:18:25",
"2015-12-17T01:4... | [
0,
2,
0,
0,
0,
1,
0,
2,
0,
0,
2,
0,
0,
2,
1,
0,
2,
0,
0,
2,
0,
0,
0,
0,
2,
0,
1,
2,
0,
0,
0,
0,
2,
0,
1,
1,
2,
2,
1,
2,
0,
2,
2,
2,
1,
2,
1,
0,
0,
2,
2,
0,
2,
0,
2,
0,
0,
2,
2,
2,
0,
0,
2,
2... | [
0,
0.05333333333333334,
0.09472222222222222,
0.03777777777777778,
0.015833333333333335,
0.036944444444444446,
0.010277777777777778,
0.19277777777777777,
0.11916666666666667,
0.30916666666666665,
0.21138888888888888,
0.024444444444444446,
0.4702777777777778,
0.0025,
0.009722222222222222,
... | [
0,
0.05333333333333334,
0.14805555555555555,
0.18583333333333332,
0.20166666666666666,
0.2386111111111111,
0.24888888888888888,
0.44166666666666665,
0.5608333333333333,
0.87,
1.081388888888889,
1.1058333333333332,
1.576111111111111,
1.5786111111111112,
1.5883333333333334,
1.59083333333... | 7 | 3 | 387 |
[
"2015-12-18T05:27:31",
"2015-12-18T06:34:13",
"2015-12-18T06:41:53",
"2015-12-18T07:08:17",
"2015-12-18T07:15:13",
"2015-12-18T07:21:07",
"2015-12-18T07:28:03",
"2015-12-18T07:32:55",
"2015-12-18T07:40:21",
"2015-12-18T07:46:44",
"2015-12-18T07:47:38",
"2015-12-18T07:53:36",
"2015-12-18T07:5... | [
2,
0,
0,
0,
2,
2,
0,
0,
2,
1,
0,
2,
2,
0,
2,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
2,
2,
0,
1,
0,
2,
2,
0,
0,
0,
0,
0,
0,
0,
2,
2,
2,
2,
0,
0,
2,
0,
0,
0,
0,
0,
0,
2,
0,
1,
1,
0,
0,
0,
1,
2,
0,
1... | [
0,
1.1116666666666666,
0.12777777777777777,
0.44,
0.11555555555555555,
0.09833333333333333,
0.11555555555555555,
0.0811111111111111,
0.1238888888888889,
0.1063888888888889,
0.015,
0.09944444444444445,
0,
0.049444444444444444,
0.095,
0.030555555555555555,
0.21305555555555555,
0.0583... | [
0,
1.1116666666666666,
1.2394444444444443,
1.6794444444444445,
1.795,
1.8933333333333333,
2.008888888888889,
2.09,
2.213888888888889,
2.3202777777777777,
2.335277777777778,
2.4347222222222222,
2.4347222222222222,
2.484166666666667,
2.5791666666666666,
2.609722222222222,
2.82277777777... | 8 | 3 | 346 |
["2015-12-19T06:08:04","2015-12-19T06:08:04","2015-12-19T08:08:53","2015-12-19T08:19:40","2015-12-19(...TRUNCATED) | [1,1,0,0,0,0,2,0,2,0,2,0,2,0,0,0,2,0,2,2,2,0,0,2,0,2,2,2,0,2,2,0,2,2,0,1,0,0,1,0,2,2,2,2,0,0,2,2,1,0(...TRUNCATED) | [0.0,0.0,2.013611111111111,0.17972222222222223,0.009166666666666667,0.1597222222222222,0.16611111111(...TRUNCATED) | [0.0,0.0,2.013611111111111,2.1933333333333334,2.2025,2.362222222222222,2.5283333333333333,2.55527777(...TRUNCATED) | 9 | 3 | 272 |
MUSES: a benchmark for Marked Unevenly Spaced Event Sequences
MUSES, a benchmark for Marked Unevenly Spaced Event Sequences, is a collection of unevenly spaced time series datasets from various domains, containing marked events for training and evaluating prediction approaches.
Languages
All the columns and classes (when textual) in MUSES are in English (BCP-47 en)
Dataset Structure
Data Instances
All datasets formatting follows a similar structure.
{
"datetime": [2020-06-10T14:27:37.963000000, 2020-06-20T21:23:06.980000000, 2020-10-15T13:36:16.687000000],
"type_event": [7, 2, 15],
"time_since_last_event": [0.0, 246.92472694444444, 2800.2193630555557],
"time_since_start": [0.0, 246.92472694444444, 3047.14409],
"seq_idx": 1,
"dim_process": 50,
"seq_len": 3
}
earthquake
- Size of downloaded dataset files: 2.77 MB
- Size of the generated dataset: 3.06 MB
- Total amount of disk used: 5.83 MB
memetrack
- Size of downloaded dataset files: 569.01 MB
- Size of the generated dataset: 993.72 MB
- Total amount of disk used: 1562.72 MB
stackoverflow
- Size of downloaded dataset files: 110.71 MB
- Size of the generated dataset: 143.91 MB
- Total amount of disk used: 254.62 MB
taxi_nyc_neighborhoods
- Size of downloaded dataset files: 968.44 MB
- Size of the generated dataset: 2031.69 MB
- Total amount of disk used: 3000.14 MB
synthea
- Size of downloaded dataset files: 7.65 MB
- Size of the generated dataset: 14.56 MB
- Total amount of disk used: 22.21 MB
spiketrains
- Size of downloaded dataset files: 8.14 MB
- Size of the generated dataset: 20.57 MB
- Total amount of disk used: 28.71 MB
crypto_transactions
- Size of downloaded dataset files: 393.35 MB
- Size of the generated dataset: 748.14 MB
- Total amount of disk used: 1141.49 MB
human_activity
- Size of downloaded dataset files: 0.20 MB
- Size of the generated dataset: 0.22 MB
- Total amount of disk used: 0.42 MB
911
- Size of downloaded dataset files: 8.42 MB
- Size of the generated dataset: 21.63 MB
- Total amount of disk used: 30.05 MB
mooc
- Size of downloaded dataset files: 3.97 MB
- Size of the generated dataset: 10.29 MB
- Total amount of disk used: 14.26 MB
amazon_easytpp
- Size of downloaded dataset files: 7.81 MB
- Size of the generated dataset: 10.41 MB
- Total amount of disk used: 18.22 MB
wikipedia
- Size of downloaded dataset files: 127.20 MB
- Size of the generated dataset: 238.24 MB
- Total amount of disk used: 365.43 MB
retweet_easytpp
- Size of downloaded dataset files: 15.12 MB
- Size of the generated dataset: 64.49 MB
- Total amount of disk used: 79.62 MB
taobao_easytpp
- Size of downloaded dataset files: 1.60 MB
- Size of the generated dataset: 2.84 MB
- Total amount of disk used: 4.43 MB
taxi_easytpp
- Size of downloaded dataset files: 0.59 MB
- Size of the generated dataset: 1.88 MB
- Total amount of disk used: 2.47 MB
volcano_easytpp
- Size of downloaded dataset files: 0.15 MB
- Size of the generated dataset: 0.23 MB
- Total amount of disk used: 0.38 MB
hawkes_dependent
- Size of downloaded dataset files: 11.10 MB
- Size of the generated dataset: 15.84 MB
- Total amount of disk used: 26.94 MB
hawkes_1
- Size of downloaded dataset files: 2.04 MB
- Size of the generated dataset: 2.49 MB
- Total amount of disk used: 4.54 MB
Data Fields
The data fields are the same among all datasets and splits. The datetime column is optional and left out when no data is available.
type_event: a classification label (int64) sequence.time_since_last_event: afloat64sequence of inter-event times.time_since_start: afloat64sequence of event timestamps since sequence start.seq_idx: anint64feature uniquely identifying sequences in the split.dim_process: anint64feature representing total number of classes in the dataset.seq_len: anint64feature containing sequence length.datetime: (if avail.) atimestamp[ns]sequence, containing event datetime if available.
The classification labels differ for each dataset. The corresponding labels for each event type can be read from metadata like this
>>> import datasets as ds
>>> meta_data = ds.load_dataset_builder("ddrg/MUSES", "DATASET-NAME")
>>> [label for label in enumerate(meta_data.info.features["type_event"].feature.names)]
Data Splits
| train | validation | test | |
|---|---|---|---|
| earthquake | 1724 | 216 | 215 |
| memetrack | 1264794 | 158100 | 158099 |
| stackoverflow | 42838 | 5354 | 5355 |
| synthea | 9198 | 1150 | 1149 |
| spiketrains | 480 | 60 | 60 |
| crypto_transactions | 1408527 | 176066 | 176066 |
| human_activity | 175 | 22 | 22 |
| 911 | 1355 | 169 | 170 |
| mooc | 5638 | 704 | 705 |
| amazon_easytpp | 7382 | 922 | 923 |
| wikipedia | 907765 | 113470 | 113471 |
| retweet_easytpp | 19200 | 2400 | 2400 |
| taobao_easytpp | 1600 | 200 | 200 |
| taxi_easytpp | 1600 | 200 | 200 |
| volcano_easytpp | 333 | 42 | 41 |
| hawkes_dependent | 19661 | 2458 | 2457 |
| hawkes_1 | 800 | 100 | 100 |
taxi_nyc_neighborhoods
| train | validation | test | extra | |
|---|---|---|---|---|
| taxi_nyc_neighborhoods | 273060 | 34132 | 34133 | 106020 |
Dataset Creation
Curation Rationale
A dataset has to comply with the following hard requirements to be eligible for MUSES.
- Inherently unevenly spaced
- Time series minimum length of 2
- Known license
- Used in Literature
- Replicably described data collection and processing
- Dataset has practical relevance or is commonly used for evaluation
- Maximum of 100 M events
- Nominal classes or common class analogy
- Commonly used in unevenly spaced time or TPP forecasting
Further, we aim to fulfull the following variety constraints by dataset composition.
- Coverage of many different domains
- Variety in dataset sizes
- Variety in sequence lengths
- Variety in problem difficulties
- Variety in problem complexity
- Variety in class balances
See our paper for further information.
Source Data
Initial Data Collection and Normalization
Data was sourced and preprocessed like in other works, unless explicitly stated otherwise in our paper. See Citation Information for details.
Personal and Sensitive Information
Data is deidentified or all personal and sensitive information was removed.
Considerations for Using the Data
Discussion of Biases
There are no ethical biases we are aware of.
Other Known Limitations
Class imbalances are common in event sequences.
See the following paper for further information.
TODO JIMMY
Additional Information
Dataset Curators
Licensing Information
The primary MUSES tasks are built on and derived from existing datasets. We refer users to the original licenses accompanying each dataset.
| License | Other | |
|---|---|---|
| earthquake | Public Domain | |
| memetrack | 3-Clause BSD | see LICENSE/memetrack |
| stackoverflow | CC BY-SA | see LICENSE/stackoverflow |
| taxi_nyc_neighborhoods | Public Domain | |
| synthea | Apache 2.0 | see LICENSE/synthea |
| spiketrains | CC0 | see LICENSE/spiketrains |
| crypto_transactions | 3-Clause BSD | see LICENSE/crypto_transactions |
| human_activity | CC-BY | see LICENSE/human_activity |
| 911 | ODC ODbL | see LICENSE/911 |
| mooc | Scientific Use Only | see LICENSE/mooc |
| amazon_easytpp | Apache 2.0 | see LICENSE/amazon_easytpp |
| wikipedia | CC BY-SA | see LICENSE/wikipedia |
| retweet_easytpp | Apache 2.0 | see LICENSE/retweet_easytpp |
| taobao_easytpp | Apache 2.0 | see LICENSE/taobao_easytpp |
| taxi_easytpp | Apache 2.0 | see LICENSE/taxi_easytpp |
| volcano_easytpp | Apache 2.0 | see LICENSE/volcano_easytpp |
| hawkes_dependent | Public Domain | |
| hawkes_1 | Public Domain |
Citation Information
We encourage you to use the following BibTeX citation for MUSES itself:
# TODO @Jimmy Add here Final Paper
If you use MUSES, please also cite all the individual datasets you use, both to give the original authors their due credit and because venues will expect papers to describe the data they evaluate on. The following provides BibTeX for all of the MUSES tasks.
earthquake
@article{stockman2024earthquakenpp,
title={EarthquakeNPP: A Benchmark for Earthquake Forecasting with Neural Point Processes},
author={Stockman, Samuel and Lawson, Daniel and Werner, Maximilian},
journal={arXiv preprint arXiv:2410.08226},
year={2024}
}
memetrack
@inproceedings{leskovec2009meme,
title={Meme-tracking and the dynamics of the news cycle},
author={Leskovec, Jure and Backstrom, Lars and Kleinberg, Jon},
booktitle={Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining},
pages={497--506},
year={2009}
}
@article{mei2017neural,
title={The neural hawkes process: A neurally self-modulating multivariate point process},
author={Mei, Hongyuan and Eisner, Jason M},
journal={Advances in neural information processing systems},
volume={30},
year={2017}
}
stackoverflow
@inproceedings{du2016recurrent,
title={Recurrent marked temporal point processes: Embedding event history to vector},
author={Du, Nan and Dai, Hanjun and Trivedi, Rakshit and Upadhyay, Utkarsh and Gomez-Rodriguez, Manuel and Song, Le},
booktitle={Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
pages={1555--1564},
year={2016}
}
taxi_nyc_neighborhoods
@inproceedings{du2016recurrent,
title={Recurrent marked temporal point processes: Embedding event history to vector},
author={Du, Nan and Dai, Hanjun and Trivedi, Rakshit and Upadhyay, Utkarsh and Gomez-Rodriguez, Manuel and Song, Le},
booktitle={Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
pages={1555--1564},
year={2016}
}
synthea
@article{walonoski2018synthea,
title={Synthea: An approach, method, and software mechanism for generating synthetic patients and the synthetic electronic health care record},
author={Walonoski, Jason and Kramer, Mark and Nichols, Joseph and Quina, Andre and Moesel, Chris and Hall, Dylan and Duffett, Carlton and Dube, Kudakwashe and Gallagher, Thomas and McLachlan, Scott},
journal={Journal of the American Medical Informatics Association},
volume={25},
number={3},
pages={230--238},
year={2018},
publisher={Oxford University Press}
}
@inproceedings{enguehard2020neural,
title={Neural temporal point processes for modelling electronic health records},
author={Enguehard, Joseph and Busbridge, Dan and Bozson, Adam and Woodcock, Claire and Hammerla, Nils},
booktitle={Machine Learning for Health},
pages={85--113},
year={2020},
organization={PMLR}
}
spiketrains
@article{stetter2012model,
title={Model-free reconstruction of excitatory neuronal connectivity from calcium imaging signals},
author={Stetter, Olav and Battaglia, Demian and Soriano, Jordi and Geisel, Theo},
year={2012},
publisher={Public Library of Science San Francisco, USA}
}
@article{linderman2015scalable,
title={Scalable bayesian inference for excitatory point process networks},
author={Linderman, Scott W and Adams, Ryan P},
journal={arXiv preprint arXiv:1507.03228},
year={2015}
}
crypto_transactions
@article{shamsi2022chartalist,
title={Chartalist: Labeled graph datasets for utxo and account-based blockchains},
author={Shamsi, Kiarash and Victor, Friedhelm and Kantarcioglu, Murat and Gel, Yulia and Akcora, Cuneyt G},
journal={Advances in Neural Information Processing Systems},
volume={35},
pages={34926--34939},
year={2022}
}
@inproceedings{du2016recurrent,
title={Recurrent marked temporal point processes: Embedding event history to vector},
author={Du, Nan and Dai, Hanjun and Trivedi, Rakshit and Upadhyay, Utkarsh and Gomez-Rodriguez, Manuel and Song, Le},
booktitle={Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
pages={1555--1564},
year={2016}
}
human_activity
@article{cook2012casas,
title={CASAS: A smart home in a box},
author={Cook, Diane J and Crandall, Aaron S and Thomas, Brian L and Krishnan, Narayanan C},
journal={Computer},
volume={46},
number={7},
pages={62--69},
year={2012},
publisher={IEEE}
}
@article{fortino2021predicting,
title={Predicting activities of daily living via temporal point processes: Approaches and experimental results},
author={Fortino, Giancarlo and Guzzo, Antonella and Ianni, Michele and Leotta, Francesco and Mecella, Massimo},
journal={Computers \& Electrical Engineering},
volume={96},
pages={107567},
year={2021},
publisher={Elsevier}
}
911
@inproceedings{zuo2020transformer,
title={Transformer hawkes process},
author={Zuo, Simiao and Jiang, Haoming and Li, Zichong and Zhao, Tuo and Zha, Hongyuan},
booktitle={International conference on machine learning},
pages={11692--11702},
year={2020},
organization={PMLR}
}
mooc
@inproceedings{kumar2019predicting,
title={Predicting dynamic embedding trajectory in temporal interaction networks},
author={Kumar, Srijan and Zhang, Xikun and Leskovec, Jure},
booktitle={Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery \& data mining},
pages={1269--1278},
year={2019}
}
@article{shchur2019intensity,
title={Intensity-free learning of temporal point processes},
author={Shchur, Oleksandr and Bilo{\v{s}}, Marin and G{\"u}nnemann, Stephan},
journal={arXiv preprint arXiv:1909.12127},
year={2019}
}
amazon_easytpp
@inproceedings{ni2019justifying,
title={Justifying recommendations using distantly-labeled reviews and fine-grained aspects},
author={Ni, Jianmo and Li, Jiacheng and McAuley, Julian},
booktitle={Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP)},
pages={188--197},
year={2019}
}
@article{xue2023easytpp,
title={Easytpp: Towards open benchmarking temporal point processes},
author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
journal={arXiv preprint arXiv:2307.08097},
year={2023}
}
wikipedia
@inproceedings{kumar2019predicting,
title={Predicting dynamic embedding trajectory in temporal interaction networks},
author={Kumar, Srijan and Zhang, Xikun and Leskovec, Jure},
booktitle={Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery \& data mining},
pages={1269--1278},
year={2019}
}
@article{bosser2023predictive,
title={On the predictive accuracy of neural temporal point process models for continuous-time event data},
author={Bosser, Tanguy and Taieb, Souhaib Ben},
journal={arXiv preprint arXiv:2306.17066},
year={2023}
}
retweet_easytpp
@inproceedings{zhao2015seismic,
title={Seismic: A self-exciting point process model for predicting tweet popularity},
author={Zhao, Qingyuan and Erdogdu, Murat A and He, Hera Y and Rajaraman, Anand and Leskovec, Jure},
booktitle={Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining},
pages={1513--1522},
year={2015}
}
@article{xue2023easytpp,
title={Easytpp: Towards open benchmarking temporal point processes},
author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
journal={arXiv preprint arXiv:2307.08097},
year={2023}
}
taobao_easytpp
@article{xue2022hypro,
title={Hypro: A hybridly normalized probabilistic model for long-horizon prediction of event sequences},
author={Xue, Siqiao and Shi, Xiaoming and Zhang, James and Mei, Hongyuan},
journal={Advances in Neural Information Processing Systems},
volume={35},
pages={34641--34650},
year={2022}
}
@article{xue2023easytpp,
title={Easytpp: Towards open benchmarking temporal point processes},
author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
journal={arXiv preprint arXiv:2307.08097},
year={2023}
}
taxi_easytpp
@article{xue2023easytpp,
title={Easytpp: Towards open benchmarking temporal point processes},
author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
journal={arXiv preprint arXiv:2307.08097},
year={2023}
}
volcano_easytpp
@article{bebbington2014long,
title={Long-term forecasting of volcanic explosivity},
author={Bebbington, MS},
journal={Geophysical Journal International},
volume={197},
number={3},
pages={1500--1515},
year={2014},
publisher={Oxford University Press}
}
@misc{easytpp-github-volcano,
title = {EasyTemporalPointProcess GitHub Volcano Preprocessing},
author = {Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
url = {https://github.com/ant-research/EasyTemporalPointProcess/blob/main/examples/script_data_processing/volcano.py}
}
hawkes_dependent
@inproceedings{enguehard2020neural,
title={Neural temporal point processes for modelling electronic health records},
author={Enguehard, Joseph and Busbridge, Dan and Bozson, Adam and Woodcock, Claire and Hammerla, Nils},
booktitle={Machine Learning for Health},
pages={85--113},
year={2020},
organization={PMLR}
}
hawkes_1
@article{omi2019fully,
title={Fully neural network based model for general temporal point processes},
author={Omi, Takahiro and Aihara, Kazuyuki and others},
journal={Advances in neural information processing systems},
volume={32},
year={2019}
}
Contributions
Thanks to @Moritz Tschöpe and @Jimmy Pöhlmann for adding this dataset.
- Downloads last month
- 8,744