Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
datetime
listlengths
114
2.19k
type_event
listlengths
114
2.19k
time_since_last_event
listlengths
114
2.19k
time_since_start
listlengths
114
2.19k
seq_idx
int64
0
1.35k
dim_process
int64
3
3
seq_len
int64
114
2.19k
[ "2015-12-10T14:39:21", "2015-12-10T15:39:04", "2015-12-10T16:17:05", "2015-12-10T16:32:10", "2015-12-10T16:46:48", "2015-12-10T16:47:36", "2015-12-10T16:51:42", "2015-12-10T16:56:52", "2015-12-10T17:09:49", "2015-12-10T17:10:52", "2015-12-10T17:12:47", "2015-12-10T17:15:49", "2015-12-10T17:2...
[ 1, 0, 0, 2, 0, 0, 0, 0, 2, 0, 2, 2, 0, 2, 2, 2, 0, 0, 0, 2, 2, 2, 2, 1, 2, 0, 0, 1, 2, 2, 0, 2, 0, 2, 2, 2, 2, 0, 2, 1, 2, 0, 0, 2, 1, 0, 2, 2, 1, 0, 2, 2, 0, 2, 1, 0, 0, 0, 0, 2, 0, 0, 0, 0...
[ 0, 0.9952777777777778, 0.6336111111111111, 0.2513888888888889, 0.24388888888888888, 0.013333333333333334, 0.06833333333333333, 0.08611111111111111, 0.21583333333333332, 0.0175, 0.03194444444444444, 0.050555555555555555, 0.22555555555555556, 0.0275, 0.04722222222222222, 0.03083333333333...
[ 0, 0.9952777777777778, 1.6288888888888888, 1.8802777777777777, 2.1241666666666665, 2.1375, 2.2058333333333335, 2.2919444444444443, 2.5077777777777777, 2.5252777777777777, 2.5572222222222223, 2.6077777777777778, 2.8333333333333335, 2.8608333333333333, 2.9080555555555554, 2.9388888888888...
0
3
114
[ "2015-12-11T00:01:29", "2015-12-11T00:36:02", "2015-12-11T00:55:01", "2015-12-11T01:05:27", "2015-12-11T01:14:20", "2015-12-11T01:15:42", "2015-12-11T01:29:52", "2015-12-11T02:00:40", "2015-12-11T02:10:29", "2015-12-11T02:26:33", "2015-12-11T02:34:58", "2015-12-11T03:02:34", "2015-12-11T03:0...
[ 0, 0, 0, 0, 0, 2, 0, 0, 0, 1, 2, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 1, 1, 0, 0, 0, 2, 0, 0, 0, 2, 0, 0, 2, 2, 0, 2, 1, 2, 2, 0, 2, 0, 1, 0, 2, 2, 2, 0, 1, 2, 2, 2, 0, 1, 2...
[ 0, 0.5758333333333333, 0.3163888888888889, 0.1738888888888889, 0.14805555555555555, 0.02277777777777778, 0.2361111111111111, 0.5133333333333333, 0.16361111111111112, 0.2677777777777778, 0.14027777777777778, 0.46, 0.10944444444444444, 0.10777777777777778, 0.014444444444444444, 0.1775, ...
[ 0, 0.5758333333333333, 0.8922222222222222, 1.066111111111111, 1.2141666666666666, 1.2369444444444444, 1.4730555555555556, 1.986388888888889, 2.15, 2.417777777777778, 2.5580555555555557, 3.0180555555555557, 3.1275, 3.2352777777777777, 3.249722222222222, 3.4272222222222224, 3.765555555...
1
3
391
[ "2015-12-12T00:15:12", "2015-12-12T00:25:14", "2015-12-12T00:29:36", "2015-12-12T00:36:21", "2015-12-12T00:46:59", "2015-12-12T00:51:53", "2015-12-12T00:59:34", "2015-12-12T01:04:03", "2015-12-12T01:06:34", "2015-12-12T01:09:57", "2015-12-12T01:33:27", "2015-12-12T01:46:20", "2015-12-12T02:0...
[ 0, 0, 1, 2, 0, 1, 2, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 1, 0, 1, 2, 1, 0, 2, 0, 2, 1, 0, 2, 0, 0, 0, 1, 0, 2, 0, 2, 2, 2, 0, 0, 2, 0, 1, 0, 2, 2, 0, 2, 1, 0, 0, 0, 2, 0, 0, 0, 1, 1, 0, 2, 1, 0, 1...
[ 0, 0.16722222222222222, 0.07277777777777777, 0.1125, 0.17722222222222223, 0.08166666666666667, 0.12805555555555556, 0.07472222222222222, 0.041944444444444444, 0.05638888888888889, 0.39166666666666666, 0.21472222222222223, 0.2722222222222222, 0.08805555555555555, 0.44055555555555553, 0....
[ 0, 0.16722222222222222, 0.24, 0.3525, 0.5297222222222222, 0.6113888888888889, 0.7394444444444445, 0.8141666666666667, 0.8561111111111112, 0.9125, 1.3041666666666667, 1.518888888888889, 1.791111111111111, 1.8791666666666667, 2.319722222222222, 2.3294444444444444, 2.539722222222222, ...
2
3
402
[ "2015-12-13T00:06:54", "2015-12-13T00:17:25", "2015-12-13T00:17:59", "2015-12-13T00:22:03", "2015-12-13T00:22:03", "2015-12-13T00:23:41", "2015-12-13T00:35:23", "2015-12-13T00:51:07", "2015-12-13T00:51:36", "2015-12-13T01:09:52", "2015-12-13T01:27:38", "2015-12-13T01:38:40", "2015-12-13T01:5...
[ 0, 0, 0, 0, 0, 0, 0, 0, 2, 1, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 0, 2, 0, 0, 0, 0, 0, 0, 2, 2, 2, 2, 0, 1, 0, 0, 1, 1, 0, 0, 2, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0...
[ 0, 0.17527777777777778, 0.009444444444444445, 0.06777777777777778, 0, 0.02722222222222222, 0.195, 0.26222222222222225, 0.008055555555555555, 0.30444444444444446, 0.2961111111111111, 0.18388888888888888, 0.2238888888888889, 0.010277777777777778, 0.19583333333333333, 0.0811111111111111, ...
[ 0, 0.17527777777777778, 0.18472222222222223, 0.2525, 0.2525, 0.2797222222222222, 0.4747222222222222, 0.7369444444444444, 0.745, 1.0494444444444444, 1.3455555555555556, 1.5294444444444444, 1.7533333333333334, 1.763611111111111, 1.9594444444444445, 2.0405555555555557, 2.451111111111111...
3
3
316
[ "2015-12-14T00:12:49", "2015-12-14T00:22:14", "2015-12-14T00:38:13", "2015-12-14T00:43:45", "2015-12-14T01:04:35", "2015-12-14T01:06:34", "2015-12-14T01:07:42", "2015-12-14T01:17:28", "2015-12-14T01:28:24", "2015-12-14T01:33:31", "2015-12-14T01:35:22", "2015-12-14T01:35:22", "2015-12-14T01:4...
[ 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 2, 2, 2, 2, 2, 0, 0, 2, 2, 2, 0, 0, 1, 2, 2, 2, 2, 2, 0, 0, 0, 1, 0, 0, 1, 2, 2, 1, 1, 0, 2, 2, 0, 2, 2...
[ 0, 0.15694444444444444, 0.2663888888888889, 0.09222222222222222, 0.3472222222222222, 0.03305555555555555, 0.01888888888888889, 0.16277777777777777, 0.18222222222222223, 0.08527777777777777, 0.030833333333333334, 0, 0.09111111111111111, 0.013333333333333334, 0.10555555555555556, 0.45833...
[ 0, 0.15694444444444444, 0.42333333333333334, 0.5155555555555555, 0.8627777777777778, 0.8958333333333334, 0.9147222222222222, 1.0775, 1.2597222222222222, 1.345, 1.3758333333333332, 1.3758333333333332, 1.4669444444444444, 1.4802777777777778, 1.5858333333333334, 2.0441666666666665, 2.20...
4
3
444
[ "2015-12-15T00:08:18", "2015-12-15T00:15:28", "2015-12-15T00:45:21", "2015-12-15T01:07:16", "2015-12-15T01:08:49", "2015-12-15T01:10:45", "2015-12-15T01:17:58", "2015-12-15T01:30:27", "2015-12-15T01:40:11", "2015-12-15T01:42:27", "2015-12-15T01:44:06", "2015-12-15T01:45:48", "2015-12-15T01:4...
[ 0, 0, 0, 0, 0, 0, 0, 2, 0, 0, 0, 2, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 2, 0, 2, 0, 2, 2, 2, 2, 1, 0, 0, 2, 1, 2, 0, 2, 2, 0, 2, 2, 2, 0, 2, 2...
[ 0, 0.11944444444444445, 0.49805555555555553, 0.36527777777777776, 0.025833333333333333, 0.03222222222222222, 0.12027777777777778, 0.20805555555555555, 0.1622222222222222, 0.03777777777777778, 0.0275, 0.028333333333333332, 0.059444444444444446, 0.17833333333333334, 0.11833333333333333, ...
[ 0, 0.11944444444444445, 0.6175, 0.9827777777777778, 1.0086111111111111, 1.0408333333333333, 1.1611111111111112, 1.3691666666666666, 1.531388888888889, 1.5691666666666666, 1.5966666666666667, 1.625, 1.6844444444444444, 1.8627777777777779, 1.981111111111111, 2.2119444444444443, 2.9375,...
5
3
419
[ "2015-12-16T00:05:47", "2015-12-16T00:06:03", "2015-12-16T00:14:25", "2015-12-16T00:50:05", "2015-12-16T01:07:12", "2015-12-16T01:19:10", "2015-12-16T01:41:36", "2015-12-16T01:42:42", "2015-12-16T01:50:37", "2015-12-16T01:54:29", "2015-12-16T02:16:11", "2015-12-16T02:24:06", "2015-12-16T02:3...
[ 1, 2, 2, 2, 2, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 2, 0, 2, 0, 0, 0, 0, 0, 1, 0, 2, 2, 2, 0, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 0, 2, 1, 0, 2, 0, 2, 0, 1, 2, 2, 0, 2, 2, 0, 0, 0, 2, 0, 0, 0, 1, 2, 2, 0...
[ 0, 0.0044444444444444444, 0.13944444444444445, 0.5944444444444444, 0.2852777777777778, 0.19944444444444445, 0.3738888888888889, 0.018333333333333333, 0.13194444444444445, 0.06444444444444444, 0.3616666666666667, 0.13194444444444445, 0.1175, 0.06888888888888889, 0.12222222222222222, 0.0...
[ 0, 0.0044444444444444444, 0.1438888888888889, 0.7383333333333333, 1.023611111111111, 1.2230555555555556, 1.5969444444444445, 1.6152777777777778, 1.7472222222222222, 1.8116666666666668, 2.1733333333333333, 2.305277777777778, 2.4227777777777777, 2.4916666666666667, 2.613888888888889, 2.6...
6
3
376
[ "2015-12-17T00:12:04", "2015-12-17T00:15:16", "2015-12-17T00:20:57", "2015-12-17T00:23:13", "2015-12-17T00:24:10", "2015-12-17T00:26:23", "2015-12-17T00:27:00", "2015-12-17T00:38:34", "2015-12-17T00:45:43", "2015-12-17T01:04:16", "2015-12-17T01:16:57", "2015-12-17T01:18:25", "2015-12-17T01:4...
[ 0, 2, 0, 0, 0, 1, 0, 2, 0, 0, 2, 0, 0, 2, 1, 0, 2, 0, 0, 2, 0, 0, 0, 0, 2, 0, 1, 2, 0, 0, 0, 0, 2, 0, 1, 1, 2, 2, 1, 2, 0, 2, 2, 2, 1, 2, 1, 0, 0, 2, 2, 0, 2, 0, 2, 0, 0, 2, 2, 2, 0, 0, 2, 2...
[ 0, 0.05333333333333334, 0.09472222222222222, 0.03777777777777778, 0.015833333333333335, 0.036944444444444446, 0.010277777777777778, 0.19277777777777777, 0.11916666666666667, 0.30916666666666665, 0.21138888888888888, 0.024444444444444446, 0.4702777777777778, 0.0025, 0.009722222222222222, ...
[ 0, 0.05333333333333334, 0.14805555555555555, 0.18583333333333332, 0.20166666666666666, 0.2386111111111111, 0.24888888888888888, 0.44166666666666665, 0.5608333333333333, 0.87, 1.081388888888889, 1.1058333333333332, 1.576111111111111, 1.5786111111111112, 1.5883333333333334, 1.59083333333...
7
3
387
[ "2015-12-18T05:27:31", "2015-12-18T06:34:13", "2015-12-18T06:41:53", "2015-12-18T07:08:17", "2015-12-18T07:15:13", "2015-12-18T07:21:07", "2015-12-18T07:28:03", "2015-12-18T07:32:55", "2015-12-18T07:40:21", "2015-12-18T07:46:44", "2015-12-18T07:47:38", "2015-12-18T07:53:36", "2015-12-18T07:5...
[ 2, 0, 0, 0, 2, 2, 0, 0, 2, 1, 0, 2, 2, 0, 2, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 2, 2, 0, 1, 0, 2, 2, 0, 0, 0, 0, 0, 0, 0, 2, 2, 2, 2, 0, 0, 2, 0, 0, 0, 0, 0, 0, 2, 0, 1, 1, 0, 0, 0, 1, 2, 0, 1...
[ 0, 1.1116666666666666, 0.12777777777777777, 0.44, 0.11555555555555555, 0.09833333333333333, 0.11555555555555555, 0.0811111111111111, 0.1238888888888889, 0.1063888888888889, 0.015, 0.09944444444444445, 0, 0.049444444444444444, 0.095, 0.030555555555555555, 0.21305555555555555, 0.0583...
[ 0, 1.1116666666666666, 1.2394444444444443, 1.6794444444444445, 1.795, 1.8933333333333333, 2.008888888888889, 2.09, 2.213888888888889, 2.3202777777777777, 2.335277777777778, 2.4347222222222222, 2.4347222222222222, 2.484166666666667, 2.5791666666666666, 2.609722222222222, 2.82277777777...
8
3
346
["2015-12-19T06:08:04","2015-12-19T06:08:04","2015-12-19T08:08:53","2015-12-19T08:19:40","2015-12-19(...TRUNCATED)
[1,1,0,0,0,0,2,0,2,0,2,0,2,0,0,0,2,0,2,2,2,0,0,2,0,2,2,2,0,2,2,0,2,2,0,1,0,0,1,0,2,2,2,2,0,0,2,2,1,0(...TRUNCATED)
[0.0,0.0,2.013611111111111,0.17972222222222223,0.009166666666666667,0.1597222222222222,0.16611111111(...TRUNCATED)
[0.0,0.0,2.013611111111111,2.1933333333333334,2.2025,2.362222222222222,2.5283333333333333,2.55527777(...TRUNCATED)
9
3
272
End of preview. Expand in Data Studio

MUSES: a benchmark for Marked Unevenly Spaced Event Sequences

MUSES, a benchmark for Marked Unevenly Spaced Event Sequences, is a collection of unevenly spaced time series datasets from various domains, containing marked events for training and evaluating prediction approaches.

Languages

All the columns and classes (when textual) in MUSES are in English (BCP-47 en)

Dataset Structure

Data Instances

All datasets formatting follows a similar structure.

{
    "datetime": [2020-06-10T14:27:37.963000000, 2020-06-20T21:23:06.980000000, 2020-10-15T13:36:16.687000000],
    "type_event": [7, 2, 15],
    "time_since_last_event": [0.0, 246.92472694444444, 2800.2193630555557],
    "time_since_start": [0.0, 246.92472694444444, 3047.14409],
    "seq_idx": 1,
    "dim_process": 50,
    "seq_len": 3
}

earthquake

  • Size of downloaded dataset files: 2.77 MB
  • Size of the generated dataset: 3.06 MB
  • Total amount of disk used: 5.83 MB

memetrack

  • Size of downloaded dataset files: 569.01 MB
  • Size of the generated dataset: 993.72 MB
  • Total amount of disk used: 1562.72 MB

stackoverflow

  • Size of downloaded dataset files: 110.71 MB
  • Size of the generated dataset: 143.91 MB
  • Total amount of disk used: 254.62 MB

taxi_nyc_neighborhoods

  • Size of downloaded dataset files: 968.44 MB
  • Size of the generated dataset: 2031.69 MB
  • Total amount of disk used: 3000.14 MB

synthea

  • Size of downloaded dataset files: 7.65 MB
  • Size of the generated dataset: 14.56 MB
  • Total amount of disk used: 22.21 MB

spiketrains

  • Size of downloaded dataset files: 8.14 MB
  • Size of the generated dataset: 20.57 MB
  • Total amount of disk used: 28.71 MB

crypto_transactions

  • Size of downloaded dataset files: 393.35 MB
  • Size of the generated dataset: 748.14 MB
  • Total amount of disk used: 1141.49 MB

human_activity

  • Size of downloaded dataset files: 0.20 MB
  • Size of the generated dataset: 0.22 MB
  • Total amount of disk used: 0.42 MB

911

  • Size of downloaded dataset files: 8.42 MB
  • Size of the generated dataset: 21.63 MB
  • Total amount of disk used: 30.05 MB

mooc

  • Size of downloaded dataset files: 3.97 MB
  • Size of the generated dataset: 10.29 MB
  • Total amount of disk used: 14.26 MB

amazon_easytpp

  • Size of downloaded dataset files: 7.81 MB
  • Size of the generated dataset: 10.41 MB
  • Total amount of disk used: 18.22 MB

wikipedia

  • Size of downloaded dataset files: 127.20 MB
  • Size of the generated dataset: 238.24 MB
  • Total amount of disk used: 365.43 MB

retweet_easytpp

  • Size of downloaded dataset files: 15.12 MB
  • Size of the generated dataset: 64.49 MB
  • Total amount of disk used: 79.62 MB

taobao_easytpp

  • Size of downloaded dataset files: 1.60 MB
  • Size of the generated dataset: 2.84 MB
  • Total amount of disk used: 4.43 MB

taxi_easytpp

  • Size of downloaded dataset files: 0.59 MB
  • Size of the generated dataset: 1.88 MB
  • Total amount of disk used: 2.47 MB

volcano_easytpp

  • Size of downloaded dataset files: 0.15 MB
  • Size of the generated dataset: 0.23 MB
  • Total amount of disk used: 0.38 MB

hawkes_dependent

  • Size of downloaded dataset files: 11.10 MB
  • Size of the generated dataset: 15.84 MB
  • Total amount of disk used: 26.94 MB

hawkes_1

  • Size of downloaded dataset files: 2.04 MB
  • Size of the generated dataset: 2.49 MB
  • Total amount of disk used: 4.54 MB

Data Fields

The data fields are the same among all datasets and splits. The datetime column is optional and left out when no data is available.

  • type_event: a classification label (int64) sequence.
  • time_since_last_event: a float64 sequence of inter-event times.
  • time_since_start: a float64 sequence of event timestamps since sequence start.
  • seq_idx: an int64 feature uniquely identifying sequences in the split.
  • dim_process: an int64 feature representing total number of classes in the dataset.
  • seq_len: an int64 feature containing sequence length.
  • datetime: (if avail.) a timestamp[ns] sequence, containing event datetime if available.

The classification labels differ for each dataset. The corresponding labels for each event type can be read from metadata like this

>>> import datasets as ds
>>> meta_data = ds.load_dataset_builder("ddrg/MUSES", "DATASET-NAME")
>>> [label for label in enumerate(meta_data.info.features["type_event"].feature.names)]

Data Splits

train validation test
earthquake 1724 216 215
memetrack 1264794 158100 158099
stackoverflow 42838 5354 5355
synthea 9198 1150 1149
spiketrains 480 60 60
crypto_transactions 1408527 176066 176066
human_activity 175 22 22
911 1355 169 170
mooc 5638 704 705
amazon_easytpp 7382 922 923
wikipedia 907765 113470 113471
retweet_easytpp 19200 2400 2400
taobao_easytpp 1600 200 200
taxi_easytpp 1600 200 200
volcano_easytpp 333 42 41
hawkes_dependent 19661 2458 2457
hawkes_1 800 100 100

taxi_nyc_neighborhoods

train validation test extra
taxi_nyc_neighborhoods 273060 34132 34133 106020

Dataset Creation

Curation Rationale

A dataset has to comply with the following hard requirements to be eligible for MUSES.

  1. Inherently unevenly spaced
  2. Time series minimum length of 2
  3. Known license
  4. Used in Literature
  5. Replicably described data collection and processing
  6. Dataset has practical relevance or is commonly used for evaluation
  7. Maximum of 100 M events
  8. Nominal classes or common class analogy
  9. Commonly used in unevenly spaced time or TPP forecasting

Further, we aim to fulfull the following variety constraints by dataset composition.

  1. Coverage of many different domains
  2. Variety in dataset sizes
  3. Variety in sequence lengths
  4. Variety in problem difficulties
  5. Variety in problem complexity
  6. Variety in class balances

See our paper for further information.

Source Data

Initial Data Collection and Normalization

Data was sourced and preprocessed like in other works, unless explicitly stated otherwise in our paper. See Citation Information for details.

Personal and Sensitive Information

Data is deidentified or all personal and sensitive information was removed.

Considerations for Using the Data

Discussion of Biases

There are no ethical biases we are aware of.

Other Known Limitations

Class imbalances are common in event sequences. See the following paper for further information. TODO JIMMY

Additional Information

Dataset Curators

More Information Needed

Licensing Information

The primary MUSES tasks are built on and derived from existing datasets. We refer users to the original licenses accompanying each dataset.

License Other
earthquake Public Domain
memetrack 3-Clause BSD see LICENSE/memetrack
stackoverflow CC BY-SA see LICENSE/stackoverflow
taxi_nyc_neighborhoods Public Domain
synthea Apache 2.0 see LICENSE/synthea
spiketrains CC0 see LICENSE/spiketrains
crypto_transactions 3-Clause BSD see LICENSE/crypto_transactions
human_activity CC-BY see LICENSE/human_activity
911 ODC ODbL see LICENSE/911
mooc Scientific Use Only see LICENSE/mooc
amazon_easytpp Apache 2.0 see LICENSE/amazon_easytpp
wikipedia CC BY-SA see LICENSE/wikipedia
retweet_easytpp Apache 2.0 see LICENSE/retweet_easytpp
taobao_easytpp Apache 2.0 see LICENSE/taobao_easytpp
taxi_easytpp Apache 2.0 see LICENSE/taxi_easytpp
volcano_easytpp Apache 2.0 see LICENSE/volcano_easytpp
hawkes_dependent Public Domain
hawkes_1 Public Domain

Citation Information

We encourage you to use the following BibTeX citation for MUSES itself:

# TODO @Jimmy Add here Final Paper

If you use MUSES, please also cite all the individual datasets you use, both to give the original authors their due credit and because venues will expect papers to describe the data they evaluate on. The following provides BibTeX for all of the MUSES tasks.

earthquake

@article{stockman2024earthquakenpp,
  title={EarthquakeNPP: A Benchmark for Earthquake Forecasting with Neural Point Processes},
  author={Stockman, Samuel and Lawson, Daniel and Werner, Maximilian},
  journal={arXiv preprint arXiv:2410.08226},
  year={2024}
}

memetrack

@inproceedings{leskovec2009meme,
  title={Meme-tracking and the dynamics of the news cycle},
  author={Leskovec, Jure and Backstrom, Lars and Kleinberg, Jon},
  booktitle={Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining},
  pages={497--506},
  year={2009}
}
@article{mei2017neural,
  title={The neural hawkes process: A neurally self-modulating multivariate point process},
  author={Mei, Hongyuan and Eisner, Jason M},
  journal={Advances in neural information processing systems},
  volume={30},
  year={2017}
}

stackoverflow

@inproceedings{du2016recurrent,
  title={Recurrent marked temporal point processes: Embedding event history to vector},
  author={Du, Nan and Dai, Hanjun and Trivedi, Rakshit and Upadhyay, Utkarsh and Gomez-Rodriguez, Manuel and Song, Le},
  booktitle={Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
  pages={1555--1564},
  year={2016}
}

taxi_nyc_neighborhoods

@inproceedings{du2016recurrent,
  title={Recurrent marked temporal point processes: Embedding event history to vector},
  author={Du, Nan and Dai, Hanjun and Trivedi, Rakshit and Upadhyay, Utkarsh and Gomez-Rodriguez, Manuel and Song, Le},
  booktitle={Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
  pages={1555--1564},
  year={2016}
}

synthea

@article{walonoski2018synthea,
  title={Synthea: An approach, method, and software mechanism for generating synthetic patients and the synthetic electronic health care record},
  author={Walonoski, Jason and Kramer, Mark and Nichols, Joseph and Quina, Andre and Moesel, Chris and Hall, Dylan and Duffett, Carlton and Dube, Kudakwashe and Gallagher, Thomas and McLachlan, Scott},
  journal={Journal of the American Medical Informatics Association},
  volume={25},
  number={3},
  pages={230--238},
  year={2018},
  publisher={Oxford University Press}
}
@inproceedings{enguehard2020neural,
  title={Neural temporal point processes for modelling electronic health records},
  author={Enguehard, Joseph and Busbridge, Dan and Bozson, Adam and Woodcock, Claire and Hammerla, Nils},
  booktitle={Machine Learning for Health},
  pages={85--113},
  year={2020},
  organization={PMLR}
}

spiketrains

@article{stetter2012model,
  title={Model-free reconstruction of excitatory neuronal connectivity from calcium imaging signals},
  author={Stetter, Olav and Battaglia, Demian and Soriano, Jordi and Geisel, Theo},
  year={2012},
  publisher={Public Library of Science San Francisco, USA}
}
@article{linderman2015scalable,
  title={Scalable bayesian inference for excitatory point process networks},
  author={Linderman, Scott W and Adams, Ryan P},
  journal={arXiv preprint arXiv:1507.03228},
  year={2015}
}

crypto_transactions

@article{shamsi2022chartalist,
  title={Chartalist: Labeled graph datasets for utxo and account-based blockchains},
  author={Shamsi, Kiarash and Victor, Friedhelm and Kantarcioglu, Murat and Gel, Yulia and Akcora, Cuneyt G},
  journal={Advances in Neural Information Processing Systems},
  volume={35},
  pages={34926--34939},
  year={2022}
}
@inproceedings{du2016recurrent,
  title={Recurrent marked temporal point processes: Embedding event history to vector},
  author={Du, Nan and Dai, Hanjun and Trivedi, Rakshit and Upadhyay, Utkarsh and Gomez-Rodriguez, Manuel and Song, Le},
  booktitle={Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
  pages={1555--1564},
  year={2016}
}

human_activity

@article{cook2012casas,
  title={CASAS: A smart home in a box},
  author={Cook, Diane J and Crandall, Aaron S and Thomas, Brian L and Krishnan, Narayanan C},
  journal={Computer},
  volume={46},
  number={7},
  pages={62--69},
  year={2012},
  publisher={IEEE}
}
@article{fortino2021predicting,
  title={Predicting activities of daily living via temporal point processes: Approaches and experimental results},
  author={Fortino, Giancarlo and Guzzo, Antonella and Ianni, Michele and Leotta, Francesco and Mecella, Massimo},
  journal={Computers \& Electrical Engineering},
  volume={96},
  pages={107567},
  year={2021},
  publisher={Elsevier}
}

911

@inproceedings{zuo2020transformer,
  title={Transformer hawkes process},
  author={Zuo, Simiao and Jiang, Haoming and Li, Zichong and Zhao, Tuo and Zha, Hongyuan},
  booktitle={International conference on machine learning},
  pages={11692--11702},
  year={2020},
  organization={PMLR}
}

mooc

@inproceedings{kumar2019predicting,
  title={Predicting dynamic embedding trajectory in temporal interaction networks},
  author={Kumar, Srijan and Zhang, Xikun and Leskovec, Jure},
  booktitle={Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery \& data mining},
  pages={1269--1278},
  year={2019}
}
@article{shchur2019intensity,
  title={Intensity-free learning of temporal point processes},
  author={Shchur, Oleksandr and Bilo{\v{s}}, Marin and G{\"u}nnemann, Stephan},
  journal={arXiv preprint arXiv:1909.12127},
  year={2019}
}

amazon_easytpp

@inproceedings{ni2019justifying,
  title={Justifying recommendations using distantly-labeled reviews and fine-grained aspects},
  author={Ni, Jianmo and Li, Jiacheng and McAuley, Julian},
  booktitle={Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP)},
  pages={188--197},
  year={2019}
}
@article{xue2023easytpp,
  title={Easytpp: Towards open benchmarking temporal point processes},
  author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
  journal={arXiv preprint arXiv:2307.08097},
  year={2023}
}

wikipedia

@inproceedings{kumar2019predicting,
  title={Predicting dynamic embedding trajectory in temporal interaction networks},
  author={Kumar, Srijan and Zhang, Xikun and Leskovec, Jure},
  booktitle={Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery \& data mining},
  pages={1269--1278},
  year={2019}
}
@article{bosser2023predictive,
  title={On the predictive accuracy of neural temporal point process models for continuous-time event data},
  author={Bosser, Tanguy and Taieb, Souhaib Ben},
  journal={arXiv preprint arXiv:2306.17066},
  year={2023}
}

retweet_easytpp

@inproceedings{zhao2015seismic,
  title={Seismic: A self-exciting point process model for predicting tweet popularity},
  author={Zhao, Qingyuan and Erdogdu, Murat A and He, Hera Y and Rajaraman, Anand and Leskovec, Jure},
  booktitle={Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining},
  pages={1513--1522},
  year={2015}
}
@article{xue2023easytpp,
  title={Easytpp: Towards open benchmarking temporal point processes},
  author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
  journal={arXiv preprint arXiv:2307.08097},
  year={2023}
}

taobao_easytpp

@article{xue2022hypro,
  title={Hypro: A hybridly normalized probabilistic model for long-horizon prediction of event sequences},
  author={Xue, Siqiao and Shi, Xiaoming and Zhang, James and Mei, Hongyuan},
  journal={Advances in Neural Information Processing Systems},
  volume={35},
  pages={34641--34650},
  year={2022}
}
@article{xue2023easytpp,
  title={Easytpp: Towards open benchmarking temporal point processes},
  author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
  journal={arXiv preprint arXiv:2307.08097},
  year={2023}
}

taxi_easytpp

@article{xue2023easytpp,
  title={Easytpp: Towards open benchmarking temporal point processes},
  author={Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
  journal={arXiv preprint arXiv:2307.08097},
  year={2023}
}

volcano_easytpp

@article{bebbington2014long,
  title={Long-term forecasting of volcanic explosivity},
  author={Bebbington, MS},
  journal={Geophysical Journal International},
  volume={197},
  number={3},
  pages={1500--1515},
  year={2014},
  publisher={Oxford University Press}
}
@misc{easytpp-github-volcano,
  title = {EasyTemporalPointProcess GitHub Volcano Preprocessing},
  author = {Xue, Siqiao and Shi, Xiaoming and Chu, Zhixuan and Wang, Yan and Hao, Hongyan and Zhou, Fan and Jiang, Caigao and Pan, Chen and Zhang, James Y and Wen, Qingsong and others},
  url = {https://github.com/ant-research/EasyTemporalPointProcess/blob/main/examples/script_data_processing/volcano.py}
}

hawkes_dependent

@inproceedings{enguehard2020neural,
  title={Neural temporal point processes for modelling electronic health records},
  author={Enguehard, Joseph and Busbridge, Dan and Bozson, Adam and Woodcock, Claire and Hammerla, Nils},
  booktitle={Machine Learning for Health},
  pages={85--113},
  year={2020},
  organization={PMLR}
}

hawkes_1

@article{omi2019fully,
  title={Fully neural network based model for general temporal point processes},
  author={Omi, Takahiro and Aihara, Kazuyuki and others},
  journal={Advances in neural information processing systems},
  volume={32},
  year={2019}
}

Contributions

Thanks to @Moritz Tschöpe and @Jimmy Pöhlmann for adding this dataset.

Downloads last month
8,744

Papers for ddrg/MUSES