whenxuan commited on
Commit
a1820f1
·
1 Parent(s): e3acc6d

whenxuan: upload the time series dataset

Browse files
train/time_series/taxi_1h/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c522b3702873531efd2098abb1000b6f25eb2db46339ed87c43edea23bfe1b4b
3
+ size 7253304
train/time_series/taxi_1h/dataset_info.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "",
4
+ "features": {
5
+ "target": {
6
+ "feature": {
7
+ "dtype": "float32",
8
+ "_type": "Value"
9
+ },
10
+ "_type": "List"
11
+ },
12
+ "id": {
13
+ "dtype": "string",
14
+ "_type": "Value"
15
+ },
16
+ "length": {
17
+ "dtype": "int64",
18
+ "_type": "Value"
19
+ },
20
+ "dataset": {
21
+ "dtype": "string",
22
+ "_type": "Value"
23
+ }
24
+ },
25
+ "homepage": "",
26
+ "license": ""
27
+ }
train/time_series/taxi_1h/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "df5e25adc1a8a054",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": null
13
+ }
train/time_series/taxi_30min/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55007caaf49cf6482f99971727f1946250fa426d2c6dfbb6ca8d0a1ad678915b
3
+ size 14442624
train/time_series/taxi_30min/dataset_info.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "",
4
+ "features": {
5
+ "target": {
6
+ "feature": {
7
+ "dtype": "float32",
8
+ "_type": "Value"
9
+ },
10
+ "_type": "List"
11
+ },
12
+ "id": {
13
+ "dtype": "string",
14
+ "_type": "Value"
15
+ },
16
+ "length": {
17
+ "dtype": "int64",
18
+ "_type": "Value"
19
+ },
20
+ "dataset": {
21
+ "dtype": "string",
22
+ "_type": "Value"
23
+ }
24
+ },
25
+ "homepage": "",
26
+ "license": ""
27
+ }
train/time_series/taxi_30min/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "7ea61a945e0784cf",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": null
13
+ }
train/time_series/uber_tlc_daily/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63e14e11028bde171c4f2e972e8601820f014bffb537a683f2b995084075c419
3
+ size 389992
train/time_series/uber_tlc_daily/dataset_info.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "",
4
+ "features": {
5
+ "target": {
6
+ "feature": {
7
+ "dtype": "int64",
8
+ "_type": "Value"
9
+ },
10
+ "_type": "List"
11
+ },
12
+ "id": {
13
+ "dtype": "string",
14
+ "_type": "Value"
15
+ },
16
+ "length": {
17
+ "dtype": "int64",
18
+ "_type": "Value"
19
+ },
20
+ "dataset": {
21
+ "dtype": "string",
22
+ "_type": "Value"
23
+ }
24
+ },
25
+ "homepage": "",
26
+ "license": ""
27
+ }
train/time_series/uber_tlc_daily/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "0dbbfcaff9f72dea",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": null
13
+ }
train/time_series/uber_tlc_hourly/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f71fab8cfaa656d71199611aae4b78a27a678e50f3e5f2e85733628d8ec5407
3
+ size 9126792
train/time_series/uber_tlc_hourly/dataset_info.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "",
4
+ "features": {
5
+ "target": {
6
+ "feature": {
7
+ "dtype": "int64",
8
+ "_type": "Value"
9
+ },
10
+ "_type": "List"
11
+ },
12
+ "id": {
13
+ "dtype": "string",
14
+ "_type": "Value"
15
+ },
16
+ "length": {
17
+ "dtype": "int64",
18
+ "_type": "Value"
19
+ },
20
+ "dataset": {
21
+ "dtype": "string",
22
+ "_type": "Value"
23
+ }
24
+ },
25
+ "homepage": "",
26
+ "license": ""
27
+ }
train/time_series/uber_tlc_hourly/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "a180caadfbaf3dac",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": null
13
+ }
train/time_series/wiki_daily_100k/data-00000-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:335d2b07e4c1e3153b7ecfd9db7d74de37d84d60427a1c3e9d745a47602517ff
3
+ size 366813240
train/time_series/wiki_daily_100k/data-00001-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9195ffec7053ea51af0d38d4f51e0614f74a3d0e421378a09fb814f5a6b2b79
3
+ size 366813576
train/time_series/wiki_daily_100k/data-00002-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68bebe5f7d4fa2994cf0547d1ebc4deb35533364d35aef5c1c64827045f4aa99
3
+ size 366813464
train/time_series/wiki_daily_100k/dataset_info.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "",
4
+ "features": {
5
+ "target": {
6
+ "feature": {
7
+ "dtype": "float32",
8
+ "_type": "Value"
9
+ },
10
+ "_type": "List"
11
+ },
12
+ "id": {
13
+ "dtype": "string",
14
+ "_type": "Value"
15
+ },
16
+ "length": {
17
+ "dtype": "int64",
18
+ "_type": "Value"
19
+ },
20
+ "dataset": {
21
+ "dtype": "string",
22
+ "_type": "Value"
23
+ }
24
+ },
25
+ "homepage": "",
26
+ "license": ""
27
+ }
train/time_series/wiki_daily_100k/state.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00003.arrow"
5
+ },
6
+ {
7
+ "filename": "data-00001-of-00003.arrow"
8
+ },
9
+ {
10
+ "filename": "data-00002-of-00003.arrow"
11
+ }
12
+ ],
13
+ "_fingerprint": "a011dcb0dbf21dba",
14
+ "_format_columns": null,
15
+ "_format_kwargs": {},
16
+ "_format_type": null,
17
+ "_output_all_columns": false,
18
+ "_split": null
19
+ }