aractingi commited on
Commit
43b4ae5
·
verified ·
1 Parent(s): a9447b8

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -53,3 +53,11 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
56
+ cmu_stretch-train.tfrecord-00000-of-00008 filter=lfs diff=lfs merge=lfs -text
57
+ cmu_stretch-train.tfrecord-00001-of-00008 filter=lfs diff=lfs merge=lfs -text
58
+ cmu_stretch-train.tfrecord-00002-of-00008 filter=lfs diff=lfs merge=lfs -text
59
+ cmu_stretch-train.tfrecord-00003-of-00008 filter=lfs diff=lfs merge=lfs -text
60
+ cmu_stretch-train.tfrecord-00004-of-00008 filter=lfs diff=lfs merge=lfs -text
61
+ cmu_stretch-train.tfrecord-00005-of-00008 filter=lfs diff=lfs merge=lfs -text
62
+ cmu_stretch-train.tfrecord-00006-of-00008 filter=lfs diff=lfs merge=lfs -text
63
+ cmu_stretch-train.tfrecord-00007-of-00008 filter=lfs diff=lfs merge=lfs -text
cmu_stretch-train.tfrecord-00000-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17852298a91f1b8d0c192b3327e8c10860f78d7a72c53c5b718377cbfd66ce70
3
+ size 90732337
cmu_stretch-train.tfrecord-00001-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e5559db1e9688fa5b6d3b321d7c833d3059787ac61da68ad36ea4194d4ac43a
3
+ size 157792547
cmu_stretch-train.tfrecord-00002-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8b40b85e22e9fce42d1c45a410dbce57c151388699044f346e69454f305e379
3
+ size 111868925
cmu_stretch-train.tfrecord-00003-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9f9a09deb315664001e95da712bc6fe1d79d0f65d714257b564e2c04a2eae7c
3
+ size 91971290
cmu_stretch-train.tfrecord-00004-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d159db1974017f372320f5079e4490d98c1e3552f99b5b0c0c92bbb0d535435
3
+ size 102700566
cmu_stretch-train.tfrecord-00005-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88ed4960d299e60a1a0a5acdaad07908a6a70ceb12e2929458d80e1301106a2a
3
+ size 61104180
cmu_stretch-train.tfrecord-00006-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce51d3d38e00072c2f4028a8ba6c16d17dd7364f2944a66172e2298d32e2c6ca
3
+ size 69391046
cmu_stretch-train.tfrecord-00007-of-00008 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78c8d1a8f0fbe54de2bdf64a99e8eb68c0ccec465acff6144c7ae08b764f8d76
3
+ size 77869146
dataset_info.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "@inproceedings{bahl2023affordances,\n title={Affordances from Human Videos as a Versatile Representation for Robotics},\n author={Bahl, Shikhar and Mendonca, Russell and Chen, Lili and Jain, Unnat and Pathak, Deepak},\n booktitle={CVPR},\n year={2023}\n}\n@article{mendonca2023structured,\n title={Structured World Models from Human Videos},\n author={Mendonca, Russell and Bahl, Shikhar and Pathak, Deepak},\n journal={CoRL},\n year={2023}\n}",
3
+ "description": "Hello stretch robot kitchen interactions",
4
+ "fileFormat": "tfrecord",
5
+ "location": {
6
+ "urls": [
7
+ "https://robo-affordances.github.io/"
8
+ ]
9
+ },
10
+ "moduleName": "tensorflow_datasets.robotics.rtx.rtx",
11
+ "name": "cmu_stretch",
12
+ "releaseNotes": {
13
+ "0.1.0": "Initial release."
14
+ },
15
+ "splits": [
16
+ {
17
+ "filepathTemplate": "{DATASET}-{SPLIT}.{FILEFORMAT}-{SHARD_X_OF_Y}",
18
+ "name": "train",
19
+ "numBytes": "763427877",
20
+ "shardLengths": [
21
+ "16",
22
+ "28",
23
+ "20",
24
+ "16",
25
+ "18",
26
+ "11",
27
+ "12",
28
+ "14"
29
+ ]
30
+ }
31
+ ],
32
+ "version": "0.1.0"
33
+ }
features.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "featuresDict": {
3
+ "features": {
4
+ "episode_metadata": {
5
+ "featuresDict": {
6
+ "features": {
7
+ "file_path": {
8
+ "description": "Path to the original data file.",
9
+ "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text",
10
+ "text": {}
11
+ }
12
+ }
13
+ },
14
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict"
15
+ },
16
+ "steps": {
17
+ "pythonClassName": "tensorflow_datasets.core.features.dataset_feature.Dataset",
18
+ "sequence": {
19
+ "feature": {
20
+ "featuresDict": {
21
+ "features": {
22
+ "action": {
23
+ "description": "Robot action, consists of [3x ee pos, 3x ee rot 1x gripper binary action, 1x terminate episode].",
24
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
25
+ "tensor": {
26
+ "dtype": "float32",
27
+ "encoding": "none",
28
+ "shape": {
29
+ "dimensions": [
30
+ "8"
31
+ ]
32
+ }
33
+ }
34
+ },
35
+ "discount": {
36
+ "description": "Discount if provided, default to 1.",
37
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
38
+ "tensor": {
39
+ "dtype": "float32",
40
+ "encoding": "none",
41
+ "shape": {}
42
+ }
43
+ },
44
+ "is_first": {
45
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
46
+ "tensor": {
47
+ "dtype": "bool",
48
+ "encoding": "none",
49
+ "shape": {}
50
+ }
51
+ },
52
+ "is_last": {
53
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
54
+ "tensor": {
55
+ "dtype": "bool",
56
+ "encoding": "none",
57
+ "shape": {}
58
+ }
59
+ },
60
+ "is_terminal": {
61
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
62
+ "tensor": {
63
+ "dtype": "bool",
64
+ "encoding": "none",
65
+ "shape": {}
66
+ }
67
+ },
68
+ "language_embedding": {
69
+ "description": "Kona language embedding. See https://tfhub.dev/google/universal-sentence-encoder-large/5",
70
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
71
+ "tensor": {
72
+ "dtype": "float32",
73
+ "encoding": "none",
74
+ "shape": {
75
+ "dimensions": [
76
+ "512"
77
+ ]
78
+ }
79
+ }
80
+ },
81
+ "language_instruction": {
82
+ "description": "Language Instruction.",
83
+ "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text",
84
+ "text": {}
85
+ },
86
+ "observation": {
87
+ "featuresDict": {
88
+ "features": {
89
+ "image": {
90
+ "description": "Main camera RGB observation.",
91
+ "image": {
92
+ "dtype": "uint8",
93
+ "encodingFormat": "png",
94
+ "shape": {
95
+ "dimensions": [
96
+ "128",
97
+ "128",
98
+ "3"
99
+ ]
100
+ }
101
+ },
102
+ "pythonClassName": "tensorflow_datasets.core.features.image_feature.Image"
103
+ },
104
+ "state": {
105
+ "description": "Robot state, consists of [3x robot joint angles/ee pos, 1x gripper position].",
106
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
107
+ "tensor": {
108
+ "dtype": "float32",
109
+ "encoding": "none",
110
+ "shape": {
111
+ "dimensions": [
112
+ "4"
113
+ ]
114
+ }
115
+ }
116
+ }
117
+ }
118
+ },
119
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict"
120
+ },
121
+ "reward": {
122
+ "description": "Reward if provided, 1 on final step for demos.",
123
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
124
+ "tensor": {
125
+ "dtype": "float32",
126
+ "encoding": "none",
127
+ "shape": {}
128
+ }
129
+ }
130
+ }
131
+ },
132
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict"
133
+ },
134
+ "length": "-1"
135
+ }
136
+ }
137
+ }
138
+ },
139
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict"
140
+ }