Add task category, link to paper

#2
by nielsr HF Staff - opened
Files changed (2) hide show
  1. LICENSE.txt +21 -0
  2. README.md +9 -4
LICENSE.txt ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ MIT License
2
+
3
+ Copyright (c) 2024 Daniel McGann
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in all
13
+ copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21
+ SOFTWARE.
README.md CHANGED
@@ -1,12 +1,20 @@
1
  ---
2
  license: mit
3
  viewer: false
 
 
4
  ---
 
5
  # TactileTracking: A tactile-based object tracking dataset
6
  [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT)
7
 
8
- We present a benchmark dataset for tactile-based object tracking, featuring 12 distinct objects and 84 tracking trials—7 trials per object, each lasting an average of 10.2 seconds. The dataset includes tactile video, per-frame 6DoF ground truth sensor poses, and pre-processed surface geometry constructed from each tactile video frame. For a robust, real-time, and accurate tactile-based object tracking solution, explore our work [NormalFlow](https://github.com/rpl-cmu/normalflow). To compare NormalFlow with other methods on this dataset,use the [NormalFlow Experiments](https://github.com/rpl-cmu/normalflow_experiment) repository.
 
 
 
 
9
 
 
10
 
11
  ## Collection Setup
12
 
@@ -35,9 +43,6 @@ Each data collection trial directory contains the following components:
35
  - **contact_masks.npy**: An (N, H, W) array of the computed contact masks for each frame in `gelsight.mp4`, derived solely from the tactile images.
36
  - **gradient_maps.npy**: An (N, H, W, 2) array of the computed gradient maps for each frame in `gelsight.mp4`, based only on the tactile images.
37
 
38
- [TODO: Visualize the surface geometry information here]
39
-
40
-
41
  ## Dataset Statistics
42
 
43
  Our benchmark dataset focuses on frame-to-frame object pose tracking, with each trial ensuring overlap between the first (reference) frame and subsequent (target) frames. This setup restricts the object to local movement without long-distance shifts. The table below details the average 6DoF movement range for each object. This dataset prioritizes rotational movement, as excessive translational sliding risks damaging the sensor’s gel.
 
1
  ---
2
  license: mit
3
  viewer: false
4
+ task_categories:
5
+ - robotics
6
  ---
7
+
8
  # TactileTracking: A tactile-based object tracking dataset
9
  [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT)
10
 
11
+ <div style="text-align: center;">
12
+ <video width="100%" onmouseover="this.pause()" onmouseout="this.play()" autoplay="" loop="" muted="">
13
+ <source src="https://joehjhuang.github.io/normalflow/videos/dataset.mp4" type="video/mp4">
14
+ </video>
15
+ </div>
16
 
17
+ We present a benchmark dataset for tactile-based object tracking, featuring 12 distinct objects and 84 tracking trials—7 trials per object, each lasting an average of 10.2 seconds. The dataset includes tactile video, per-frame 6DoF ground truth sensor poses, and pre-processed surface geometry constructed from each tactile video frame. For a robust, real-time, and accurate tactile-based object tracking solution, explore our work [NormalFlow](https://huggingface.co/papers/2412.09617). To compare NormalFlow with other methods on this dataset,use the [NormalFlow Experiments](https://github.com/rpl-cmu/normalflow_experiment) repository.
18
 
19
  ## Collection Setup
20
 
 
43
  - **contact_masks.npy**: An (N, H, W) array of the computed contact masks for each frame in `gelsight.mp4`, derived solely from the tactile images.
44
  - **gradient_maps.npy**: An (N, H, W, 2) array of the computed gradient maps for each frame in `gelsight.mp4`, based only on the tactile images.
45
 
 
 
 
46
  ## Dataset Statistics
47
 
48
  Our benchmark dataset focuses on frame-to-frame object pose tracking, with each trial ensuring overlap between the first (reference) frame and subsequent (target) frames. This setup restricts the object to local movement without long-distance shifts. The table below details the average 6DoF movement range for each object. This dataset prioritizes rotational movement, as excessive translational sliding risks damaging the sensor’s gel.