Datasets:

Modalities:
Image
ArXiv:
Libraries:
Datasets
License:
xifen527 commited on
Commit
9b0f234
·
verified ·
1 Parent(s): a90565a

update readme

Browse files
Files changed (1) hide show
  1. README.md +161 -3
README.md CHANGED
@@ -1,3 +1,161 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+ # QuadTrack
6
+
7
+ <!-- Provide a quick summary of the dataset. -->
8
+
9
+ QuadTrack is a dataset designed for multi-object tracking (MOT) research, with a focus on panoramic and long-span scenarios. It provides challenging tracking sequences that include drastic appearance variations, prolonged occlusions, and wide field-of-view distortions, enabling the development and evaluation of robust MOT algorithms.
10
+
11
+ ## Dataset Details
12
+
13
+ ### Dataset Description
14
+
15
+ <!-- Provide a longer summary of what this dataset is. -->
16
+
17
+
18
+
19
+ - **Curated by:** [HNU CVPU]
20
+ - **Funded by [optional]:** [National Natural Science Foundation of China (No.62473139 and No.12174341), Zhejiang Provincial Natural Science Foundation of China (Grant No. LZ24F050003) and Shanghai SUPREMIND Technology Co. Ltd.]
21
+ - **Shared by [optional]:** [HNU CVPU]
22
+ - **License:** [CC BY-NC 4.0]
23
+
24
+ ### Dataset Sources [optional]
25
+
26
+ <!-- Provide the basic links for the dataset. -->
27
+
28
+ - **Repository:** [https://github.com/xifen523/OmniTrack]
29
+ - **Paper:** [https://arxiv.org/abs/2503.04565]
30
+ - **Demo:** [https://www.youtube.com/watch?v=Q3mvzBtkkeU]
31
+
32
+ ## Uses
33
+
34
+ <!-- Address questions around how the dataset is intended to be used. -->
35
+
36
+ ### Direct Use
37
+
38
+ <!-- This section describes suitable use cases for the dataset. -->
39
+
40
+ QuadTrack is designed for multi-object tracking (MOT) research, particularly in panoramic.
41
+
42
+
43
+ ## Dataset Structure
44
+
45
+ <!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
46
+ The dataset is organized into two main splits: train and test.
47
+ ```bash
48
+ QuadTrack/
49
+ ├── train/ # Training set
50
+ │ ├── img1/ # Training images (video frames)
51
+ │ └── gt/ # Ground-truth annotations (bounding boxes, IDs, etc.)
52
+
53
+ └── test/ # Test set
54
+ └── img1/ # Test images (no ground-truth provided)
55
+ ```
56
+
57
+
58
+
59
+ ## Dataset Creation
60
+
61
+ ### Curation Rationale
62
+
63
+ <!-- Motivation for the creation of this dataset. -->
64
+
65
+ QuadTrack was created to address the limitations of existing multi-object tracking (MOT) datasets, which often focus on narrow field-of-view scenarios and short-term associations. In contrast, panoramic and long-span tracking poses unique challenges such as:
66
+
67
+ + Prolonged occlusions leading to identity switches.
68
+
69
+ + Wide field-of-view distortions caused by panoramic cameras.
70
+
71
+ + Dramatic appearance variations across long sequences.
72
+
73
+ The dataset aims to provide a benchmark for developing algorithms that achieve long-term identity stability and robust re-identification in real-world panoramic environments.
74
+
75
+ ### Source Data
76
+
77
+ <!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
78
+
79
+ #### Data Collection and Processing
80
+
81
+ <!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
82
+
83
+ Collection: The video sequences were captured using panoramic and wide-angle cameras in complex real-world environments (e.g., urban traffic, crowded public areas).
84
+
85
+ + Annotation:
86
+
87
+ + Bounding boxes and unique object IDs were assigned frame-by-frame.
88
+
89
+ + Annotations follow the standard MOTChallenge format for compatibility.
90
+
91
+ + Processing:
92
+
93
+ + Frames were extracted at fixed intervals to balance temporal resolution and storage.
94
+
95
+ + Quality checks ensured consistency in ID assignment across long occlusions.
96
+
97
+ + Tools used: https://www.cvat.ai/
98
+ #### Who are the source data producers?
99
+
100
+ <!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
101
+
102
+ The source videos were collected and annotated by the QuadTrack research team.
103
+
104
+ + Producers: Internal annotation team trained for MOT labeling tasks.
105
+
106
+ + Demographics: Not applicable, as the dataset focuses on object trajectories rather than personal or sensitive identity information.
107
+
108
+ + Note: No personally identifiable information (PII) is included. The dataset is curated strictly for research purposes.
109
+
110
+
111
+ ## Bias, Risks, and Limitations
112
+
113
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
114
+
115
+ While QuadTrack provides challenging panoramic multi-object tracking scenarios, several limitations and risks should be noted:
116
+
117
+ + Domain bias: The dataset primarily consists of panoramic and wide field-of-view sequences. Models trained on QuadTrack may not generalize well to conventional narrow-angle tracking datasets.
118
+
119
+ + Scene diversity: Although collected across different environments, the dataset may not cover all possible real-world scenarios (e.g., extreme weather, night-time, or thermal imagery).
120
+
121
+ + Annotation errors: Despite quality control, occasional inaccuracies in bounding boxes or identity switches may exist, especially under heavy occlusion.
122
+
123
+ + Ethical risks: As a vision dataset, improper use in surveillance or privacy-intrusive applications could raise ethical concerns.
124
+
125
+ ### Recommendations
126
+
127
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
128
+
129
+ Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
130
+
131
+ ## Citation [optional]
132
+
133
+ <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
134
+
135
+ **BibTeX:**
136
+ ```bibtex
137
+ @inproceedings{luo2025omnidirectional,
138
+ title={Omnidirectional Multi-Object Tracking},
139
+ author={Luo, Kai and Shi, Hao and Wu, Sheng and Teng, Fei and Duan, Mengfei and Huang, Chang and Wang, Yuhang and Wang, Kaiwei and Yang, Kailun},
140
+ booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
141
+ pages={21959--21969},
142
+ year={2025}
143
+ }
144
+ ```
145
+
146
+ **APA:**
147
+
148
+ ```mathematica
149
+ Luo, K., Shi, H., Wu, S., Teng, F., Duan, M., Huang, C., Wang, Y., Wang, K., & Yang, K. (2025). Omnidirectional multi-object tracking. *Proceedings of the Computer Vision and Pattern Recognition Conference*, 21959–21969.
150
+
151
+ ```
152
+
153
+
154
+
155
+ ## Dataset Card Authors [optional]
156
+
157
+ xifen527
158
+
159
+ ## Dataset Card Contact
160
+
161
+ kailun.yang@hnu.edu.cn, luokai@hnu.edu.cn