harpreetsahota commited on
Commit
b48e3b1
·
verified ·
1 Parent(s): 0670a96

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +193 -105
README.md CHANGED
@@ -11,18 +11,19 @@ tags:
11
  - fiftyone
12
  - image
13
  - object-detection
14
- dataset_summary: '
15
 
16
 
17
 
18
 
19
- This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 5308 samples.
 
20
 
21
 
22
  ## Installation
23
 
24
 
25
- If you haven''t already, install FiftyOne:
26
 
27
 
28
  ```bash
@@ -44,9 +45,9 @@ dataset_summary: '
44
 
45
  # Load the dataset
46
 
47
- # Note: other available arguments include ''max_samples'', etc
48
 
49
- dataset = load_from_hub("harpreetsahota/FloorPlanCAD")
50
 
51
 
52
  # Launch the App
@@ -54,16 +55,12 @@ dataset_summary: '
54
  session = fo.launch_app(dataset)
55
 
56
  ```
57
-
58
- '
59
  ---
60
 
61
- # Dataset Card for FloorPlanCAD
62
-
63
- <!-- Provide a quick summary of the dataset. -->
64
-
65
-
66
 
 
67
 
68
 
69
  This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 5308 samples.
@@ -84,141 +81,232 @@ from fiftyone.utils.huggingface import load_from_hub
84
 
85
  # Load the dataset
86
  # Note: other available arguments include 'max_samples', etc
87
- dataset = load_from_hub("harpreetsahota/FloorPlanCAD")
88
 
89
  # Launch the App
90
  session = fo.launch_app(dataset)
91
  ```
92
 
93
-
94
  ## Dataset Details
95
 
96
  ### Dataset Description
97
 
98
- <!-- Provide a longer summary of what this dataset is. -->
99
-
100
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
101
 
102
- - **Curated by:** [More Information Needed]
103
- - **Funded by [optional]:** [More Information Needed]
104
- - **Shared by [optional]:** [More Information Needed]
105
- - **Language(s) (NLP):** en
106
- - **License:** [More Information Needed]
107
 
108
- ### Dataset Sources [optional]
 
 
 
 
 
109
 
110
- <!-- Provide the basic links for the dataset. -->
111
 
112
- - **Repository:** [More Information Needed]
113
- - **Paper [optional]:** [More Information Needed]
114
- - **Demo [optional]:** [More Information Needed]
115
 
116
- ## Uses
117
 
118
- <!-- Address questions around how the dataset is intended to be used. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
119
 
120
- ### Direct Use
121
 
122
- <!-- This section describes suitable use cases for the dataset. -->
123
 
124
- [More Information Needed]
125
 
126
- ### Out-of-Scope Use
127
 
128
- <!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
129
 
130
- [More Information Needed]
131
 
132
- ## Dataset Structure
133
 
134
- <!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
135
 
136
- [More Information Needed]
137
 
138
  ## Dataset Creation
139
 
140
  ### Curation Rationale
141
 
142
- <!-- Motivation for the creation of this dataset. -->
143
 
144
- [More Information Needed]
 
 
 
145
 
146
  ### Source Data
147
 
148
- <!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
149
-
150
  #### Data Collection and Processing
151
 
152
- <!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
 
 
 
 
153
 
154
- [More Information Needed]
 
 
 
 
155
 
156
  #### Who are the source data producers?
157
 
158
- <!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
159
-
160
- [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
161
 
162
- ### Annotations [optional]
163
-
164
- <!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
165
-
166
- #### Annotation process
167
-
168
- <!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
169
-
170
- [More Information Needed]
171
 
172
  #### Who are the annotators?
173
 
174
- <!-- This section describes the people or systems who created the annotations. -->
175
-
176
- [More Information Needed]
177
-
178
- #### Personal and Sensitive Information
179
-
180
- <!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
181
-
182
- [More Information Needed]
183
-
184
- ## Bias, Risks, and Limitations
185
-
186
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
187
-
188
- [More Information Needed]
189
-
190
- ### Recommendations
191
-
192
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
193
-
194
- Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
195
-
196
- ## Citation [optional]
197
-
198
- <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
199
-
200
- **BibTeX:**
201
-
202
- [More Information Needed]
203
-
204
- **APA:**
205
-
206
- [More Information Needed]
207
-
208
- ## Glossary [optional]
209
-
210
- <!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
211
-
212
- [More Information Needed]
213
-
214
- ## More Information [optional]
215
-
216
- [More Information Needed]
217
-
218
- ## Dataset Card Authors [optional]
219
-
220
- [More Information Needed]
221
 
222
- ## Dataset Card Contact
223
 
224
- [More Information Needed]
 
11
  - fiftyone
12
  - image
13
  - object-detection
14
+ dataset_summary: >
15
 
16
 
17
 
18
 
19
+ This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 5308
20
+ samples.
21
 
22
 
23
  ## Installation
24
 
25
 
26
+ If you haven't already, install FiftyOne:
27
 
28
 
29
  ```bash
 
45
 
46
  # Load the dataset
47
 
48
+ # Note: other available arguments include 'max_samples', etc
49
 
50
+ dataset = load_from_hub("Voxel51/FloorPlanCAD")
51
 
52
 
53
  # Launch the App
 
55
  session = fo.launch_app(dataset)
56
 
57
  ```
58
+ license: cc-by-sa-4.0
 
59
  ---
60
 
61
+ # Dataset Card for FloorPlanCAD (test split)
 
 
 
 
62
 
63
+ ![image](floorplancad.gif)
64
 
65
 
66
  This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 5308 samples.
 
81
 
82
  # Load the dataset
83
  # Note: other available arguments include 'max_samples', etc
84
+ dataset = load_from_hub("Voxel51/FloorPlanCAD")
85
 
86
  # Launch the App
87
  session = fo.launch_app(dataset)
88
  ```
89
 
 
90
  ## Dataset Details
91
 
92
  ### Dataset Description
93
 
94
+ FloorPlanCAD is a large-scale real-world CAD drawing dataset containing over 15,000 annotated floor plans for panoptic symbol spotting in architectural drawings. The dataset provides line-grained vector annotations for 30 object categories across residential and commercial buildings.
95
+
96
+ **Key Features:**
97
+ - **Format**: Vector graphics (SVG) with corresponding PNG rasterizations
98
+ - **Scale**: 15,663 CAD drawings (originally 10,094 in v1, updated to 15,663)
99
+ - **Categories**: 30 classes total
100
+ - 28 "thing" classes (countable instances): doors, windows, furniture, appliances, equipment
101
+ - 2 "stuff" classes (semantic regions): wall, parking
102
+ - **Annotation Type**: Line-grained primitive-level annotations with semantic and instance labels
103
+ - **Original Split**: 6,382 training / 3,712 testing drawings
104
+ - **Privacy Protected**: Cropped into 20m × 20m blocks, 50% retention rate, sensitive text removed
105
+
106
+ * **Curated by**: Zhiwen Fan†, Lingjie Zhu†, Honghua Li, Xiaohao Chen, Siyu Zhu, Ping Tan (Alibaba A.I. Labs & Simon Fraser University, †Equal contribution)
107
+ * **Funded by**: Alibaba A.I. Labs
108
+ * **Language(s)**: Not applicable (architectural vector graphics)
109
+ * **License**: Creative Commons Attribution-NonCommercial 4.0 License
110
+ * **Project shutdown notice**: As of January 2023, the project was shut down and most participants left the company
111
+
112
+ ### Dataset Sources
113
+
114
+ * **Repository**: https://floorplancad.github.io/ (Note: Project shut down in 2022)
115
+ * **Paper**: Fan et al. "FloorPlanCAD: A Large-Scale CAD Drawing Dataset for Panoptic Symbol Spotting" (ICCV 2021)
116
+
117
+ ## Uses
118
 
119
+ ### Direct Use
 
 
 
 
120
 
121
+ This dataset is designed for:
122
+ - **Panoptic symbol spotting**: Detecting both countable object instances and semantic regions in architectural drawings
123
+ - **Instance segmentation**: Identifying individual furniture, fixtures, and building elements
124
+ - **Semantic segmentation**: Recognizing structural elements like walls and parking areas
125
+ - **CAD drawing analysis**: Training models for automated floor plan understanding
126
+ - **Architecture/Engineering/Construction (AEC) applications**: Automated 3D modeling from 2D CAD drawings
127
 
128
+ ### Out-of-Scope Use
129
 
130
+ - **Commercial applications**: Dataset is licensed for non-commercial use only
131
+ - **Privacy-sensitive reconstruction**: The dataset is intentionally cropped and anonymized; attempting to reconstruct original complete floor plans or identify building locations violates privacy protections
132
+ - **As-is architectural design**: The cropped 20m × 20m blocks are not complete floor plans suitable for construction
133
 
134
+ ## Dataset Structure
135
 
136
+ The converted FiftyOne dataset contains the following structure:
137
+
138
+ ```text
139
+ <Sample: {
140
+ 'id': '690a547c0420c654cb79d521',
141
+ 'media_type': 'image',
142
+ 'filepath': '../image_data/0000-0003.png',
143
+ 'tags': [],
144
+ 'metadata': <ImageMetadata: {
145
+ 'size_bytes': 7803,
146
+ 'mime_type': 'image/png',
147
+ 'width': 1000,
148
+ 'height': 1000,
149
+ 'num_channels': 4,
150
+ }>,
151
+ 'created_at': datetime.datetime(2025, 11, 4, 19, 31, 8, 427000),
152
+ 'last_modified_at': datetime.datetime(2025, 11, 4, 19, 39, 58, 326000),
153
+ 'ground_truth': <Detections: {
154
+ 'detections': [
155
+ <Detection: {
156
+ 'id': '690a547c0420c654cb79d520',
157
+ 'attributes': {},
158
+ 'tags': [],
159
+ 'label': 'wall',
160
+ 'bounding_box': [0.30975255, 0.0, 0.69024745, 0.7205705549999999],
161
+ 'mask': array([[255, 255, 255, ..., 0, 0, 0],
162
+ [255, 255, 255, ..., 0, 0, 0],
163
+ [255, 255, 255, ..., 0, 0, 0],
164
+ ...,
165
+ [255, 255, 255, ..., 0, 0, 0],
166
+ [255, 255, 255, ..., 255, 255, 255],
167
+ [255, 255, 255, ..., 255, 255, 255]], dtype=uint8),
168
+ 'mask_path': None,
169
+ 'confidence': None,
170
+ 'index': None,
171
+ }>,
172
+ ],
173
+ }>,
174
+ }>
175
+ ```
176
 
177
+ **Object Categories (30 total):**
178
 
179
+ *Doors (3):* single_door, double_door, sliding_door
180
 
181
+ *Windows (4):* window, bay_window, blind_window, opening_symbol
182
 
183
+ *Stairs (1):* stair
184
 
185
+ *Home Appliances (3):* gas_stove, refrigerator, washing_machine
186
 
187
+ *Furniture (11):* sofa, bed, chair, table, bedside_cupboard, tv_cabinet, half_height_cabinet, high_cabinet, wardrobe, sink, bath
188
 
189
+ *Equipment (6):* bath_tub, squat_toilet, urinal, toilet, elevator, escalator
190
 
191
+ *Stuff Classes (2):* wall, parking
192
 
193
+ **Note on class distribution:** Wall and parking together account for ~27% of all annotated primitives. Significant class imbalance exists across categories.
194
 
195
  ## Dataset Creation
196
 
197
  ### Curation Rationale
198
 
199
+ The FloorPlanCAD dataset was created to address critical limitations in existing symbol spotting research:
200
 
201
+ 1. **Scale**: Previous datasets (SESYD with 1,000 synthetic plans, FPLAN-POLY with 42 plans) were too small for deep learning
202
+ 2. **Real-world diversity**: Prior datasets lacked the symbol variation seen across different architectural firms and building types
203
+ 3. **Vector graphics**: Maintaining CAD's native vector format (rather than rasterization) preserves accuracy and enables graph-based methods
204
+ 4. **Panoptic scope**: Traditional symbol spotting focused only on "thing" instances; this dataset includes "stuff" classes (walls, parking) for complete scene understanding
205
 
206
  ### Source Data
207
 
 
 
208
  #### Data Collection and Processing
209
 
210
+ **Original Data Sources:**
211
+ - 100+ architectural projects from production environments
212
+ - Multiple partner companies and institutions
213
+ - Building types: residential towers, schools, hospitals, shopping malls, office buildings
214
+ - Geographic diversity: Projects from various regions (layer names include Chinese characters indicating Asian sources)
215
 
216
+ **Technical Processing:**
217
+ - Multi-layer SVG organization (dozens of layers per drawing)
218
+ - Layer-by-layer annotation to reduce clutter
219
+ - Scale handling: Entity lengths range from millimeters to tens of meters (5+ orders of magnitude)
220
+ - Coordinate systems: Metric units (meters) for real-world measurements
221
 
222
  #### Who are the source data producers?
223
 
224
+ - **Primary producers**: Architects, engineers, and CAD designers from various companies creating production floor plans
225
+ - **Data providers**: Multiple partner companies and institutions in the AEC industry (anonymized for privacy)
226
+ - **Geographic origin**: Multinational (layer names suggest significant Asian representation)
227
+
228
+ ### Annotations
229
+
230
+ #### Annotation Process
231
+
232
+ **From SVG to FiftyOne Annotations:**
233
+
234
+ The conversion from vector SVG to structured annotations involves several stages:
235
+
236
+ 1. **SVG Primitive Parsing** (using `svgpathtools`):
237
+ ```python
238
+ # For each <path>, <circle>, <ellipse> element:
239
+ - Extract semantic-id (class label 1-35)
240
+ - Extract instance-id (unique instance number or -1 for stuff)
241
+ - Parse geometry:
242
+ * Paths: start point, end point, middle point via path.point(0.5)
243
+ * Circles: center (cx, cy), radius (r)
244
+ * Ellipses: center, radii (rx, ry)
245
+ ```
246
+
247
+ 2. **Coordinate Transformation**:
248
+ - SVG coordinates scaled by 10x to match PNG dimensions
249
+ - svg_x * 10 → png_x
250
+ - Maintains accurate spatial relationships
251
+
252
+ 3. **Instance Grouping**:
253
+ - Primitives grouped by (semantic_id, instance_id) tuple
254
+ - Each unique tuple represents one object instance
255
+ - "Stuff" classes have instance_id = -1 (no individual instances)
256
+
257
+ 4. **Bounding Box Computation**:
258
+ ```python
259
+ # For each instance:
260
+ - Collect all primitive endpoints and centers
261
+ - Compute axis-aligned bounding box:
262
+ x_min = min(all_x_coordinates)
263
+ y_min = min(all_y_coordinates)
264
+ width = x_max - x_min
265
+ height = y_max - y_min
266
+ - Normalize to [0, 1] by dividing by image dimensions
267
+ ```
268
+
269
+ 5. **Segmentation Mask Rendering**:
270
+ ```python
271
+ # For each instance:
272
+ - Create blank mask (image_height × image_width)
273
+ - Render each primitive with line_width=3 pixels:
274
+ * Paths: cv2.line() or cv2.polylines()
275
+ * Circles: cv2.circle()
276
+ * Ellipses: cv2.ellipse()
277
+ - Crop mask to bounding box region
278
+ ```
279
+
280
+ 6. **FiftyOne Detection Object Creation**:
281
+ - Each instance becomes `fo.Detection()` with:
282
+ * label: mapped class name (e.g., "wall", "single_door")
283
+ * bounding_box: normalized [x, y, w, h]
284
+ * mask: binary array (if include_masks=True)
285
 
 
 
 
 
 
 
 
 
 
286
 
287
  #### Who are the annotators?
288
 
289
+ - **Number**: 11 specialist annotators
290
+ - **Time investment**: Over 1,000 hours total annotation effort
291
+ - **Expertise**: Domain specialists familiar with architectural CAD drawings
292
+ - **Quality control**: Layer-by-layer annotation methodology for accuracy
293
+ - **Employer**: Alibaba A.I. Labs (annotation team)
294
+
295
+ ## Citation
296
+
297
+ ### BibTeX
298
+
299
+ ```bibtex
300
+ @InProceedings{Fan_2021_ICCV,
301
+ author = {Fan, Zhiwen and Zhu, Lingjie and Li, Honghua and Zhu, Siyu and Tan, Ping},
302
+ title = {FloorPlanCAD: A Large-Scale CAD Drawing Dataset for Panoptic Symbol Spotting},
303
+ booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
304
+ month = {October},
305
+ year = {2021},
306
+ pages = {10128-10137}
307
+ }
308
+ ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
309
 
310
+ ### APA
311
 
312
+ Fan, Z., Zhu, L., Li, H., Zhu, S., & Tan, P. (2021). FloorPlanCAD: A Large-Scale CAD Drawing Dataset for Panoptic Symbol Spotting. In *Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)* (pp. 10128-10137).