prasatee commited on
Commit
130d7a0
·
verified ·
1 Parent(s): ee9883f

Add files using upload-large-folder tool

Browse files
README.md ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ task_categories:
4
+ - image-segmentation
5
+ - object-detection
6
+ tags:
7
+ - P&ID
8
+ - lines
9
+ - pipelines
10
+ - engineering
11
+ - diagrams
12
+ - line-detection
13
+ size_categories:
14
+ - 1K<n<10K
15
+ ---
16
+
17
+ # P&ID Line Detection Dataset
18
+
19
+ This dataset contains cropped images from P&ID (Piping and Instrumentation Diagrams)
20
+ with line segment annotations for line detection and segmentation tasks.
21
+
22
+ ## Dataset Description
23
+
24
+ - **Total source images:** 500
25
+ - **Total cropped samples:** 10,000
26
+ - **Total line segments:** 87,908
27
+ - **Crops per image:** 20
28
+ - **Image sizes:** Various sizes under 1000px (e.g., 300x500, 512x768, 900x900)
29
+
30
+ ## Dataset Splits
31
+
32
+ The dataset is split by source image to prevent data leakage:
33
+
34
+ | Split | Source Images | Samples | Line Segments |
35
+ |------------|---------------|---------|---------------|
36
+ | Train | 400 (80%) | 8,000 | 69,483 |
37
+ | Validation | 50 (10%) | 1,000 | 9,485 |
38
+ | Test | 50 (10%) | 1,000 | 8,940 |
39
+
40
+ ## Dataset Structure
41
+
42
+ ```
43
+ lines_dataset/
44
+ ├── train/
45
+ │ ├── metadata.jsonl
46
+ │ └── *.jpg (8,000 images)
47
+ ├── validation/
48
+ │ ├── metadata.jsonl
49
+ │ └── *.jpg (1,000 images)
50
+ ├── test/
51
+ │ ├── metadata.jsonl
52
+ │ └── *.jpg (1,000 images)
53
+ └── visualizer/
54
+ └── (visualization app)
55
+ ```
56
+
57
+ Each sample contains:
58
+ - `file_name`: Image filename
59
+ - `source_image_idx`: Index of the original P&ID image
60
+ - `crop_idx`: Index of this crop from the source image
61
+ - `width`: Crop width in pixels
62
+ - `height`: Crop height in pixels
63
+ - `lines`: Dictionary with:
64
+ - `segments`: List of line segments as [x1, y1, x2, y2] (start and end points)
65
+ - `line_types`: List of line types ("solid" or "dashed")
66
+ - `pipelines`: List of pipeline names for each line
67
+
68
+ ## Usage
69
+
70
+ ```python
71
+ from datasets import load_dataset
72
+
73
+ # Load the dataset
74
+ dataset = load_dataset("imagefolder", data_dir="path/to/lines_dataset")
75
+
76
+ # Access splits
77
+ train_data = dataset["train"]
78
+ val_data = dataset["validation"]
79
+ test_data = dataset["test"]
80
+
81
+ # Access a sample
82
+ sample = train_data[0]
83
+ image = sample["image"]
84
+ lines = sample["lines"]
85
+ segments = lines["segments"] # [[x1, y1, x2, y2], ...]
86
+ line_types = lines["line_types"] # ["solid", "dashed", ...]
87
+ pipelines = lines["pipelines"] # ["5\"-EK-2648", ...]
88
+
89
+ # Draw lines on image
90
+ from PIL import ImageDraw
91
+ draw = ImageDraw.Draw(image)
92
+ for seg in segments:
93
+ x1, y1, x2, y2 = seg
94
+ draw.line([(x1, y1), (x2, y2)], fill="blue", width=3)
95
+ image.show()
96
+ ```
97
+
98
+ ## Visualizer
99
+
100
+ A built-in visualizer is included to explore the dataset:
101
+
102
+ ```bash
103
+ cd lines_dataset/visualizer
104
+ pip install flask
105
+ python app.py
106
+ ```
107
+
108
+ Then open http://localhost:5051 in your browser.
109
+
110
+ ## Line Segment Format
111
+
112
+ Each line segment is represented as `[x1, y1, x2, y2]` where:
113
+ - `(x1, y1)` is the start point
114
+ - `(x2, y2)` is the end point
115
+ - Coordinates are in pixels, relative to the cropped image
116
+ - Lines are **clipped** to crop boundaries - partial lines that extend beyond the crop are included with endpoints adjusted to the crop edges
117
+
118
+ ## Line Types
119
+
120
+ - `solid`: Continuous pipeline lines
121
+ - `dashed`: Dashed lines (often representing signal/instrument lines)
122
+
123
+ ## License
124
+
125
+ MIT License
split_dataset.py ADDED
@@ -0,0 +1,182 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Script to split the lines_dataset into train/validation/test splits.
3
+ - 80% train
4
+ - 10% validation
5
+ - 10% test
6
+
7
+ The split is done by source image to prevent data leakage - all crops from the
8
+ same source image go into the same split.
9
+ """
10
+
11
+ import json
12
+ import shutil
13
+ import random
14
+ from pathlib import Path
15
+ from collections import defaultdict
16
+
17
+ # Configuration
18
+ RANDOM_SEED = 42
19
+ TRAIN_RATIO = 0.8
20
+ VAL_RATIO = 0.1
21
+ TEST_RATIO = 0.1
22
+
23
+ BASE_DIR = Path(__file__).parent
24
+ CURRENT_TRAIN_DIR = BASE_DIR / "train"
25
+ METADATA_FILE = CURRENT_TRAIN_DIR / "metadata.jsonl"
26
+
27
+
28
+ def load_metadata():
29
+ """Load all metadata from the jsonl file."""
30
+ if not METADATA_FILE.exists():
31
+ print(f"Metadata file not found: {METADATA_FILE}")
32
+ return []
33
+
34
+ metadata = []
35
+ with open(METADATA_FILE, 'r') as f:
36
+ for line in f:
37
+ if line.strip():
38
+ metadata.append(json.loads(line))
39
+ return metadata
40
+
41
+
42
+ def group_by_source(metadata):
43
+ """Group samples by their source image index."""
44
+ groups = defaultdict(list)
45
+ for item in metadata:
46
+ source_idx = item.get("source_image_idx", 0)
47
+ groups[source_idx].append(item)
48
+ return groups
49
+
50
+
51
+ def split_sources(source_indices, train_ratio, val_ratio, test_ratio):
52
+ """Split source indices into train/val/test sets."""
53
+ random.shuffle(source_indices)
54
+
55
+ n = len(source_indices)
56
+ n_train = int(n * train_ratio)
57
+ n_val = int(n * val_ratio)
58
+
59
+ train_sources = source_indices[:n_train]
60
+ val_sources = source_indices[n_train:n_train + n_val]
61
+ test_sources = source_indices[n_train + n_val:]
62
+
63
+ return train_sources, val_sources, test_sources
64
+
65
+
66
+ def create_split_directory(split_name, samples, base_dir, source_dir):
67
+ """Create a split directory with images and metadata."""
68
+ split_dir = base_dir / split_name
69
+ split_dir.mkdir(parents=True, exist_ok=True)
70
+
71
+ # Copy images and prepare metadata
72
+ metadata_entries = []
73
+
74
+ for sample in samples:
75
+ file_name = sample["file_name"]
76
+ src_path = source_dir / file_name
77
+ dst_path = split_dir / file_name
78
+
79
+ if src_path.exists():
80
+ shutil.copy2(src_path, dst_path)
81
+ metadata_entries.append(sample)
82
+ else:
83
+ print(f"Warning: Image not found: {src_path}")
84
+
85
+ # Write metadata
86
+ metadata_path = split_dir / "metadata.jsonl"
87
+ with open(metadata_path, 'w') as f:
88
+ for entry in metadata_entries:
89
+ f.write(json.dumps(entry) + '\n')
90
+
91
+ return len(metadata_entries)
92
+
93
+
94
+ def main():
95
+ random.seed(RANDOM_SEED)
96
+
97
+ print("Loading metadata...")
98
+ metadata = load_metadata()
99
+ print(f"Total samples: {len(metadata)}")
100
+
101
+ # Group by source image
102
+ print("Grouping by source image...")
103
+ groups = group_by_source(metadata)
104
+ source_indices = list(groups.keys())
105
+ print(f"Total source images: {len(source_indices)}")
106
+
107
+ # Split source indices
108
+ print(f"\nSplitting sources ({TRAIN_RATIO:.0%} train, {VAL_RATIO:.0%} val, {TEST_RATIO:.0%} test)...")
109
+ train_sources, val_sources, test_sources = split_sources(
110
+ source_indices, TRAIN_RATIO, VAL_RATIO, TEST_RATIO
111
+ )
112
+
113
+ print(f" Train sources: {len(train_sources)}")
114
+ print(f" Validation sources: {len(val_sources)}")
115
+ print(f" Test sources: {len(test_sources)}")
116
+
117
+ # Gather samples for each split
118
+ train_samples = []
119
+ val_samples = []
120
+ test_samples = []
121
+
122
+ for src_idx in train_sources:
123
+ train_samples.extend(groups[src_idx])
124
+ for src_idx in val_sources:
125
+ val_samples.extend(groups[src_idx])
126
+ for src_idx in test_sources:
127
+ test_samples.extend(groups[src_idx])
128
+
129
+ print(f"\nSamples per split:")
130
+ print(f" Train: {len(train_samples)} ({len(train_samples)/len(metadata)*100:.1f}%)")
131
+ print(f" Validation: {len(val_samples)} ({len(val_samples)/len(metadata)*100:.1f}%)")
132
+ print(f" Test: {len(test_samples)} ({len(test_samples)/len(metadata)*100:.1f}%)")
133
+
134
+ # Create temporary directory for new structure
135
+ temp_dir = BASE_DIR / "_temp_splits"
136
+ temp_dir.mkdir(exist_ok=True)
137
+
138
+ print("\nCreating split directories...")
139
+
140
+ # Create each split
141
+ n_train = create_split_directory("train", train_samples, temp_dir, CURRENT_TRAIN_DIR)
142
+ print(f" Created train split: {n_train} samples")
143
+
144
+ n_val = create_split_directory("validation", val_samples, temp_dir, CURRENT_TRAIN_DIR)
145
+ print(f" Created validation split: {n_val} samples")
146
+
147
+ n_test = create_split_directory("test", test_samples, temp_dir, CURRENT_TRAIN_DIR)
148
+ print(f" Created test split: {n_test} samples")
149
+
150
+ # Remove old train directory and move new splits
151
+ print("\nReorganizing directory structure...")
152
+
153
+ # Remove old train directory
154
+ shutil.rmtree(CURRENT_TRAIN_DIR)
155
+
156
+ # Move new splits from temp to base
157
+ for split_name in ["train", "validation", "test"]:
158
+ src = temp_dir / split_name
159
+ dst = BASE_DIR / split_name
160
+ shutil.move(str(src), str(dst))
161
+
162
+ # Remove temp directory
163
+ temp_dir.rmdir()
164
+
165
+ print("\nDone! New directory structure:")
166
+ print(f" {BASE_DIR}/train/ ({n_train} samples)")
167
+ print(f" {BASE_DIR}/validation/ ({n_val} samples)")
168
+ print(f" {BASE_DIR}/test/ ({n_test} samples)")
169
+
170
+ # Count total lines per split
171
+ def count_lines(samples):
172
+ return sum(len(s.get("lines", {}).get("segments", [])) for s in samples)
173
+
174
+ print(f"\nTotal lines per split:")
175
+ print(f" Train: {count_lines(train_samples)}")
176
+ print(f" Validation: {count_lines(val_samples)}")
177
+ print(f" Test: {count_lines(test_samples)}")
178
+
179
+
180
+ if __name__ == "__main__":
181
+ main()
182
+
test/111_3.jpg ADDED

Git LFS Details

  • SHA256: 44e025d9f1c2021f99f28dd3d4e3cc6bdfc5562b6aa6fd84f05872ed71e035fa
  • Pointer size: 130 Bytes
  • Size of remote file: 15.7 kB
test/112_18.jpg ADDED

Git LFS Details

  • SHA256: 8e41d511e758fae03c984c18c13a1707f2182a635cecf5c1e8f9b80ca25903b0
  • Pointer size: 130 Bytes
  • Size of remote file: 23.5 kB
test/119_11.jpg ADDED

Git LFS Details

  • SHA256: 8d30172f1abdcf747b5b04bd3fe069d0fd3d8522ea2a86057c62c34a7beacf87
  • Pointer size: 130 Bytes
  • Size of remote file: 16.1 kB
test/119_13.jpg ADDED

Git LFS Details

  • SHA256: 5a09e2749582aa2687c96965d2d4657f3c8b5f74445fc2d8a45ee032a590775f
  • Pointer size: 130 Bytes
  • Size of remote file: 16.3 kB
test/13_10.jpg ADDED

Git LFS Details

  • SHA256: 7d4a57508ab2d7d4c480a34783676b935a567264215c1b08f150b8b928fd6522
  • Pointer size: 130 Bytes
  • Size of remote file: 38.1 kB
test/13_12.jpg ADDED

Git LFS Details

  • SHA256: 26bbd8ee1ca21427b2de197a76d52bc962be7d109219739d2e0d5054e650faee
  • Pointer size: 130 Bytes
  • Size of remote file: 13.5 kB
test/140_10.jpg ADDED

Git LFS Details

  • SHA256: 5907e7ef093c4eb152e83f025a3c601ac2f85464e7c31a4e7643aa78dc652991
  • Pointer size: 130 Bytes
  • Size of remote file: 16.9 kB
test/142_15.jpg ADDED

Git LFS Details

  • SHA256: eff36243f1b3be46949805e4618e41b152b866ee52a297cd45c68aeebcdb5d1c
  • Pointer size: 130 Bytes
  • Size of remote file: 39.8 kB
test/142_17.jpg ADDED

Git LFS Details

  • SHA256: 8ae7ddc7e00c0f9cf8bf1c9f9a823f02387038681e88bd6531816afb7b9b96a5
  • Pointer size: 130 Bytes
  • Size of remote file: 39.2 kB
test/15_9.jpg ADDED

Git LFS Details

  • SHA256: 67df004301cd3366b7b83feb086fff5ec128da3bb3632abff04386b793bdcd6c
  • Pointer size: 130 Bytes
  • Size of remote file: 27.3 kB
test/172_0.jpg ADDED

Git LFS Details

  • SHA256: fb5482e5990bd431d30a9401558507abb81416abbcb5fee9efbe8aa3653516a7
  • Pointer size: 129 Bytes
  • Size of remote file: 3.84 kB
test/174_14.jpg ADDED

Git LFS Details

  • SHA256: c8ab5a74dc94c45d72c0ac82911c34e0114c38dbe426280367a25894e230e681
  • Pointer size: 130 Bytes
  • Size of remote file: 43.3 kB
test/174_16.jpg ADDED

Git LFS Details

  • SHA256: 02e6ea7b6e021dad9450aa996f306effe8980474ac295643a998cb624bda8a22
  • Pointer size: 130 Bytes
  • Size of remote file: 65.1 kB
test/174_6.jpg ADDED

Git LFS Details

  • SHA256: 9f5ac57bef45870c1d27c614aeeae1efb380e56e13d7994ff1d7584dd14a4766
  • Pointer size: 129 Bytes
  • Size of remote file: 4.25 kB
test/214_14.jpg ADDED

Git LFS Details

  • SHA256: 52fbc9d773a85e66f42eb3984b3e230a8372781f5cda5f85bcbd6aa7ee13215d
  • Pointer size: 129 Bytes
  • Size of remote file: 3.99 kB
test/214_16.jpg ADDED

Git LFS Details

  • SHA256: 18f0b402abbb3f572f6a72a4db8db2a403b8ef670be8a619a2887aa5f3be56ae
  • Pointer size: 130 Bytes
  • Size of remote file: 18.7 kB
test/214_7.jpg ADDED

Git LFS Details

  • SHA256: 016da85b091641bc8feae04ee8545d9337cec66d58b5b26b6ad818f150a03542
  • Pointer size: 130 Bytes
  • Size of remote file: 32.8 kB
test/216_13.jpg ADDED

Git LFS Details

  • SHA256: ce146e76c3822944bf964c423b327bc09dd0397a98af472398cda2f3a593bd29
  • Pointer size: 130 Bytes
  • Size of remote file: 16.3 kB
test/216_5.jpg ADDED

Git LFS Details

  • SHA256: d340b3328af4cd75d0b6d22ba89ca3b0e7752fe1ef609d7288c179d2f64b0acb
  • Pointer size: 130 Bytes
  • Size of remote file: 81.3 kB
test/229_5.jpg ADDED

Git LFS Details

  • SHA256: 54a9ed824b054af0e5e7e9a6690e496777c5da1ecf51b76d1b4f96ae4bec7d60
  • Pointer size: 130 Bytes
  • Size of remote file: 22.6 kB
test/229_7.jpg ADDED

Git LFS Details

  • SHA256: 69d0b82867dd517aebab6b83c731e382543b431a59d0bc9d20e4a9f7069ec600
  • Pointer size: 130 Bytes
  • Size of remote file: 26.4 kB
test/279_13.jpg ADDED

Git LFS Details

  • SHA256: ef3dd49e8055b4f147e7e78359f80945049222f109d4c72315d1eeeff573c34b
  • Pointer size: 130 Bytes
  • Size of remote file: 65.1 kB
test/308_13.jpg ADDED

Git LFS Details

  • SHA256: e54c69a1c6a48a2afe3da7734848f86e3740794676dfb7b120fa38be215510a3
  • Pointer size: 130 Bytes
  • Size of remote file: 20.5 kB
test/332_1.jpg ADDED

Git LFS Details

  • SHA256: 45a3b9d97f40ee3a26cf3e22f8255bc887309da4178375a3a193299a16c1478b
  • Pointer size: 130 Bytes
  • Size of remote file: 13.4 kB
test/332_10.jpg ADDED

Git LFS Details

  • SHA256: 047bfbf2dee5b354cf9c5ba3fca0ce871015d7d7b659ce3c5d757b5d8f58f28e
  • Pointer size: 129 Bytes
  • Size of remote file: 8.16 kB
test/332_12.jpg ADDED

Git LFS Details

  • SHA256: d842756c8ee8ae6551df39682197196432ffcd9f784b8b44787482f6b02b5a04
  • Pointer size: 129 Bytes
  • Size of remote file: 6.99 kB
test/357_4.jpg ADDED

Git LFS Details

  • SHA256: 2916b1fd9d5fc31c0c5dbf14294362d9a2690474a452bf720160dc76750ce065
  • Pointer size: 130 Bytes
  • Size of remote file: 30.6 kB
test/377_1.jpg ADDED

Git LFS Details

  • SHA256: ea811ccd2df76abc789050a9f47d82597253d5b3f22a36b64954dc497ddbeb65
  • Pointer size: 129 Bytes
  • Size of remote file: 8.06 kB
test/377_14.jpg ADDED

Git LFS Details

  • SHA256: e347ee5f306dd19b4c09e89d7eede21c35cf84c93820d442c327f72bd798a9e5
  • Pointer size: 130 Bytes
  • Size of remote file: 26.1 kB
test/377_16.jpg ADDED

Git LFS Details

  • SHA256: eb3a581c0bcd4866bb6ecfbb743aa9e1835c3306875e65a85bac1e5cf3e96904
  • Pointer size: 130 Bytes
  • Size of remote file: 50.5 kB
test/379_13.jpg ADDED

Git LFS Details

  • SHA256: e2b8bdabdf08a72d51aa4f71bb9928de36d7432aa0ae0895a9304ceb4cb124ce
  • Pointer size: 130 Bytes
  • Size of remote file: 61 kB
test/388_15.jpg ADDED

Git LFS Details

  • SHA256: aa127f9d7627e7f36721eccd9a4a247ae5e4b6d2f4fa7e694cfaf22af1c67d09
  • Pointer size: 130 Bytes
  • Size of remote file: 29.2 kB
test/388_3.jpg ADDED

Git LFS Details

  • SHA256: c6c3fa5329446e20976cee1be4d244c4fa5a66f5ff8aa31b37445cb4cafffdf3
  • Pointer size: 130 Bytes
  • Size of remote file: 22.3 kB
test/3_0.jpg ADDED

Git LFS Details

  • SHA256: 90a2fde3c5f9ee4699ec9d18f4858088a5b295cdd823169dfc492f43f4be5292
  • Pointer size: 129 Bytes
  • Size of remote file: 8.81 kB
test/412_12.jpg ADDED

Git LFS Details

  • SHA256: 179d0e649117661a2c8e89b821c817dbd18e7a137edf7619b0d59a87f64145e7
  • Pointer size: 130 Bytes
  • Size of remote file: 25 kB
test/445_14.jpg ADDED

Git LFS Details

  • SHA256: 7a7872ef617e4d20cae36b6964e12b03f1f8baf6b10283b317a6c6b51cebc312
  • Pointer size: 130 Bytes
  • Size of remote file: 24.8 kB
test/445_16.jpg ADDED

Git LFS Details

  • SHA256: c170ea79718624ce68ae2b6a9aee3c9b5cf3757d5e7e6ca3b73e65f90f12a2b5
  • Pointer size: 130 Bytes
  • Size of remote file: 12.2 kB
test/445_6.jpg ADDED

Git LFS Details

  • SHA256: c51951b6aaaea011999a62a9b040acc3ed66d6bba10f627d8bb99228751b9491
  • Pointer size: 129 Bytes
  • Size of remote file: 7.58 kB
test/44_14.jpg ADDED

Git LFS Details

  • SHA256: a99cf3e21aee4cd80468a735dcb8fb150dbc30bc929b20b68ab4ef5aa8ff92e0
  • Pointer size: 130 Bytes
  • Size of remote file: 15.9 kB
test/464_19.jpg ADDED

Git LFS Details

  • SHA256: 9687a9c92946ae83a7aab811d73f38f421276bad0593b34eab41577cfd10de07
  • Pointer size: 130 Bytes
  • Size of remote file: 24.5 kB
test/47_7.jpg ADDED

Git LFS Details

  • SHA256: 40a9e9c8084a2dc19bafdff8f60272dd374e7ceb198d056491e5f5a9a92aad3c
  • Pointer size: 130 Bytes
  • Size of remote file: 43.8 kB
test/483_0.jpg ADDED

Git LFS Details

  • SHA256: c32533ecad5eba182380cce8923acd493eb2fa589eb775c47dbe9dcd71abb559
  • Pointer size: 130 Bytes
  • Size of remote file: 42.6 kB
test/52_9.jpg ADDED

Git LFS Details

  • SHA256: fae8df013229661d6a81c8f57db65e9175e07a691b274489390d385180c7a1f5
  • Pointer size: 130 Bytes
  • Size of remote file: 12.7 kB
test/81_1.jpg ADDED

Git LFS Details

  • SHA256: b3909db9687f1c0a10dee090c423a0550d7f92bbe5be7b5409613a15800d7c86
  • Pointer size: 129 Bytes
  • Size of remote file: 8.29 kB
test/81_14.jpg ADDED

Git LFS Details

  • SHA256: eb96bd74bbee2e059c67c95220e946e9a6bca76a55c5d90a041855657228e863
  • Pointer size: 130 Bytes
  • Size of remote file: 11.4 kB
test/81_16.jpg ADDED

Git LFS Details

  • SHA256: 89ed84bd849ffd41dba045850ca45441ef657b5b5efd797550f7d758e8bf8ae5
  • Pointer size: 130 Bytes
  • Size of remote file: 42.5 kB
test/metadata.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
train/metadata.jsonl ADDED
The diff for this file is too large to render. See raw diff