Upload folder using huggingface_hub
Browse files- .gitattributes +14 -0
- args.yaml +105 -0
- confusion_matrix.png +0 -0
- confusion_matrix_normalized.png +3 -0
- results.csv +101 -0
- results.png +3 -0
- train_batch0.jpg +3 -0
- train_batch1.jpg +3 -0
- train_batch1440.jpg +3 -0
- train_batch1441.jpg +3 -0
- train_batch1442.jpg +3 -0
- train_batch2.jpg +3 -0
- val_batch0_labels.jpg +3 -0
- val_batch0_pred.jpg +3 -0
- val_batch1_labels.jpg +3 -0
- val_batch1_pred.jpg +3 -0
- val_batch2_labels.jpg +3 -0
- val_batch2_pred.jpg +3 -0
- weights/best.pt +3 -0
- weights/last.pt +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
confusion_matrix_normalized.png filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
results.png filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
train_batch0.jpg filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
train_batch1.jpg filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
train_batch1440.jpg filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
train_batch1441.jpg filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
train_batch1442.jpg filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
train_batch2.jpg filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
val_batch0_labels.jpg filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
val_batch0_pred.jpg filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
val_batch1_labels.jpg filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
val_batch1_pred.jpg filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
val_batch2_labels.jpg filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
val_batch2_pred.jpg filter=lfs diff=lfs merge=lfs -text
|
args.yaml
ADDED
|
@@ -0,0 +1,105 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
task: classify
|
| 2 |
+
mode: train
|
| 3 |
+
model: yolo11n-cls.pt
|
| 4 |
+
data: dataset
|
| 5 |
+
epochs: 100
|
| 6 |
+
time: null
|
| 7 |
+
patience: 100
|
| 8 |
+
batch: 64
|
| 9 |
+
imgsz: 448
|
| 10 |
+
save: true
|
| 11 |
+
save_period: -1
|
| 12 |
+
cache: false
|
| 13 |
+
device: null
|
| 14 |
+
workers: 8
|
| 15 |
+
project: null
|
| 16 |
+
name: train
|
| 17 |
+
exist_ok: false
|
| 18 |
+
pretrained: true
|
| 19 |
+
optimizer: auto
|
| 20 |
+
verbose: true
|
| 21 |
+
seed: 0
|
| 22 |
+
deterministic: true
|
| 23 |
+
single_cls: false
|
| 24 |
+
rect: false
|
| 25 |
+
cos_lr: false
|
| 26 |
+
close_mosaic: 10
|
| 27 |
+
resume: false
|
| 28 |
+
amp: true
|
| 29 |
+
fraction: 1.0
|
| 30 |
+
profile: false
|
| 31 |
+
freeze: null
|
| 32 |
+
multi_scale: false
|
| 33 |
+
overlap_mask: true
|
| 34 |
+
mask_ratio: 4
|
| 35 |
+
dropout: 0.0
|
| 36 |
+
val: true
|
| 37 |
+
split: val
|
| 38 |
+
save_json: false
|
| 39 |
+
conf: null
|
| 40 |
+
iou: 0.7
|
| 41 |
+
max_det: 300
|
| 42 |
+
half: false
|
| 43 |
+
dnn: false
|
| 44 |
+
plots: true
|
| 45 |
+
source: null
|
| 46 |
+
vid_stride: 1
|
| 47 |
+
stream_buffer: false
|
| 48 |
+
visualize: false
|
| 49 |
+
augment: false
|
| 50 |
+
agnostic_nms: false
|
| 51 |
+
classes: null
|
| 52 |
+
retina_masks: false
|
| 53 |
+
embed: null
|
| 54 |
+
show: false
|
| 55 |
+
save_frames: false
|
| 56 |
+
save_txt: false
|
| 57 |
+
save_conf: false
|
| 58 |
+
save_crop: false
|
| 59 |
+
show_labels: true
|
| 60 |
+
show_conf: true
|
| 61 |
+
show_boxes: true
|
| 62 |
+
line_width: null
|
| 63 |
+
format: torchscript
|
| 64 |
+
keras: false
|
| 65 |
+
optimize: false
|
| 66 |
+
int8: false
|
| 67 |
+
dynamic: false
|
| 68 |
+
simplify: true
|
| 69 |
+
opset: null
|
| 70 |
+
workspace: null
|
| 71 |
+
nms: false
|
| 72 |
+
lr0: 0.01
|
| 73 |
+
lrf: 0.01
|
| 74 |
+
momentum: 0.937
|
| 75 |
+
weight_decay: 0.0005
|
| 76 |
+
warmup_epochs: 3.0
|
| 77 |
+
warmup_momentum: 0.8
|
| 78 |
+
warmup_bias_lr: 0.1
|
| 79 |
+
box: 7.5
|
| 80 |
+
cls: 0.5
|
| 81 |
+
dfl: 1.5
|
| 82 |
+
pose: 12.0
|
| 83 |
+
kobj: 1.0
|
| 84 |
+
nbs: 64
|
| 85 |
+
hsv_h: 0.015
|
| 86 |
+
hsv_s: 0.7
|
| 87 |
+
hsv_v: 0.4
|
| 88 |
+
degrees: 0.0
|
| 89 |
+
translate: 0.1
|
| 90 |
+
scale: 0.5
|
| 91 |
+
shear: 0.0
|
| 92 |
+
perspective: 0.0
|
| 93 |
+
flipud: 0.0
|
| 94 |
+
fliplr: 0.5
|
| 95 |
+
bgr: 0.0
|
| 96 |
+
mosaic: 1.0
|
| 97 |
+
mixup: 0.0
|
| 98 |
+
cutmix: 0.0
|
| 99 |
+
copy_paste: 0.0
|
| 100 |
+
copy_paste_mode: flip
|
| 101 |
+
auto_augment: randaugment
|
| 102 |
+
erasing: 0.4
|
| 103 |
+
cfg: null
|
| 104 |
+
tracker: botsort.yaml
|
| 105 |
+
save_dir: runs/classify/train
|
confusion_matrix.png
ADDED
|
confusion_matrix_normalized.png
ADDED
|
Git LFS Details
|
results.csv
ADDED
|
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,time,train/loss,metrics/accuracy_top1,metrics/accuracy_top5,val/loss,lr/pg0,lr/pg1,lr/pg2
|
| 2 |
+
1,13.7442,0.5474,0.8825,1,0.23692,0.00025005,0.00025005,0.00025005
|
| 3 |
+
2,18.6025,0.223,0.93,1,0.20056,0.000511654,0.000511654,0.000511654
|
| 4 |
+
3,22.6051,0.13535,0.93,1,0.19147,0.000767977,0.000767977,0.000767977
|
| 5 |
+
4,26.2284,0.14566,0.9025,1,0.23131,0.00101902,0.00101902,0.00101902
|
| 6 |
+
5,30.0531,0.14428,0.84,1,0.48279,0.00126478,0.00126478,0.00126478
|
| 7 |
+
6,33.6127,0.16988,0.91,1,0.32916,0.00150526,0.00150526,0.00150526
|
| 8 |
+
7,38.1383,0.1653,0.9025,1,0.44763,0.00156798,0.00156798,0.00156798
|
| 9 |
+
8,42.2768,0.17871,0.915,1,0.34949,0.00155148,0.00155148,0.00155148
|
| 10 |
+
9,45.8697,0.1518,0.6125,1,2.24731,0.00153497,0.00153497,0.00153497
|
| 11 |
+
10,49.1888,0.13781,0.9275,1,0.1926,0.00151847,0.00151847,0.00151847
|
| 12 |
+
11,52.5864,0.13231,0.93,1,0.26221,0.00150197,0.00150197,0.00150197
|
| 13 |
+
12,56.2213,0.10945,0.9075,1,0.28439,0.00148546,0.00148546,0.00148546
|
| 14 |
+
13,60.2303,0.13172,0.9225,1,0.17477,0.00146896,0.00146896,0.00146896
|
| 15 |
+
14,64.5234,0.14571,0.92,1,0.24792,0.00145246,0.00145246,0.00145246
|
| 16 |
+
15,68.3285,0.1254,0.9175,1,0.17751,0.00143595,0.00143595,0.00143595
|
| 17 |
+
16,72.3052,0.09455,0.9275,1,0.20151,0.00141945,0.00141945,0.00141945
|
| 18 |
+
17,77.2002,0.10403,0.9025,1,0.33157,0.00140295,0.00140295,0.00140295
|
| 19 |
+
18,81.1883,0.08363,0.8775,1,0.42993,0.00138644,0.00138644,0.00138644
|
| 20 |
+
19,85.329,0.0805,0.935,1,0.28903,0.00136994,0.00136994,0.00136994
|
| 21 |
+
20,88.5917,0.06396,0.9175,1,0.18219,0.00135344,0.00135344,0.00135344
|
| 22 |
+
21,92.2626,0.05752,0.935,1,0.2247,0.00133693,0.00133693,0.00133693
|
| 23 |
+
22,95.6249,0.05155,0.925,1,0.15441,0.00132043,0.00132043,0.00132043
|
| 24 |
+
23,99.8108,0.05892,0.93,1,0.20683,0.00130393,0.00130393,0.00130393
|
| 25 |
+
24,103.099,0.03977,0.94,1,0.23221,0.00128742,0.00128742,0.00128742
|
| 26 |
+
25,106.621,0.0431,0.93,1,0.26459,0.00127092,0.00127092,0.00127092
|
| 27 |
+
26,110.412,0.02385,0.9225,1,0.29141,0.00125442,0.00125442,0.00125442
|
| 28 |
+
27,113.574,0.03993,0.9,1,0.48111,0.00123791,0.00123791,0.00123791
|
| 29 |
+
28,117.808,0.06877,0.91,1,0.33588,0.00122141,0.00122141,0.00122141
|
| 30 |
+
29,121.081,0.06794,0.85,1,0.47186,0.00120491,0.00120491,0.00120491
|
| 31 |
+
30,124.214,0.05524,0.925,1,0.21754,0.0011884,0.0011884,0.0011884
|
| 32 |
+
31,128.483,0.04108,0.9275,1,0.24187,0.0011719,0.0011719,0.0011719
|
| 33 |
+
32,131.63,0.03585,0.93,1,0.38281,0.0011554,0.0011554,0.0011554
|
| 34 |
+
33,135.348,0.02222,0.9375,1,0.25282,0.00113889,0.00113889,0.00113889
|
| 35 |
+
34,138.935,0.06157,0.9225,1,0.34286,0.00112239,0.00112239,0.00112239
|
| 36 |
+
35,142.543,0.03983,0.9325,1,0.2171,0.00110589,0.00110589,0.00110589
|
| 37 |
+
36,146.474,0.02118,0.935,1,0.1955,0.00108938,0.00108938,0.00108938
|
| 38 |
+
37,149.78,0.0157,0.915,1,0.39688,0.00107288,0.00107288,0.00107288
|
| 39 |
+
38,153.654,0.02329,0.9275,1,0.26883,0.00105638,0.00105638,0.00105638
|
| 40 |
+
39,157.288,0.01694,0.9375,1,0.2084,0.00103987,0.00103987,0.00103987
|
| 41 |
+
40,160.895,0.01443,0.9275,1,0.22192,0.00102337,0.00102337,0.00102337
|
| 42 |
+
41,164.25,0.01364,0.9425,1,0.22223,0.00100687,0.00100687,0.00100687
|
| 43 |
+
42,167.854,0.0148,0.9375,1,0.37137,0.000990365,0.000990365,0.000990365
|
| 44 |
+
43,171.574,0.01727,0.925,1,0.29317,0.000973861,0.000973861,0.000973861
|
| 45 |
+
44,174.858,0.01027,0.9425,1,0.21136,0.000957358,0.000957358,0.000957358
|
| 46 |
+
45,178.923,0.01504,0.9325,1,0.54095,0.000940855,0.000940855,0.000940855
|
| 47 |
+
46,183.4,0.01515,0.92,1,0.32422,0.000924352,0.000924352,0.000924352
|
| 48 |
+
47,186.642,0.01672,0.945,1,0.44019,0.000907848,0.000907848,0.000907848
|
| 49 |
+
48,189.997,0.00696,0.9325,1,0.27787,0.000891345,0.000891345,0.000891345
|
| 50 |
+
49,194.009,0.00546,0.945,1,0.26839,0.000874842,0.000874842,0.000874842
|
| 51 |
+
50,197.906,0.01493,0.9325,1,0.41754,0.000858338,0.000858338,0.000858338
|
| 52 |
+
51,201.729,0.01219,0.9225,1,0.44385,0.000841835,0.000841835,0.000841835
|
| 53 |
+
52,205.3,0.02749,0.925,1,0.51117,0.000825332,0.000825332,0.000825332
|
| 54 |
+
53,208.746,0.02109,0.9275,1,0.36407,0.000808828,0.000808828,0.000808828
|
| 55 |
+
54,212.44,0.01382,0.915,1,0.52063,0.000792325,0.000792325,0.000792325
|
| 56 |
+
55,216.071,0.01026,0.94,1,0.27217,0.000775822,0.000775822,0.000775822
|
| 57 |
+
56,219.406,0.02082,0.915,1,0.43329,0.000759318,0.000759318,0.000759318
|
| 58 |
+
57,223.056,0.03126,0.935,1,0.34738,0.000742815,0.000742815,0.000742815
|
| 59 |
+
58,226.238,0.01668,0.9225,1,0.26819,0.000726312,0.000726312,0.000726312
|
| 60 |
+
59,230.166,0.01366,0.935,1,0.59033,0.000709809,0.000709809,0.000709809
|
| 61 |
+
60,234.263,0.01286,0.925,1,0.25999,0.000693305,0.000693305,0.000693305
|
| 62 |
+
61,237.663,0.00735,0.93,1,0.2685,0.000676802,0.000676802,0.000676802
|
| 63 |
+
62,240.959,0.00943,0.93,1,0.28235,0.000660299,0.000660299,0.000660299
|
| 64 |
+
63,244.919,0.00315,0.9375,1,0.26724,0.000643795,0.000643795,0.000643795
|
| 65 |
+
64,249.316,0.00576,0.9425,1,0.23121,0.000627292,0.000627292,0.000627292
|
| 66 |
+
65,252.642,0.00172,0.945,1,0.25946,0.000610789,0.000610789,0.000610789
|
| 67 |
+
66,256.813,0.00303,0.9425,1,0.40546,0.000594286,0.000594286,0.000594286
|
| 68 |
+
67,260.062,0.00174,0.9525,1,0.27037,0.000577782,0.000577782,0.000577782
|
| 69 |
+
68,263.395,0.00126,0.9525,1,0.26799,0.000561279,0.000561279,0.000561279
|
| 70 |
+
69,266.703,0.00283,0.9475,1,0.33087,0.000544776,0.000544776,0.000544776
|
| 71 |
+
70,270.76,0.00092,0.945,1,0.32608,0.000528272,0.000528272,0.000528272
|
| 72 |
+
71,274.487,0.00848,0.95,1,0.19709,0.000511769,0.000511769,0.000511769
|
| 73 |
+
72,277.95,0.00266,0.9375,1,0.28043,0.000495266,0.000495266,0.000495266
|
| 74 |
+
73,281.396,0.00246,0.945,1,0.21303,0.000478762,0.000478762,0.000478762
|
| 75 |
+
74,284.67,0.00378,0.94,1,0.5094,0.000462259,0.000462259,0.000462259
|
| 76 |
+
75,288.229,0.00137,0.94,1,0.34906,0.000445756,0.000445756,0.000445756
|
| 77 |
+
76,291.937,0.00114,0.94,1,0.32184,0.000429253,0.000429253,0.000429253
|
| 78 |
+
77,295.36,0.00325,0.9375,1,0.39532,0.000412749,0.000412749,0.000412749
|
| 79 |
+
78,298.6,0.00235,0.9375,1,0.25989,0.000396246,0.000396246,0.000396246
|
| 80 |
+
79,302.213,0.00117,0.9375,1,0.26505,0.000379743,0.000379743,0.000379743
|
| 81 |
+
80,305.873,0.00049,0.94,1,0.4093,0.000363239,0.000363239,0.000363239
|
| 82 |
+
81,309.497,0.00051,0.9375,1,0.36206,0.000346736,0.000346736,0.000346736
|
| 83 |
+
82,312.897,0.00126,0.945,1,0.49448,0.000330233,0.000330233,0.000330233
|
| 84 |
+
83,316.347,0.00098,0.945,1,0.24756,0.000313729,0.000313729,0.000313729
|
| 85 |
+
84,319.728,0.00242,0.94,1,0.48669,0.000297226,0.000297226,0.000297226
|
| 86 |
+
85,323.036,0.00079,0.945,1,0.22728,0.000280723,0.000280723,0.000280723
|
| 87 |
+
86,326.247,0.00064,0.945,1,0.27292,0.00026422,0.00026422,0.00026422
|
| 88 |
+
87,330.128,0.00186,0.9475,1,0.32642,0.000247716,0.000247716,0.000247716
|
| 89 |
+
88,333.542,0.0002,0.9475,1,0.22529,0.000231213,0.000231213,0.000231213
|
| 90 |
+
89,336.786,0.00038,0.945,1,0.23324,0.00021471,0.00021471,0.00021471
|
| 91 |
+
90,340.359,0.00024,0.9475,1,0.3385,0.000198206,0.000198206,0.000198206
|
| 92 |
+
91,347.23,0.00019,0.9475,1,0.22037,0.000181703,0.000181703,0.000181703
|
| 93 |
+
92,351.153,0.00042,0.9475,1,0.28093,0.0001652,0.0001652,0.0001652
|
| 94 |
+
93,354.712,0.00035,0.945,1,0.25102,0.000148696,0.000148696,0.000148696
|
| 95 |
+
94,358.447,0.00119,0.9425,1,0.22487,0.000132193,0.000132193,0.000132193
|
| 96 |
+
95,362.368,0.00041,0.9425,1,0.22438,0.00011569,0.00011569,0.00011569
|
| 97 |
+
96,366.387,0.00026,0.945,1,0.35329,9.91865e-05,9.91865e-05,9.91865e-05
|
| 98 |
+
97,370.231,0.00032,0.9425,1,0.22662,8.26832e-05,8.26832e-05,8.26832e-05
|
| 99 |
+
98,373.642,0.00034,0.9475,1,0.23278,6.61799e-05,6.61799e-05,6.61799e-05
|
| 100 |
+
99,377.614,0.00368,0.945,1,0.59402,4.96766e-05,4.96766e-05,4.96766e-05
|
| 101 |
+
100,382.129,0.00025,0.945,1,0.23099,3.31733e-05,3.31733e-05,3.31733e-05
|
results.png
ADDED
|
Git LFS Details
|
train_batch0.jpg
ADDED
|
Git LFS Details
|
train_batch1.jpg
ADDED
|
Git LFS Details
|
train_batch1440.jpg
ADDED
|
Git LFS Details
|
train_batch1441.jpg
ADDED
|
Git LFS Details
|
train_batch1442.jpg
ADDED
|
Git LFS Details
|
train_batch2.jpg
ADDED
|
Git LFS Details
|
val_batch0_labels.jpg
ADDED
|
Git LFS Details
|
val_batch0_pred.jpg
ADDED
|
Git LFS Details
|
val_batch1_labels.jpg
ADDED
|
Git LFS Details
|
val_batch1_pred.jpg
ADDED
|
Git LFS Details
|
val_batch2_labels.jpg
ADDED
|
Git LFS Details
|
val_batch2_pred.jpg
ADDED
|
Git LFS Details
|
weights/best.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ba3bcdf17c8466e342b18a830573cf72502a859cd4a16c1351c0491e22a6a730
|
| 3 |
+
size 3212044
|
weights/last.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:19ae1c7d0d75e6bb57fbf05c645fa8f3bc710b2ced6981246b43af6c6b841948
|
| 3 |
+
size 3214156
|