JohnWeck commited on
Commit
b068650
·
verified ·
1 Parent(s): 42f9d1b

Upload 5 files

Browse files
Dataset004_CVC/nnUNetTrainer__nnUNetPlans__2d/dataset.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "channel_names": {
3
+ "0": "R",
4
+ "1": "G",
5
+ "2": "B"
6
+ },
7
+ "labels": {
8
+ "background": 0,
9
+ "polyp": 1
10
+ },
11
+ "numTraining": 1100,
12
+ "file_ending": ".png",
13
+ "name": "Dataset004_CVC",
14
+ "num_test_cases": 62
15
+ }
Dataset004_CVC/nnUNetTrainer__nnUNetPlans__2d/dataset_fingerprint.json ADDED
The diff for this file is too large to render. See raw diff
 
Dataset004_CVC/nnUNetTrainer__nnUNetPlans__2d/fold_5/checkpoint_best.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9cdc9ad7d05a699b27e2b18e16758e463bc84d51bf4a983e73f022dde853fc8
3
+ size 268192975
Dataset004_CVC/nnUNetTrainer__nnUNetPlans__2d/fold_5/checkpoint_final.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9cdc9ad7d05a699b27e2b18e16758e463bc84d51bf4a983e73f022dde853fc8
3
+ size 268192975
Dataset004_CVC/nnUNetTrainer__nnUNetPlans__2d/plans.json ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_name": "Dataset004_CVC",
3
+ "plans_name": "nnUNetPlans",
4
+ "original_median_spacing_after_transp": [
5
+ 999.0,
6
+ 1.0,
7
+ 1.0
8
+ ],
9
+ "original_median_shape_after_transp": [
10
+ 1,
11
+ 288,
12
+ 384
13
+ ],
14
+ "image_reader_writer": "NaturalImage2DIO",
15
+ "transpose_forward": [
16
+ 0,
17
+ 1,
18
+ 2
19
+ ],
20
+ "transpose_backward": [
21
+ 0,
22
+ 1,
23
+ 2
24
+ ],
25
+ "configurations": {
26
+ "2d": {
27
+ "data_identifier": "nnUNetPlans_2d",
28
+ "preprocessor_name": "DefaultPreprocessor",
29
+ "batch_size": 26,
30
+ "patch_size": [
31
+ 320,
32
+ 384
33
+ ],
34
+ "median_image_size_in_voxels": [
35
+ 288.0,
36
+ 384.0
37
+ ],
38
+ "spacing": [
39
+ 1.0,
40
+ 1.0
41
+ ],
42
+ "normalization_schemes": [
43
+ "ZScoreNormalization",
44
+ "ZScoreNormalization",
45
+ "ZScoreNormalization"
46
+ ],
47
+ "use_mask_for_norm": [
48
+ false,
49
+ false,
50
+ false
51
+ ],
52
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
53
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
54
+ "resampling_fn_data_kwargs": {
55
+ "is_seg": false,
56
+ "order": 3,
57
+ "order_z": 0,
58
+ "force_separate_z": null
59
+ },
60
+ "resampling_fn_seg_kwargs": {
61
+ "is_seg": true,
62
+ "order": 1,
63
+ "order_z": 0,
64
+ "force_separate_z": null
65
+ },
66
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
67
+ "resampling_fn_probabilities_kwargs": {
68
+ "is_seg": false,
69
+ "order": 1,
70
+ "order_z": 0,
71
+ "force_separate_z": null
72
+ },
73
+ "architecture": {
74
+ "network_class_name": "dynamic_network_architectures.architectures.unet.PlainConvUNet",
75
+ "arch_kwargs": {
76
+ "n_stages": 7,
77
+ "features_per_stage": [
78
+ 32,
79
+ 64,
80
+ 128,
81
+ 256,
82
+ 512,
83
+ 512,
84
+ 512
85
+ ],
86
+ "conv_op": "torch.nn.modules.conv.Conv2d",
87
+ "kernel_sizes": [
88
+ [
89
+ 3,
90
+ 3
91
+ ],
92
+ [
93
+ 3,
94
+ 3
95
+ ],
96
+ [
97
+ 3,
98
+ 3
99
+ ],
100
+ [
101
+ 3,
102
+ 3
103
+ ],
104
+ [
105
+ 3,
106
+ 3
107
+ ],
108
+ [
109
+ 3,
110
+ 3
111
+ ],
112
+ [
113
+ 3,
114
+ 3
115
+ ]
116
+ ],
117
+ "strides": [
118
+ [
119
+ 1,
120
+ 1
121
+ ],
122
+ [
123
+ 2,
124
+ 2
125
+ ],
126
+ [
127
+ 2,
128
+ 2
129
+ ],
130
+ [
131
+ 2,
132
+ 2
133
+ ],
134
+ [
135
+ 2,
136
+ 2
137
+ ],
138
+ [
139
+ 2,
140
+ 2
141
+ ],
142
+ [
143
+ 2,
144
+ 2
145
+ ]
146
+ ],
147
+ "n_conv_per_stage": [
148
+ 2,
149
+ 2,
150
+ 2,
151
+ 2,
152
+ 2,
153
+ 2,
154
+ 2
155
+ ],
156
+ "n_conv_per_stage_decoder": [
157
+ 2,
158
+ 2,
159
+ 2,
160
+ 2,
161
+ 2,
162
+ 2
163
+ ],
164
+ "conv_bias": true,
165
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm2d",
166
+ "norm_op_kwargs": {
167
+ "eps": 1e-05,
168
+ "affine": true
169
+ },
170
+ "dropout_op": null,
171
+ "dropout_op_kwargs": null,
172
+ "nonlin": "torch.nn.LeakyReLU",
173
+ "nonlin_kwargs": {
174
+ "inplace": true
175
+ }
176
+ },
177
+ "_kw_requires_import": [
178
+ "conv_op",
179
+ "norm_op",
180
+ "dropout_op",
181
+ "nonlin"
182
+ ]
183
+ },
184
+ "batch_dice": true
185
+ }
186
+ },
187
+ "experiment_planner_used": "ExperimentPlanner",
188
+ "label_manager": "LabelManager",
189
+ "foreground_intensity_properties_per_channel": {
190
+ "0": {
191
+ "max": 255.0,
192
+ "mean": 178.0963592529297,
193
+ "median": 186.0,
194
+ "min": 0.0,
195
+ "percentile_00_5": 13.0,
196
+ "percentile_99_5": 255.0,
197
+ "std": 49.21238708496094
198
+ },
199
+ "1": {
200
+ "max": 255.0,
201
+ "mean": 104.5896987915039,
202
+ "median": 100.0,
203
+ "min": 0.0,
204
+ "percentile_00_5": 10.0,
205
+ "percentile_99_5": 242.0,
206
+ "std": 45.28557205200195
207
+ },
208
+ "2": {
209
+ "max": 255.0,
210
+ "mean": 78.76152038574219,
211
+ "median": 74.0,
212
+ "min": 0.0,
213
+ "percentile_00_5": 6.0,
214
+ "percentile_99_5": 231.0,
215
+ "std": 37.14423370361328
216
+ }
217
+ }
218
+ }