Upload 6 files
Browse files- cavia2024/args.yaml +273 -0
- cavia2024/model_best.pth.tar +3 -0
- cavia2024/summary.csv +53 -0
- corvi2023/args.yaml +266 -0
- corvi2023/model_best.pth.tar +3 -0
- corvi2023/summary.csv +17 -0
cavia2024/args.yaml
ADDED
|
@@ -0,0 +1,273 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
aa: null
|
| 2 |
+
amp: true
|
| 3 |
+
amp_dtype: float16
|
| 4 |
+
amp_impl: native
|
| 5 |
+
aug_repeats: 0
|
| 6 |
+
aug_splits: 0
|
| 7 |
+
batch_size: 8
|
| 8 |
+
bce_loss: true
|
| 9 |
+
bce_target_thresh: null
|
| 10 |
+
benchmark: false
|
| 11 |
+
blur_max: 2
|
| 12 |
+
blur_min: 0.1
|
| 13 |
+
blur_prob: 0.5
|
| 14 |
+
blur_sig:
|
| 15 |
+
- 0
|
| 16 |
+
- 3.0
|
| 17 |
+
bn_eps: null
|
| 18 |
+
bn_momentum: null
|
| 19 |
+
brightness_max: 2
|
| 20 |
+
brightness_min: 0.5
|
| 21 |
+
channels_last: false
|
| 22 |
+
checkpoint_hist: 10
|
| 23 |
+
cineca: true
|
| 24 |
+
class_map: ''
|
| 25 |
+
classifier:
|
| 26 |
+
- linear
|
| 27 |
+
- knn
|
| 28 |
+
- svm
|
| 29 |
+
clip_grad: null
|
| 30 |
+
clip_mode: norm
|
| 31 |
+
color_jitter: 0.4
|
| 32 |
+
contrast_max: 1.5
|
| 33 |
+
contrast_min: 0.5
|
| 34 |
+
contrastive_loss: false
|
| 35 |
+
cooldown_epochs: 0
|
| 36 |
+
corvi_resnet: false
|
| 37 |
+
crop_max: 512
|
| 38 |
+
crop_min: 64
|
| 39 |
+
crop_pct: null
|
| 40 |
+
cutmix: 0.0
|
| 41 |
+
cutmix_minmax: null
|
| 42 |
+
data: null
|
| 43 |
+
data_dir: /leonardo_scratch/large/userexternal/fcocchi0/deepfake/dataset/webdatasets_elsa_v2
|
| 44 |
+
data_dir_eval_augm: /leonardo_scratch/large/userexternal/fcocchi0/deepfake/dataset/Elsa_datasetv2_test_fix/wds_test_small/transf
|
| 45 |
+
data_dir_eval_no_augm: /leonardo_scratch/large/userexternal/fcocchi0/deepfake/dataset/Elsa_datasetv2_test_fix/wds_test_small/no_transf
|
| 46 |
+
data_generator: null
|
| 47 |
+
data_len_eval: 4800
|
| 48 |
+
data_len_linear: 9600
|
| 49 |
+
data_len_train: null
|
| 50 |
+
dataset: elsa_v2_ladeda
|
| 51 |
+
dataset_download: false
|
| 52 |
+
dataset_eval: elsa_v2_ladeda
|
| 53 |
+
decay_epochs: 90
|
| 54 |
+
decay_milestones:
|
| 55 |
+
- 90
|
| 56 |
+
- 180
|
| 57 |
+
- 270
|
| 58 |
+
decay_rate: 0.1
|
| 59 |
+
defake: false
|
| 60 |
+
deterministic: true
|
| 61 |
+
dino_crop: false
|
| 62 |
+
dino_head: false
|
| 63 |
+
dino_loss: false
|
| 64 |
+
dino_loss_weight: 0.5
|
| 65 |
+
dino_temp: 0.1
|
| 66 |
+
dist_bn: reduce
|
| 67 |
+
distance: cosine
|
| 68 |
+
double_contrastive: false
|
| 69 |
+
drop: 0.0
|
| 70 |
+
drop_block: null
|
| 71 |
+
drop_connect: null
|
| 72 |
+
drop_path: null
|
| 73 |
+
early_stopping: false
|
| 74 |
+
epoch_repeats: 0.0
|
| 75 |
+
epochs: 1000
|
| 76 |
+
epochs_classifier: 1500
|
| 77 |
+
eval_metric: accuracy
|
| 78 |
+
experiment: ladeda_d3_d3_augm-augm_prob_0.5-w8_transform_lr-0.0002-batch_size-32_epochs-1000_loss-bce_optimizer-adam_fcocchi
|
| 79 |
+
external_transform: false
|
| 80 |
+
fast_norm: false
|
| 81 |
+
fuser: ''
|
| 82 |
+
global_crops_scale:
|
| 83 |
+
- 0.4
|
| 84 |
+
- 1.0
|
| 85 |
+
gp: null
|
| 86 |
+
grad_accum_steps: 1
|
| 87 |
+
grad_checkpointing: false
|
| 88 |
+
head_bottlenck: 256
|
| 89 |
+
head_hidden_dim: 384
|
| 90 |
+
head_init_bias: null
|
| 91 |
+
head_init_scale: null
|
| 92 |
+
head_out: 192
|
| 93 |
+
hflip: 0.5
|
| 94 |
+
img_size: null
|
| 95 |
+
in_chans: null
|
| 96 |
+
infonce_loss_temperature: 0.1
|
| 97 |
+
initial_checkpoint: ''
|
| 98 |
+
input_size:
|
| 99 |
+
- 3
|
| 100 |
+
- 224
|
| 101 |
+
- 224
|
| 102 |
+
interpolation: ''
|
| 103 |
+
jitter_max: 1.5
|
| 104 |
+
jitter_min: 0.5
|
| 105 |
+
job_id: 0
|
| 106 |
+
jpeg_max: 100
|
| 107 |
+
jpeg_min: 30
|
| 108 |
+
jpeg_prob: 0.5
|
| 109 |
+
jsd_loss: false
|
| 110 |
+
ladeda: true
|
| 111 |
+
ladeda_augm: true
|
| 112 |
+
lambda_loss: 10
|
| 113 |
+
last_crop: false
|
| 114 |
+
layer_decay: null
|
| 115 |
+
linear_pretrained: null
|
| 116 |
+
linear_train_shards: dataset/shards/elsa_v2_train_transf.shards
|
| 117 |
+
load_code: false
|
| 118 |
+
local_crops_scale:
|
| 119 |
+
- 0.05
|
| 120 |
+
- 0.4
|
| 121 |
+
local_rank: 0
|
| 122 |
+
log_interval: 5
|
| 123 |
+
log_wandb: true
|
| 124 |
+
lr: 0.0002
|
| 125 |
+
lr_base: 0.1
|
| 126 |
+
lr_base_scale: ''
|
| 127 |
+
lr_base_size: 256
|
| 128 |
+
lr_cycle_decay: 0.5
|
| 129 |
+
lr_cycle_limit: 1
|
| 130 |
+
lr_cycle_mul: 1.0
|
| 131 |
+
lr_k_decay: 1.0
|
| 132 |
+
lr_noise: null
|
| 133 |
+
lr_noise_pct: 0.67
|
| 134 |
+
lr_noise_std: 1.0
|
| 135 |
+
margin: 1
|
| 136 |
+
mean:
|
| 137 |
+
- 0.485
|
| 138 |
+
- 0.456
|
| 139 |
+
- 0.406
|
| 140 |
+
min_delta: -0.001
|
| 141 |
+
min_lr: 0.0
|
| 142 |
+
mixup: 0.0
|
| 143 |
+
mixup_mode: batch
|
| 144 |
+
mixup_off_epoch: 0
|
| 145 |
+
mixup_prob: 1.0
|
| 146 |
+
mixup_switch_prob: 0.5
|
| 147 |
+
model: ladeda_d3
|
| 148 |
+
model_ema: false
|
| 149 |
+
model_ema_decay: 0.9998
|
| 150 |
+
model_ema_force_cpu: false
|
| 151 |
+
model_kwargs: {}
|
| 152 |
+
momentum: 0.9
|
| 153 |
+
multiple_evaluations: true
|
| 154 |
+
n_crops: 1
|
| 155 |
+
no_aug: false
|
| 156 |
+
no_ddp_bb: false
|
| 157 |
+
no_prefetcher: true
|
| 158 |
+
no_resume_opt: false
|
| 159 |
+
not_load_input_size: false
|
| 160 |
+
num_classes: null
|
| 161 |
+
num_jobs: 1
|
| 162 |
+
num_step: 6800
|
| 163 |
+
num_transform: 2
|
| 164 |
+
ojha_d3: false
|
| 165 |
+
only_linear_training: false
|
| 166 |
+
only_validate: false
|
| 167 |
+
opacity_max: 1.0
|
| 168 |
+
opacity_min: 0.2
|
| 169 |
+
opt: adam
|
| 170 |
+
opt_betas: null
|
| 171 |
+
opt_eps: null
|
| 172 |
+
opt_kwargs: {}
|
| 173 |
+
output: /leonardo_scratch/large/userexternal/fcocchi0/deepfake/runs
|
| 174 |
+
overlay_max: 0.35
|
| 175 |
+
overlay_min: 0.05
|
| 176 |
+
pad_max: 0.25
|
| 177 |
+
pad_min: 0.01
|
| 178 |
+
patience_counter: 0
|
| 179 |
+
patience_epochs: 15
|
| 180 |
+
permutation_fake: false
|
| 181 |
+
permutation_real: false
|
| 182 |
+
pin_mem: false
|
| 183 |
+
pix_max: 1
|
| 184 |
+
pix_min: 0.3
|
| 185 |
+
plot_freq: 5
|
| 186 |
+
pretrained: true
|
| 187 |
+
random_crop: false
|
| 188 |
+
random_grayscale_prob: 0.01
|
| 189 |
+
ratio:
|
| 190 |
+
- 0.75
|
| 191 |
+
- 1.3333333333333333
|
| 192 |
+
ratio_max: 2
|
| 193 |
+
ratio_min: 0.75
|
| 194 |
+
real_centering_loss: false
|
| 195 |
+
recount: 1
|
| 196 |
+
recovery_interval: 0
|
| 197 |
+
remode: pixel
|
| 198 |
+
reprob: 0.0
|
| 199 |
+
resize_max: 512
|
| 200 |
+
resize_min: 64
|
| 201 |
+
resplit: false
|
| 202 |
+
resume: ''
|
| 203 |
+
rotatio_max: 270
|
| 204 |
+
rotatio_min: 90
|
| 205 |
+
saturation_max: 1.5
|
| 206 |
+
saturation_min: 0.5
|
| 207 |
+
save_images: false
|
| 208 |
+
save_model_linear: false
|
| 209 |
+
scale:
|
| 210 |
+
- 0.08
|
| 211 |
+
- 1.0
|
| 212 |
+
scale_max: 1.5
|
| 213 |
+
scale_min: 0.5
|
| 214 |
+
sched: plateau
|
| 215 |
+
sched_on_updates: false
|
| 216 |
+
seed: 42
|
| 217 |
+
sharp_max: 2.0
|
| 218 |
+
sharp_min: 1.2
|
| 219 |
+
shuffle_max: 0.35
|
| 220 |
+
shuffle_min: 0.0
|
| 221 |
+
skew_max: 1.0
|
| 222 |
+
skew_min: -1.0
|
| 223 |
+
smoothing: 0.0
|
| 224 |
+
split_bn: false
|
| 225 |
+
start_epoch: null
|
| 226 |
+
std:
|
| 227 |
+
- 0.229
|
| 228 |
+
- 0.224
|
| 229 |
+
- 0.225
|
| 230 |
+
step: 5
|
| 231 |
+
sup_contrastive_loss: false
|
| 232 |
+
suppl: false
|
| 233 |
+
sync_bn: true
|
| 234 |
+
synchronize_step: false
|
| 235 |
+
teacher_temp: 0.07
|
| 236 |
+
teacher_temp_fix: false
|
| 237 |
+
test_augm: false
|
| 238 |
+
test_shards_augm: dataset/shards/coco-test-dict.shards
|
| 239 |
+
test_shards_no_augm: dataset/shards/coco-test-dict.shards
|
| 240 |
+
threshold_plateau: 0.001
|
| 241 |
+
torchcompile: null
|
| 242 |
+
torchscript: false
|
| 243 |
+
train_interpolation: random
|
| 244 |
+
train_shards: dataset/shards/elsav2-training.shards
|
| 245 |
+
train_split: train
|
| 246 |
+
triplet_loss: false
|
| 247 |
+
tta: 0
|
| 248 |
+
use_multi_epochs_loader: false
|
| 249 |
+
val: false
|
| 250 |
+
val_shards_augm: dataset/shards/validation_set-transf-elsav2.shards
|
| 251 |
+
val_shards_no_augm: dataset/shards/validation_set-no_transf-elsav2.shards
|
| 252 |
+
val_split: validation
|
| 253 |
+
validation_batch_size: 50
|
| 254 |
+
vflip: 0.0
|
| 255 |
+
wandb_entity: lorenzo_b_master_thesis
|
| 256 |
+
wandb_group: null
|
| 257 |
+
wandb_id: null
|
| 258 |
+
wandb_logging: false
|
| 259 |
+
wandb_name: null
|
| 260 |
+
wandb_notes: null
|
| 261 |
+
wandb_project_name: contrastive-fake
|
| 262 |
+
wandb_resume: allow
|
| 263 |
+
warmup_epochs: 0
|
| 264 |
+
warmup_lr: 1.0e-06
|
| 265 |
+
warmup_prefix: false
|
| 266 |
+
warmup_teacher_temp: 0.04
|
| 267 |
+
warmup_teacher_temp_epochs: 30
|
| 268 |
+
watermark_prob: 0.2
|
| 269 |
+
weight_decay: 2.0e-05
|
| 270 |
+
weight_decay_end: null
|
| 271 |
+
worker_seeding: all
|
| 272 |
+
workers: 8
|
| 273 |
+
workers_validate: 3
|
cavia2024/model_best.pth.tar
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e30979838c8dd9a3e1a62411237c222ee9c8698230eafed0f26b70463919224
|
| 3 |
+
size 164164132
|
cavia2024/summary.csv
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,train_loss_contrastive_1,train_loss_contrastive_2,train_loss_dino,train_loss_mse,eval_loss,eval_loss_contrastive_1,eval_loss_contrastive_2,eval_loss_dino,eval_top1_tot,eval_accuracy,lr
|
| 2 |
+
0,0.4715742940521416,0.4715742940521416,0.0,0.0,0.0,1.0348793665568035,1.0348793665568035,0.0,0.0,0.5662499939401945,0.5662499939401945,0.0002
|
| 3 |
+
1,0.41957348761532237,0.41957348761532237,0.0,0.0,0.0,0.4376497281094392,0.4376497281094392,0.0,0.0,0.7814583306511244,0.7814583306511244,0.0002
|
| 4 |
+
2,0.4039742007084629,0.4039742007084629,0.0,0.0,0.0,0.43792665128906566,0.43792665128906566,0.0,0.0,0.7924999867876371,0.7924999867876371,0.0002
|
| 5 |
+
3,0.38984661175705054,0.38984661175705054,0.0,0.0,0.0,0.4219825491309166,0.4219825491309166,0.0,0.0,0.7964583237965902,0.7964583237965902,0.0002
|
| 6 |
+
4,0.38123743336866883,0.38123743336866883,0.0,0.0,0.0,0.5754485875368118,0.5754485875368118,0.0,0.0,0.7589583347241083,0.7589583347241083,0.0002
|
| 7 |
+
5,0.3721761410836788,0.3721761410836788,0.0,0.0,0.0,0.47601031387845677,0.47601031387845677,0.0,0.0,0.784166673819224,0.784166673819224,0.0002
|
| 8 |
+
6,0.36888879357891924,0.36888879357891924,0.0,0.0,0.0,0.5046001076698303,0.5046001076698303,0.0,0.0,0.7785416593154272,0.7785416593154272,0.0002
|
| 9 |
+
7,0.36127805593697465,0.36127805593697465,0.0,0.0,0.0,0.49871913840373355,0.49871913840373355,0.0,0.0,0.7799999987085661,0.7799999987085661,0.0002
|
| 10 |
+
8,0.355624608531156,0.355624608531156,0.0,0.0,0.0,0.8060868283112844,0.8060868283112844,0.0,0.0,0.7156250054637591,0.7156250054637591,0.0002
|
| 11 |
+
9,0.35819693353465376,0.35819693353465376,0.0,0.0,0.0,0.5123565718531609,0.5123565718531609,0.0,0.0,0.7872916584213575,0.7872916584213575,0.0002
|
| 12 |
+
10,0.35487239067716636,0.35487239067716636,0.0,0.0,0.0,0.6208437109986941,0.6208437109986941,0.0,0.0,0.7504166637857755,0.7504166637857755,0.0002
|
| 13 |
+
11,0.34664950236568554,0.34664950236568554,0.0,0.0,0.0,0.6312317748864492,0.6312317748864492,0.0,0.0,0.7618749961256981,0.7618749961256981,0.0002
|
| 14 |
+
12,0.35307861270163865,0.35307861270163865,0.0,0.0,0.0,0.49351882189512253,0.49351882189512253,0.0,0.0,0.7891666640837988,0.7891666640837988,0.0002
|
| 15 |
+
13,0.34500516020857236,0.34500516020857236,0.0,0.0,0.0,0.5925510463615259,0.5925510463615259,0.0,0.0,0.7616666580239931,0.7616666580239931,0.0002
|
| 16 |
+
14,0.338449774967397,0.338449774967397,0.0,0.0,0.0,0.6316928118467331,0.6316928118467331,0.0,0.0,0.762916661798954,0.762916661798954,0.0002
|
| 17 |
+
15,0.3389449470621698,0.3389449470621698,0.0,0.0,0.0,0.5894950131575266,0.5894950131575266,0.0,0.0,0.7647916600108147,0.7647916600108147,0.0002
|
| 18 |
+
16,0.3380515977247235,0.3380515977247235,0.0,0.0,0.0,0.5934893501301607,0.5934893501301607,0.0,0.0,0.7770833323399226,0.7770833323399226,0.0002
|
| 19 |
+
17,0.3352091408937293,0.3352091408937293,0.0,0.0,0.0,0.9127591873208681,0.9127591873208681,0.0,0.0,0.7164583330353101,0.7164583330353101,0.0002
|
| 20 |
+
18,0.3356964987319182,0.3356964987319182,0.0,0.0,0.0,0.6280725635588169,0.6280725635588169,0.0,0.0,0.7677083338300387,0.7677083338300387,0.0002
|
| 21 |
+
19,0.3330101035973605,0.3330101035973605,0.0,0.0,0.0,0.7506165206432343,0.7506165206432343,0.0,0.0,0.7462499961256981,0.7462499961256981,0.0002
|
| 22 |
+
20,0.31619833306245065,0.31619833306245065,0.0,0.0,0.0,0.7563606550296148,0.7563606550296148,0.0,0.0,0.7464583367109299,0.7464583367109299,2e-05
|
| 23 |
+
21,0.3006130123675308,0.3006130123675308,0.0,0.0,0.0,0.742191178103288,0.742191178103288,0.0,0.0,0.7537500038743019,0.7537500038743019,2e-05
|
| 24 |
+
22,0.30464976418434697,0.30464976418434697,0.0,0.0,0.0,0.709917259713014,0.709917259713014,0.0,0.0,0.7639583349227905,0.7639583349227905,2e-05
|
| 25 |
+
23,0.30154833327431013,0.30154833327431013,0.0,0.0,0.0,0.7742489576339722,0.7742489576339722,0.0,0.0,0.7602083384990692,0.7602083384990692,2e-05
|
| 26 |
+
24,0.30147681120673525,0.30147681120673525,0.0,0.0,0.0,0.8164870118101438,0.8164870118101438,0.0,0.0,0.7529166763027509,0.7529166763027509,2e-05
|
| 27 |
+
25,0.30200591095011026,0.30200591095011026,0.0,0.0,0.0,0.8715417385101318,0.8715417385101318,0.0,0.0,0.7402083327372869,0.7402083327372869,2e-05
|
| 28 |
+
26,0.2986566265606705,0.2986566265606705,0.0,0.0,0.0,0.7886857589085897,0.7886857589085897,0.0,0.0,0.7508333449562391,0.7508333449562391,2e-05
|
| 29 |
+
27,0.29696609591517376,0.29696609591517376,0.0,0.0,0.0,0.780015729367733,0.780015729367733,0.0,0.0,0.757083331545194,0.757083331545194,2e-05
|
| 30 |
+
28,0.2984225282116848,0.2984225282116848,0.0,0.0,0.0,0.8414804016550382,0.8414804016550382,0.0,0.0,0.7495833337306976,0.7495833337306976,2e-05
|
| 31 |
+
29,0.29967677641440843,0.29967677641440843,0.0,0.0,0.0,0.7911093135674795,0.7911093135674795,0.0,0.0,0.7554166615009308,0.7554166615009308,2e-05
|
| 32 |
+
30,0.295477515135837,0.295477515135837,0.0,0.0,0.0,0.7634668747584025,0.7634668747584025,0.0,0.0,0.7574999928474426,0.7574999928474426,2e-05
|
| 33 |
+
31,0.2921763726290973,0.2921763726290973,0.0,0.0,0.0,0.7987801333268484,0.7987801333268484,0.0,0.0,0.7520833387970924,0.7520833387970924,2e-05
|
| 34 |
+
32,0.2914208925274365,0.2914208925274365,0.0,0.0,0.0,0.8135101199150085,0.8135101199150085,0.0,0.0,0.7518750031789144,0.7518750031789144,2e-05
|
| 35 |
+
33,0.29439522144110764,0.29439522144110764,0.0,0.0,0.0,0.8285359119375547,0.8285359119375547,0.0,0.0,0.7562499990065893,0.7562499990065893,2e-05
|
| 36 |
+
34,0.29300388195278015,0.29300388195278015,0.0,0.0,0.0,0.8260350326697031,0.8260350326697031,0.0,0.0,0.7547916769981384,0.7547916769981384,2e-05
|
| 37 |
+
35,0.2883887961369884,0.2883887961369884,0.0,0.0,0.0,0.808680015305678,0.808680015305678,0.0,0.0,0.7566666627923647,0.7566666627923647,2e-05
|
| 38 |
+
36,0.2957167236791814,0.2957167236791814,0.0,0.0,0.0,0.8322958101828893,0.8322958101828893,0.0,0.0,0.7549999927481016,0.7549999927481016,2.0000000000000003e-06
|
| 39 |
+
37,0.2907253118670162,0.2907253118670162,0.0,0.0,0.0,0.8748250255982081,0.8748250255982081,0.0,0.0,0.7410416727264723,0.7410416727264723,2.0000000000000003e-06
|
| 40 |
+
38,0.2867253391291289,0.2867253391291289,0.0,0.0,0.0,0.8294427916407585,0.8294427916407585,0.0,0.0,0.7543749958276749,0.7543749958276749,2.0000000000000003e-06
|
| 41 |
+
39,0.28846490932080676,0.28846490932080676,0.0,0.0,0.0,0.831823588659366,0.831823588659366,0.0,0.0,0.75583333025376,0.75583333025376,2.0000000000000003e-06
|
| 42 |
+
40,0.28739439499400116,0.28739439499400116,0.0,0.0,0.0,0.8560531189044317,0.8560531189044317,0.0,0.0,0.7531250069538752,0.7531250069538752,2.0000000000000003e-06
|
| 43 |
+
41,0.2903684458147515,0.2903684458147515,0.0,0.0,0.0,0.8575630386670431,0.8575630386670431,0.0,0.0,0.7531250044703484,0.7531250044703484,2.0000000000000003e-06
|
| 44 |
+
42,0.2837417893604759,0.2837417893604759,0.0,0.0,0.0,0.8910363912582397,0.8910363912582397,0.0,0.0,0.7466666748126348,0.7466666748126348,2.0000000000000003e-06
|
| 45 |
+
43,0.2835005822701051,0.2835005822701051,0.0,0.0,0.0,0.8428035080432892,0.8428035080432892,0.0,0.0,0.7545833364129066,0.7545833364129066,2.0000000000000003e-06
|
| 46 |
+
44,0.2873177314779776,0.2873177314779776,0.0,0.0,0.0,0.8845499530434608,0.8845499530434608,0.0,0.0,0.7566666702429453,0.7566666702429453,2.0000000000000003e-06
|
| 47 |
+
45,0.2848608064574792,0.2848608064574792,0.0,0.0,0.0,0.8590305397907892,0.8590305397907892,0.0,0.0,0.7558333352208138,0.7558333352208138,2.0000000000000003e-06
|
| 48 |
+
46,0.28449650667059945,0.28449650667059945,0.0,0.0,0.0,0.9118799045681953,0.9118799045681953,0.0,0.0,0.7425000121196111,0.7425000121196111,2.0000000000000003e-06
|
| 49 |
+
47,0.2949172469622949,0.2949172469622949,0.0,0.0,0.0,0.8296625837683678,0.8296625837683678,0.0,0.0,0.7602083260814348,0.7602083260814348,2.0000000000000003e-06
|
| 50 |
+
48,0.28397094216197727,0.28397094216197727,0.0,0.0,0.0,0.8702390491962433,0.8702390491962433,0.0,0.0,0.7516666675607363,0.7516666675607363,2.0000000000000003e-06
|
| 51 |
+
49,0.28254622190862017,0.28254622190862017,0.0,0.0,0.0,0.8600285823146502,0.8600285823146502,0.0,0.0,0.753541665772597,0.753541665772597,2.0000000000000003e-06
|
| 52 |
+
50,0.28421684513964196,0.28421684513964196,0.0,0.0,0.0,0.9056417221824328,0.9056417221824328,0.0,0.0,0.7431249916553497,0.7431249916553497,2.0000000000000003e-06
|
| 53 |
+
51,0.2835257639396278,0.2835257639396278,0.0,0.0,0.0,0.8019334400693575,0.8019334400693575,0.0,0.0,0.7664583300550779,0.7664583300550779,2.0000000000000003e-06
|
corvi2023/args.yaml
ADDED
|
@@ -0,0 +1,266 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
aa: null
|
| 2 |
+
amp: true
|
| 3 |
+
amp_dtype: float16
|
| 4 |
+
amp_impl: native
|
| 5 |
+
aug_repeats: 0
|
| 6 |
+
aug_splits: 0
|
| 7 |
+
batch_size: 16
|
| 8 |
+
bce_loss: true
|
| 9 |
+
bce_target_thresh: null
|
| 10 |
+
benchmark: false
|
| 11 |
+
blur_max: 2
|
| 12 |
+
blur_min: 0.1
|
| 13 |
+
blur_prob: 0.5
|
| 14 |
+
blur_sig:
|
| 15 |
+
- 0
|
| 16 |
+
- 3.0
|
| 17 |
+
bn_eps: null
|
| 18 |
+
bn_momentum: null
|
| 19 |
+
brightness_max: 2
|
| 20 |
+
brightness_min: 0.5
|
| 21 |
+
channels_last: false
|
| 22 |
+
checkpoint_hist: 10
|
| 23 |
+
cineca: true
|
| 24 |
+
class_map: ''
|
| 25 |
+
classifier:
|
| 26 |
+
- linear
|
| 27 |
+
- knn
|
| 28 |
+
- svm
|
| 29 |
+
clip_grad: null
|
| 30 |
+
clip_mode: norm
|
| 31 |
+
color_jitter: 0.4
|
| 32 |
+
contrast_max: 1.5
|
| 33 |
+
contrast_min: 0.5
|
| 34 |
+
contrastive_loss: false
|
| 35 |
+
cooldown_epochs: 0
|
| 36 |
+
corvi_resnet: true
|
| 37 |
+
crop_max: 512
|
| 38 |
+
crop_min: 64
|
| 39 |
+
crop_pct: null
|
| 40 |
+
cutmix: 0.0
|
| 41 |
+
cutmix_minmax: null
|
| 42 |
+
data: null
|
| 43 |
+
data_dir: /leonardo_scratch/large/userexternal/lbaraldi/webdatasets_elsa_v2
|
| 44 |
+
data_dir_eval_augm: /leonardo_scratch/large/userexternal/fcocchi0/deepfake/test_wds_fix/wds_test_small/transf
|
| 45 |
+
data_dir_eval_no_augm: /leonardo_scratch/large/userexternal/fcocchi0/deepfake/test_wds_fix/wds_test_small/no_transf
|
| 46 |
+
data_generator: null
|
| 47 |
+
data_len_eval: 4800
|
| 48 |
+
data_len_linear: 9600
|
| 49 |
+
data_len_train: null
|
| 50 |
+
dataset: elsa_v2_binarycrossentropy
|
| 51 |
+
dataset_download: false
|
| 52 |
+
dataset_eval: elsa_v2_binarycrossentropy_all_gen
|
| 53 |
+
decay_epochs: 90
|
| 54 |
+
decay_milestones:
|
| 55 |
+
- 90
|
| 56 |
+
- 180
|
| 57 |
+
- 270
|
| 58 |
+
decay_rate: 0.1
|
| 59 |
+
deterministic: true
|
| 60 |
+
dino_crop: false
|
| 61 |
+
dino_head: false
|
| 62 |
+
dino_loss: false
|
| 63 |
+
dino_loss_weight: 0.5
|
| 64 |
+
dino_temp: 0.1
|
| 65 |
+
dist_bn: reduce
|
| 66 |
+
distance: cosine
|
| 67 |
+
double_contrastive: false
|
| 68 |
+
drop: 0.0
|
| 69 |
+
drop_block: null
|
| 70 |
+
drop_connect: null
|
| 71 |
+
drop_path: null
|
| 72 |
+
early_stopping: false
|
| 73 |
+
epoch_repeats: 0.0
|
| 74 |
+
epochs: 150
|
| 75 |
+
epochs_classifier: 1500
|
| 76 |
+
eval_metric: accuracy
|
| 77 |
+
experiment: eccv_rebuttal_corvi_d3-augm_prob_0.5-resnet50.tv_in1k_w8_transform_lr-1e-4-batch_size-64_epochs-150_loss-bce_optimizer-adam_fcocchi
|
| 78 |
+
external_transform: true
|
| 79 |
+
fast_norm: false
|
| 80 |
+
fuser: ''
|
| 81 |
+
global_crops_scale:
|
| 82 |
+
- 0.4
|
| 83 |
+
- 1.0
|
| 84 |
+
gp: null
|
| 85 |
+
grad_accum_steps: 1
|
| 86 |
+
grad_checkpointing: false
|
| 87 |
+
head_bottlenck: 256
|
| 88 |
+
head_hidden_dim: 384
|
| 89 |
+
head_init_bias: null
|
| 90 |
+
head_init_scale: null
|
| 91 |
+
head_out: 192
|
| 92 |
+
hflip: 0.5
|
| 93 |
+
img_size: null
|
| 94 |
+
in_chans: null
|
| 95 |
+
infonce_loss_temperature: 0.1
|
| 96 |
+
initial_checkpoint: ''
|
| 97 |
+
input_size:
|
| 98 |
+
- 3
|
| 99 |
+
- 224
|
| 100 |
+
- 224
|
| 101 |
+
interpolation: ''
|
| 102 |
+
jitter_max: 1.5
|
| 103 |
+
jitter_min: 0.5
|
| 104 |
+
job_id: 0
|
| 105 |
+
jpeg_max: 100
|
| 106 |
+
jpeg_min: 30
|
| 107 |
+
jpeg_prob: 0.5
|
| 108 |
+
jsd_loss: false
|
| 109 |
+
lambda_loss: 10
|
| 110 |
+
last_crop: false
|
| 111 |
+
layer_decay: null
|
| 112 |
+
linear_pretrained: null
|
| 113 |
+
linear_train_shards: dataset/shards/elsa_v2_train_transf.shards
|
| 114 |
+
local_crops_scale:
|
| 115 |
+
- 0.05
|
| 116 |
+
- 0.4
|
| 117 |
+
local_rank: 0
|
| 118 |
+
log_interval: 5
|
| 119 |
+
log_wandb: true
|
| 120 |
+
lr: 0.0001
|
| 121 |
+
lr_base: 0.1
|
| 122 |
+
lr_base_scale: ''
|
| 123 |
+
lr_base_size: 256
|
| 124 |
+
lr_cycle_decay: 0.5
|
| 125 |
+
lr_cycle_limit: 1
|
| 126 |
+
lr_cycle_mul: 1.0
|
| 127 |
+
lr_k_decay: 1.0
|
| 128 |
+
lr_noise: null
|
| 129 |
+
lr_noise_pct: 0.67
|
| 130 |
+
lr_noise_std: 1.0
|
| 131 |
+
margin: 1
|
| 132 |
+
mean:
|
| 133 |
+
- 0.485
|
| 134 |
+
- 0.456
|
| 135 |
+
- 0.406
|
| 136 |
+
min_lr: 0.0
|
| 137 |
+
mixup: 0.0
|
| 138 |
+
mixup_mode: batch
|
| 139 |
+
mixup_off_epoch: 0
|
| 140 |
+
mixup_prob: 1.0
|
| 141 |
+
mixup_switch_prob: 0.5
|
| 142 |
+
model: resnet50.tv_in1k
|
| 143 |
+
model_ema: false
|
| 144 |
+
model_ema_decay: 0.9998
|
| 145 |
+
model_ema_force_cpu: false
|
| 146 |
+
model_kwargs: {}
|
| 147 |
+
momentum: 0.9
|
| 148 |
+
multiple_evaluations: true
|
| 149 |
+
n_crops: 1
|
| 150 |
+
no_aug: false
|
| 151 |
+
no_ddp_bb: false
|
| 152 |
+
no_prefetcher: true
|
| 153 |
+
no_resume_opt: false
|
| 154 |
+
not_load_input_size: false
|
| 155 |
+
num_classes: null
|
| 156 |
+
num_jobs: 1
|
| 157 |
+
num_step: 27200
|
| 158 |
+
num_transform: 2
|
| 159 |
+
only_validate: false
|
| 160 |
+
opacity_max: 1.0
|
| 161 |
+
opacity_min: 0.2
|
| 162 |
+
opt: adam
|
| 163 |
+
opt_betas: null
|
| 164 |
+
opt_eps: null
|
| 165 |
+
opt_kwargs: {}
|
| 166 |
+
output: /leonardo_scratch/large/userexternal/fcocchi0/deepfake/runs
|
| 167 |
+
overlay_max: 0.35
|
| 168 |
+
overlay_min: 0.05
|
| 169 |
+
pad_max: 0.25
|
| 170 |
+
pad_min: 0.01
|
| 171 |
+
patience_counter: 0
|
| 172 |
+
patience_epochs: 6
|
| 173 |
+
permutation_fake: false
|
| 174 |
+
permutation_real: false
|
| 175 |
+
pin_mem: false
|
| 176 |
+
pix_max: 1
|
| 177 |
+
pix_min: 0.3
|
| 178 |
+
plot_freq: 5
|
| 179 |
+
pretrained: true
|
| 180 |
+
random_crop: false
|
| 181 |
+
random_grayscale_prob: 0.01
|
| 182 |
+
ratio:
|
| 183 |
+
- 0.75
|
| 184 |
+
- 1.3333333333333333
|
| 185 |
+
ratio_max: 2
|
| 186 |
+
ratio_min: 0.75
|
| 187 |
+
real_centering_loss: false
|
| 188 |
+
recount: 1
|
| 189 |
+
recovery_interval: 0
|
| 190 |
+
remode: pixel
|
| 191 |
+
reprob: 0.0
|
| 192 |
+
resize_max: 512
|
| 193 |
+
resize_min: 64
|
| 194 |
+
resplit: false
|
| 195 |
+
resume: ''
|
| 196 |
+
rotatio_max: 270
|
| 197 |
+
rotatio_min: 90
|
| 198 |
+
saturation_max: 1.5
|
| 199 |
+
saturation_min: 0.5
|
| 200 |
+
save_images: false
|
| 201 |
+
save_model_linear: false
|
| 202 |
+
scale:
|
| 203 |
+
- 0.08
|
| 204 |
+
- 1.0
|
| 205 |
+
scale_max: 1.5
|
| 206 |
+
scale_min: 0.5
|
| 207 |
+
sched: plateau
|
| 208 |
+
sched_on_updates: false
|
| 209 |
+
seed: 42
|
| 210 |
+
sharp_max: 2.0
|
| 211 |
+
sharp_min: 1.2
|
| 212 |
+
shuffle_max: 0.35
|
| 213 |
+
shuffle_min: 0.0
|
| 214 |
+
skew_max: 1.0
|
| 215 |
+
skew_min: -1.0
|
| 216 |
+
smoothing: 0.0
|
| 217 |
+
split_bn: false
|
| 218 |
+
start_epoch: null
|
| 219 |
+
std:
|
| 220 |
+
- 0.229
|
| 221 |
+
- 0.224
|
| 222 |
+
- 0.225
|
| 223 |
+
step: 5
|
| 224 |
+
sup_contrastive_loss: false
|
| 225 |
+
suppl: false
|
| 226 |
+
sync_bn: true
|
| 227 |
+
synchronize_step: false
|
| 228 |
+
teacher_temp: 0.07
|
| 229 |
+
teacher_temp_fix: false
|
| 230 |
+
test_augm: false
|
| 231 |
+
test_shards_augm: dataset/shards/coco-test-dict.shards
|
| 232 |
+
test_shards_no_augm: dataset/shards/coco-test-dict.shards
|
| 233 |
+
threshold_plateau: 0.001
|
| 234 |
+
torchcompile: null
|
| 235 |
+
torchscript: false
|
| 236 |
+
train_interpolation: random
|
| 237 |
+
train_shards: dataset/shards/elsav2-training.shards
|
| 238 |
+
train_split: train
|
| 239 |
+
triplet_loss: false
|
| 240 |
+
tta: 0
|
| 241 |
+
use_multi_epochs_loader: false
|
| 242 |
+
val: false
|
| 243 |
+
val_shards_augm: dataset/shards/validation_set-transf-elsav2.shards
|
| 244 |
+
val_shards_no_augm: dataset/shards/validation_set-no_transf-elsav2.shards
|
| 245 |
+
val_split: validation
|
| 246 |
+
validation_batch_size: 50
|
| 247 |
+
vflip: 0.0
|
| 248 |
+
wandb_entity: lorenzo_b_master_thesis
|
| 249 |
+
wandb_group: null
|
| 250 |
+
wandb_id: null
|
| 251 |
+
wandb_logging: false
|
| 252 |
+
wandb_name: null
|
| 253 |
+
wandb_notes: null
|
| 254 |
+
wandb_project_name: contrastive-fake
|
| 255 |
+
wandb_resume: allow
|
| 256 |
+
warmup_epochs: 0
|
| 257 |
+
warmup_lr: 1.0e-06
|
| 258 |
+
warmup_prefix: false
|
| 259 |
+
warmup_teacher_temp: 0.04
|
| 260 |
+
warmup_teacher_temp_epochs: 30
|
| 261 |
+
watermark_prob: 0.2
|
| 262 |
+
weight_decay: 2.0e-05
|
| 263 |
+
weight_decay_end: null
|
| 264 |
+
worker_seeding: all
|
| 265 |
+
workers: 8
|
| 266 |
+
workers_validate: 3
|
corvi2023/model_best.pth.tar
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:983714bf0597edc3aa716df41b345b3b87e77c46fbb7626b22a80ac9bae8aa8b
|
| 3 |
+
size 282584748
|
corvi2023/summary.csv
ADDED
|
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,train_loss_contrastive_1,train_loss_contrastive_2,train_loss_dino,train_loss_mse,eval_loss,eval_loss_contrastive_1,eval_loss_contrastive_2,eval_loss_dino,eval_top1_tot,eval_top1_real,eval_top1_f0,eval_top1_f1,eval_top1_f2,eval_top1_f3,eval_accuracy,lr
|
| 2 |
+
0,0.1425799236715059,0.1425799236715059,0.0,0.0,0.0,0.07369088536749284,0.07369088536749284,0.0,0.0,0.9738333423932394,0.9662500272194544,0.9400000125169754,0.9927083402872086,0.980625015993913,0.9895833432674408,0.9738333423932394,0.0001
|
| 3 |
+
1,0.10152659070291145,0.10152659070291145,0.0,0.0,0.0,0.07265529626359542,0.07265529626359542,0.0,0.0,0.9741249854365984,0.9550000205636024,0.9537500267227491,0.9931250065565109,0.9802083497246107,0.9885416775941849,0.9741249854365984,0.0001
|
| 4 |
+
2,0.08957478656906032,0.08957478656906032,0.0,0.0,0.0,0.10243250305453937,0.10243250305453937,0.0,0.0,0.966499999165535,0.9766666864355406,0.9379166662693024,0.9747916907072067,0.9635416865348816,0.979583352804184,0.966499999165535,0.0001
|
| 5 |
+
3,0.08228224498931976,0.08228224498931976,0.0,0.0,0.0,0.12307979383816321,0.12307979383816321,0.0,0.0,0.9627499952912331,0.9356250166893005,0.9456250220537186,0.9827083498239517,0.9704166899124781,0.979375014702479,0.9627499952912331,0.0001
|
| 6 |
+
4,0.07740738441951929,0.07740738441951929,0.0,0.0,0.0,0.1594851796204845,0.1594851796204845,0.0,0.0,0.9563333218296369,0.9633333583672842,0.918541669845581,0.973750022550424,0.9568750237425169,0.9691666836539904,0.9563333218296369,0.0001
|
| 7 |
+
5,0.07453417447194464,0.07453417447194464,0.0,0.0,0.0,0.2358288150280714,0.2358288150280714,0.0,0.0,0.9400833348433176,0.9766666889190674,0.895833320915699,0.9422916869322459,0.9306249991059303,0.9550000230471293,0.9400833348433176,0.0001
|
| 8 |
+
6,0.07097108407662531,0.07097108407662531,0.0,0.0,0.0,0.3129023462533951,0.3129023462533951,0.0,0.0,0.9381666655341784,0.9602083563804626,0.9404166663686434,0.928125003973643,0.9156249910593033,0.9464583496252695,0.9381666655341784,0.0001
|
| 9 |
+
7,0.06843958382070174,0.06843958382070174,0.0,0.0,0.0,0.13095395577450594,0.13095395577450594,0.0,0.0,0.9624166761835417,0.9412500262260437,0.9677083566784859,0.9685416941841444,0.9602083538969358,0.9743750194708506,0.9624166761835417,0.0001
|
| 10 |
+
8,0.0501526239916797,0.0501526239916797,0.0,0.0,0.0,0.3366285723944505,0.3366285723944505,0.0,0.0,0.9330416594942411,0.979375017186006,0.8827083185315132,0.9383333449562391,0.926458328962326,0.9383333499232928,0.9330416594942411,1e-05
|
| 11 |
+
9,0.044897408176192956,0.044897408176192956,0.0,0.0,0.0,0.3811467296133439,0.3811467296133439,0.0,0.0,0.9304999932646751,0.9810416847467422,0.8858333230018616,0.9345833336313566,0.9181249986092249,0.9329166834553083,0.9304999932646751,1e-05
|
| 12 |
+
10,0.04343909534966181,0.04343909534966181,0.0,0.0,0.0,0.5174223159750303,0.5174223159750303,0.0,0.0,0.9125416701038679,0.9812500153978666,0.8581249887744585,0.9110416521628698,0.8943749864896139,0.9179166704416275,0.9125416701038679,1e-05
|
| 13 |
+
11,0.0425043538421211,0.0425043538421211,0.0,0.0,0.0,0.40790603185693425,0.40790603185693425,0.0,0.0,0.9270416547854742,0.9810416847467422,0.8843749910593033,0.9272916639844576,0.910624993344148,0.9318750177820524,0.9270416547854742,1e-05
|
| 14 |
+
12,0.04150725317051504,0.04150725317051504,0.0,0.0,0.0,0.6614557256301244,0.6614557256301244,0.0,0.0,0.9044166778524717,0.979583352804184,0.838541661699613,0.9079166601101557,0.8866666480898857,0.9093750094374021,0.9044166778524717,1e-05
|
| 15 |
+
13,0.04046413996368702,0.04046413996368702,0.0,0.0,0.0,0.7155438661575317,0.7155438661575317,0.0,0.0,0.897833339869976,0.9827083498239517,0.8383333260814348,0.8962499896685282,0.8710416530569395,0.9008333211143812,0.897833339869976,1e-05
|
| 16 |
+
14,0.03931607918702331,0.03931607918702331,0.0,0.0,0.0,0.854375367363294,0.854375367363294,0.0,0.0,0.893208327392737,0.9789583534002304,0.8308333257834116,0.8899999956289927,0.8658333271741867,0.9004166622956594,0.893208327392737,1e-05
|
| 17 |
+
15,0.037304425933250646,0.037304425933250646,0.0,0.0,0.0,0.6599167610208193,0.6599167610208193,0.0,0.0,0.9045833249886831,0.9822916835546494,0.8424999912579855,0.9054166600108147,0.8839583272735277,0.908750000099341,0.9045833249886831,1.0000000000000002e-06
|