james-ocf commited on
Commit
c4d1d47
·
verified ·
1 Parent(s): bd82eee

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. README.md +8 -2
  2. data_config.yaml +4 -3
  3. model_config.yaml +401 -0
  4. pytorch_model.bin +2 -2
README.md CHANGED
@@ -39,7 +39,7 @@ Data is prepared with the `ocf_data_sampler/torch_datasets/datasets/pvnet_uk` Da
39
  ## Results
40
 
41
  The training logs for the current model can be found here:
42
- - [https://wandb.ai/openclimatefix/pvnet2.1/runs/aswcdyxy](https://wandb.ai/openclimatefix/pvnet2.1/runs/aswcdyxy)
43
 
44
 
45
  The training logs for all model runs of PVNet2 can be found [here](https://wandb.ai/openclimatefix/pvnet2.1).
@@ -53,5 +53,11 @@ Trained on a single NVIDIA Tesla T4
53
 
54
  ### Software
55
 
 
 
56
  - [1] https://github.com/openclimatefix/PVNet
57
- - [2] https://github.com/openclimatefix/ocf-data-sampler
 
 
 
 
 
39
  ## Results
40
 
41
  The training logs for the current model can be found here:
42
+ - [https://wandb.ai/openclimatefix/pvnet2.1/runs/rydw0jo0](https://wandb.ai/openclimatefix/pvnet2.1/runs/rydw0jo0)
43
 
44
 
45
  The training logs for all model runs of PVNet2 can be found [here](https://wandb.ai/openclimatefix/pvnet2.1).
 
53
 
54
  ### Software
55
 
56
+ This model was trained using the following Open Climate Fix packages:
57
+
58
  - [1] https://github.com/openclimatefix/PVNet
59
+ - [2] https://github.com/openclimatefix/ocf-data-sampler
60
+
61
+ The versions of these packages can be found below:
62
+ - pvnet==4.1.6
63
+ - ocf-data-sampler==0.2.17.post0+git.563e6370.dirty
data_config.yaml CHANGED
@@ -3,6 +3,7 @@ general:
3
  name: PVNet current
4
  input_data:
5
  gsp:
 
6
  dropout_fraction: 0
7
  dropout_timedeltas_minutes: []
8
  interval_end_minutes: 480
@@ -24,7 +25,7 @@ input_data:
24
  - mcc
25
  - lcc
26
  - tcc
27
- - sde
28
  - sr
29
  - duvrs
30
  - u10
@@ -59,7 +60,7 @@ input_data:
59
  mcc:
60
  mean: 0.3288780450820923
61
  std: 0.38039860129356384
62
- sde:
63
  mean: 8.107526082312688e-05
64
  std: 0.000913831521756947
65
  t2m:
@@ -191,6 +192,6 @@ input_data:
191
  time_resolution_minutes: 5
192
  zarr_path: PLACEHOLDER.zarr
193
  solar_position:
194
- interval_end_minutes: 480
195
  interval_start_minutes: -120
196
  time_resolution_minutes: 30
 
3
  name: PVNet current
4
  input_data:
5
  gsp:
6
+ boundaries_version: '20250109'
7
  dropout_fraction: 0
8
  dropout_timedeltas_minutes: []
9
  interval_end_minutes: 480
 
25
  - mcc
26
  - lcc
27
  - tcc
28
+ - sd
29
  - sr
30
  - duvrs
31
  - u10
 
60
  mcc:
61
  mean: 0.3288780450820923
62
  std: 0.38039860129356384
63
+ sd:
64
  mean: 8.107526082312688e-05
65
  std: 0.000913831521756947
66
  t2m:
 
192
  time_resolution_minutes: 5
193
  zarr_path: PLACEHOLDER.zarr
194
  solar_position:
195
+ interval_end_minutes: 2160
196
  interval_start_minutes: -120
197
  time_resolution_minutes: 30
model_config.yaml ADDED
@@ -0,0 +1,401 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _target_: pvnet.models.multimodal.multimodal.Model
2
+ output_quantiles:
3
+ - 0.02
4
+ - 0.1
5
+ - 0.25
6
+ - 0.5
7
+ - 0.75
8
+ - 0.9
9
+ - 0.98
10
+ nwp_encoders_dict:
11
+ ukv:
12
+ _target_: pvnet.models.multimodal.encoders.encoders3d.DefaultPVNet
13
+ _partial_: true
14
+ in_channels: 11
15
+ out_features: 256
16
+ number_of_conv3d_layers: 6
17
+ conv3d_channels: 32
18
+ image_size_pixels: 24
19
+ ecmwf:
20
+ _target_: pvnet.models.multimodal.encoders.encoders3d.DefaultPVNet
21
+ _partial_: true
22
+ in_channels: 12
23
+ out_features: 256
24
+ number_of_conv3d_layers: 4
25
+ conv3d_channels: 32
26
+ image_size_pixels: 12
27
+ sat_encoder:
28
+ _target_: pvnet.models.multimodal.encoders.encoders3d.ResConv3DNet2
29
+ _partial_: true
30
+ in_channels: 11
31
+ out_features: 256
32
+ image_size_pixels: 24
33
+ hidden_channels: 32
34
+ n_res_blocks: 3
35
+ res_block_layers: 3
36
+ batch_norm: true
37
+ dropout_frac: 0.0
38
+ add_image_embedding_channel: false
39
+ pv_encoder: null
40
+ output_network:
41
+ _target_: pvnet.models.multimodal.linear_networks.networks.ResFCNet2
42
+ _partial_: true
43
+ fc_hidden_features: 128
44
+ n_res_blocks: 6
45
+ res_block_layers: 2
46
+ dropout_frac: 0.0
47
+ embedding_dim: 16
48
+ include_sun: true
49
+ include_gsp_yield_history: false
50
+ forecast_minutes: 480
51
+ history_minutes: 120
52
+ min_sat_delay_minutes: 30
53
+ sat_history_minutes: 60
54
+ pv_history_minutes: 180
55
+ nwp_history_minutes:
56
+ ukv: 120
57
+ ecmwf: 120
58
+ nwp_forecast_minutes:
59
+ ukv: 480
60
+ ecmwf: 480
61
+ optimizer:
62
+ _target_: pvnet.optimizers.EmbAdamWReduceLROnPlateau
63
+ lr: 0.0001
64
+ weight_decay: 0.25
65
+ amsgrad: true
66
+ patience: 10
67
+ factor: 0.1
68
+ threshold: 0.002
69
+ location_id_mapping:
70
+ 1: 1
71
+ 2: 2
72
+ 3: 3
73
+ 4: 4
74
+ 6: 5
75
+ 7: 6
76
+ 8: 7
77
+ 9: 8
78
+ 10: 9
79
+ 11: 10
80
+ 12: 11
81
+ 13: 12
82
+ 14: 13
83
+ 15: 14
84
+ 16: 15
85
+ 18: 16
86
+ 19: 17
87
+ 20: 18
88
+ 21: 19
89
+ 22: 20
90
+ 23: 21
91
+ 24: 22
92
+ 25: 23
93
+ 26: 24
94
+ 27: 25
95
+ 28: 26
96
+ 29: 27
97
+ 30: 28
98
+ 31: 29
99
+ 32: 30
100
+ 33: 31
101
+ 34: 32
102
+ 35: 33
103
+ 36: 34
104
+ 37: 35
105
+ 38: 36
106
+ 39: 37
107
+ 40: 38
108
+ 41: 39
109
+ 42: 40
110
+ 43: 41
111
+ 44: 42
112
+ 45: 43
113
+ 46: 44
114
+ 47: 45
115
+ 48: 46
116
+ 49: 47
117
+ 50: 48
118
+ 51: 49
119
+ 52: 50
120
+ 54: 51
121
+ 55: 52
122
+ 56: 53
123
+ 57: 54
124
+ 58: 55
125
+ 59: 56
126
+ 60: 57
127
+ 61: 58
128
+ 62: 59
129
+ 63: 60
130
+ 64: 61
131
+ 65: 62
132
+ 66: 63
133
+ 67: 64
134
+ 68: 65
135
+ 69: 66
136
+ 70: 67
137
+ 71: 68
138
+ 72: 69
139
+ 73: 70
140
+ 74: 71
141
+ 76: 72
142
+ 77: 73
143
+ 78: 74
144
+ 79: 75
145
+ 80: 76
146
+ 81: 77
147
+ 82: 78
148
+ 83: 79
149
+ 84: 80
150
+ 85: 81
151
+ 86: 82
152
+ 87: 83
153
+ 88: 84
154
+ 89: 85
155
+ 90: 86
156
+ 91: 87
157
+ 92: 88
158
+ 93: 89
159
+ 94: 90
160
+ 95: 91
161
+ 96: 92
162
+ 97: 93
163
+ 98: 94
164
+ 99: 95
165
+ 100: 96
166
+ 101: 97
167
+ 102: 98
168
+ 103: 99
169
+ 104: 100
170
+ 105: 101
171
+ 106: 102
172
+ 107: 103
173
+ 108: 104
174
+ 109: 105
175
+ 110: 106
176
+ 111: 107
177
+ 112: 108
178
+ 113: 109
179
+ 114: 110
180
+ 115: 111
181
+ 116: 112
182
+ 117: 113
183
+ 118: 114
184
+ 119: 115
185
+ 120: 116
186
+ 121: 117
187
+ 122: 118
188
+ 123: 119
189
+ 124: 120
190
+ 125: 121
191
+ 126: 122
192
+ 127: 123
193
+ 128: 124
194
+ 129: 125
195
+ 130: 126
196
+ 131: 127
197
+ 132: 128
198
+ 133: 129
199
+ 134: 130
200
+ 135: 131
201
+ 136: 132
202
+ 137: 133
203
+ 138: 134
204
+ 141: 135
205
+ 142: 136
206
+ 144: 137
207
+ 145: 138
208
+ 146: 139
209
+ 147: 140
210
+ 148: 141
211
+ 149: 142
212
+ 150: 143
213
+ 151: 144
214
+ 152: 145
215
+ 153: 146
216
+ 154: 147
217
+ 155: 148
218
+ 156: 149
219
+ 158: 150
220
+ 159: 151
221
+ 160: 152
222
+ 161: 153
223
+ 162: 154
224
+ 164: 155
225
+ 165: 156
226
+ 166: 157
227
+ 167: 158
228
+ 168: 159
229
+ 169: 160
230
+ 170: 161
231
+ 171: 162
232
+ 172: 163
233
+ 173: 164
234
+ 174: 165
235
+ 175: 166
236
+ 176: 167
237
+ 177: 168
238
+ 178: 169
239
+ 179: 170
240
+ 180: 171
241
+ 181: 172
242
+ 182: 173
243
+ 183: 174
244
+ 184: 175
245
+ 185: 176
246
+ 186: 177
247
+ 187: 178
248
+ 188: 179
249
+ 189: 180
250
+ 190: 181
251
+ 191: 182
252
+ 192: 183
253
+ 193: 184
254
+ 194: 185
255
+ 195: 186
256
+ 196: 187
257
+ 197: 188
258
+ 198: 189
259
+ 199: 190
260
+ 200: 191
261
+ 201: 192
262
+ 202: 193
263
+ 203: 194
264
+ 204: 195
265
+ 205: 196
266
+ 206: 197
267
+ 207: 198
268
+ 208: 199
269
+ 209: 200
270
+ 210: 201
271
+ 211: 202
272
+ 212: 203
273
+ 213: 204
274
+ 214: 205
275
+ 215: 206
276
+ 216: 207
277
+ 217: 208
278
+ 218: 209
279
+ 219: 210
280
+ 220: 211
281
+ 221: 212
282
+ 222: 213
283
+ 223: 214
284
+ 224: 215
285
+ 226: 216
286
+ 227: 217
287
+ 228: 218
288
+ 229: 219
289
+ 230: 220
290
+ 231: 221
291
+ 232: 222
292
+ 233: 223
293
+ 234: 224
294
+ 235: 225
295
+ 236: 226
296
+ 237: 227
297
+ 238: 228
298
+ 239: 229
299
+ 240: 230
300
+ 241: 231
301
+ 242: 232
302
+ 243: 233
303
+ 244: 234
304
+ 245: 235
305
+ 246: 236
306
+ 247: 237
307
+ 248: 238
308
+ 249: 239
309
+ 250: 240
310
+ 251: 241
311
+ 252: 242
312
+ 253: 243
313
+ 254: 244
314
+ 255: 245
315
+ 256: 246
316
+ 257: 247
317
+ 258: 248
318
+ 259: 249
319
+ 260: 250
320
+ 261: 251
321
+ 262: 252
322
+ 263: 253
323
+ 264: 254
324
+ 265: 255
325
+ 266: 256
326
+ 267: 257
327
+ 268: 258
328
+ 269: 259
329
+ 270: 260
330
+ 271: 261
331
+ 272: 262
332
+ 273: 263
333
+ 274: 264
334
+ 275: 265
335
+ 276: 266
336
+ 277: 267
337
+ 278: 268
338
+ 279: 269
339
+ 280: 270
340
+ 281: 271
341
+ 282: 272
342
+ 283: 273
343
+ 284: 274
344
+ 285: 275
345
+ 286: 276
346
+ 287: 277
347
+ 288: 278
348
+ 289: 279
349
+ 290: 280
350
+ 291: 281
351
+ 292: 282
352
+ 293: 283
353
+ 294: 284
354
+ 295: 285
355
+ 296: 286
356
+ 297: 287
357
+ 298: 288
358
+ 299: 289
359
+ 300: 290
360
+ 301: 291
361
+ 302: 292
362
+ 303: 293
363
+ 304: 294
364
+ 305: 295
365
+ 306: 296
366
+ 307: 297
367
+ 308: 298
368
+ 309: 299
369
+ 311: 300
370
+ 312: 301
371
+ 313: 302
372
+ 314: 303
373
+ 315: 304
374
+ 316: 305
375
+ 317: 306
376
+ 318: 307
377
+ 319: 308
378
+ 320: 309
379
+ 321: 310
380
+ 322: 311
381
+ 323: 312
382
+ 324: 313
383
+ 325: 314
384
+ 326: 315
385
+ 327: 316
386
+ 328: 317
387
+ 329: 318
388
+ 330: 319
389
+ 331: 320
390
+ 332: 321
391
+ 333: 322
392
+ 334: 323
393
+ 335: 324
394
+ 336: 325
395
+ 337: 326
396
+ 338: 327
397
+ 339: 328
398
+ 340: 329
399
+ 341: 330
400
+ 342: 331
401
+ adapt_batches: true
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a5dfc56237cf1f93d62d552622f81049316d0209a343da8adef0008e54d470d
3
- size 34546610
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b167289e37df64574c50d76fe466c8d6423ac9f105250229a5c85d152a6ae6fe
3
+ size 34547915