Lovre commited on
Commit
a5fdb19
·
verified ·
1 Parent(s): eeee45c

Update FINAL_TRAINING_REPORT.md

Browse files
Files changed (1) hide show
  1. FINAL_TRAINING_REPORT.md +6 -6
FINAL_TRAINING_REPORT.md CHANGED
@@ -19,10 +19,10 @@ training:
19
  batch_size: 64
20
 
21
  # DataLoader optimizations
22
- num_workers: 4 # Number of parallel data loading processes (0 for main thread only)
23
- pin_memory: true # Pin memory for faster CPU->GPU transfer
24
- persistent_workers: true # Keep workers alive between epochs (only if num_workers > 0)
25
- prefetch_factor: 2 # Number of batches to prefetch per worker (only if num_workers > 0)
26
 
27
  # Smooth curriculum over digit lengths (optional)
28
  curriculum:
@@ -73,8 +73,8 @@ evaluation:
73
  final_samples_per_combination: 1000 # Samples per digit combination for final evaluation
74
 
75
  logging:
76
- interval_examples: 5000 # Log training loss every N examples (was: 250 batches * 32 batch_size = 8000 examples)
77
- save_path: "models/minimal_example_lora" # Model will be uploaded to HuggingFace as your-username/minimal-addition-lora
78
  ```
79
 
80
  ## Training Summary
 
19
  batch_size: 64
20
 
21
  # DataLoader optimizations
22
+ num_workers: 4
23
+ pin_memory: true
24
+ persistent_workers: true
25
+ prefetch_factor: 2
26
 
27
  # Smooth curriculum over digit lengths (optional)
28
  curriculum:
 
73
  final_samples_per_combination: 1000 # Samples per digit combination for final evaluation
74
 
75
  logging:
76
+ interval_examples: 5000 # Log training loss every N examples
77
+ save_path: "models/minimal_example_lora"
78
  ```
79
 
80
  ## Training Summary