ProGamerGov commited on
Commit
df68f4a
·
verified ·
1 Parent(s): 6d0dd7e

Fix training details

Browse files
Files changed (1) hide show
  1. README.md +9 -7
README.md CHANGED
@@ -35,16 +35,18 @@ The model is designed to be capable of producing equirectangular images that can
35
 
36
  ### Training Details
37
 
38
- Training was first performed using nf4 for 8 epochs.
39
- - `qwen-360-diffusion-int4-bf16-v1.safetensors` was trained for 7 epochs.
40
- - `qwen-360-diffusion-int4-bf16-v1-b.safetensors` was trained for 8 epochs.
41
 
42
- Training then continued at int8 for another 4 epochs.
43
- - `qwen-360-diffusion-int8-bf16-v1.safetensors` was trained for a total of 12 epochs.
44
 
45
- Each 360 degree training image was randomly rotated horizontally 3 times for data augmentation (original + 3 rotations).
46
 
47
- For regularization, 64k images were randomly selected from the [pexels-568k-internvl2](https://huggingface.co/datasets/CaptionEmporium/pexels-568k-internvl2) dataset.
 
 
 
 
 
48
 
49
  ---
50
 
 
35
 
36
  ### Training Details
37
 
38
+ The training dataset consists of 32k unique 360 degree equirectangular images. Each image was randomly rotated horizontally 3 times for data augmentation (original + 3 rotations), providing a total of 128k training images. All 32k original 360 images were manually checked by humans for seams, polar artifacts, incorrect distortions, and other problems before their inclusion in the dataset.
 
 
39
 
40
+ For regularization, 64k images were randomly selected from the [pexels-568k-internvl2](https://huggingface.co/datasets/CaptionEmporium/pexels-568k-internvl2) dataset and added to the training set.
 
41
 
42
+ **Training timeline:** 3 months and 23 days
43
 
44
+ Training was first performed using nf4 quantization for 32 epochs (8 epochs counting the original + augmentations as a single epoch):
45
+ - `qwen-360-diffusion-int4-bf16-v1.safetensors` was trained for 28 epochs (1,344,000 steps)
46
+ - `qwen-360-diffusion-int4-bf16-v1-b.safetensors` was trained for 32 epochs (1,536,000 steps)
47
+
48
+ Training then continued at int8 quantization for another 16 epochs (4 epochs counting the original + augmentations as a single epoch):
49
+ - `qwen-360-diffusion-int8-bf16-v1.safetensors` was trained for a total of 48 epochs (2,304,000 steps)
50
 
51
  ---
52