Fgdfgfthgr commited on
Commit
5932b89
·
verified ·
1 Parent(s): 9381b0e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -7
README.md CHANGED
@@ -5,7 +5,13 @@ datasets:
5
  ---
6
  ## Check out my [blog](https://huggingface.co/blog/Fgdfgfthgr/typical-anime-image-style-dim)!
7
 
8
- # You can use 6 numbers to fully describe the style of an (anime) image!
 
 
 
 
 
 
9
 
10
  ## What's it and what could it do?
11
  Many diffusion models, though, choose to use artist tags to control the style of output images.
@@ -33,16 +39,16 @@ With current version (v3):
33
 
34
  Training was done using [PyTorch Lightning](https://lightning.ai/).
35
 
36
- lr = 0.0001
37
 
38
- weight_decay = 0.0001
39
 
40
- [AdEMAMix](https://github.com/apple/ml-ademamix) optimizer
41
 
42
  ExponentialLR scheduler, with a gamma of 0.99, applied every epoch.
43
 
44
- Batch size of 1. [accumulate_grad_batches](https://lightning.ai/docs/pytorch/stable/advanced/training_tricks.html) of 16.
45
 
46
- With every anchor image, 16 positive images and 16 negative images are used.
47
 
48
- Trained for 15 epoches. On 2 A100 GPUs. A total of 3434 optimizer updates.
 
5
  ---
6
  ## Check out my [blog](https://huggingface.co/blog/Fgdfgfthgr/typical-anime-image-style-dim)!
7
 
8
+ # Update 17/10/2025
9
+ V4 released! This time instead of training a vision model from scratch,
10
+ it uses a simple mlp that takes the cls token from a [DINOv3](https://huggingface.co/collections/facebook/dinov3-68924841bd6b561778e31009) model to get the embedding.
11
+
12
+ Far more accurate than the previous V3! You do need access to the DINOv3 with your HuggingFace token, though.
13
+
14
+ # You can use 6/7 numbers to fully describe the style of an (anime) image!
15
 
16
  ## What's it and what could it do?
17
  Many diffusion models, though, choose to use artist tags to control the style of output images.
 
39
 
40
  Training was done using [PyTorch Lightning](https://lightning.ai/).
41
 
42
+ lr = 0.0005
43
 
44
+ weight_decay = 0.01
45
 
46
+ AdamW optimizer
47
 
48
  ExponentialLR scheduler, with a gamma of 0.99, applied every epoch.
49
 
50
+ Batch size of 9999 (so all data goes through the network at once).
51
 
52
+ With every anchor image, 4 positive images and 16 negative images are used.
53
 
54
+ Trained for 150 epoches. On a single RTX 3080 GPUs. A total of 150 optimizer updates.