Spaces:
Paused
Paused
Upload 2 files
Browse files- env.py +5 -0
- requirements.txt +3 -2
env.py
CHANGED
|
@@ -30,6 +30,8 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 30 |
'BlueDancer/Artisanica_XL',
|
| 31 |
'neta-art/neta-noob-1.0',
|
| 32 |
'OnomaAIResearch/Illustrious-xl-early-release-v0',
|
|
|
|
|
|
|
| 33 |
'Raelina/Rae-Diffusion-XL-V2',
|
| 34 |
'Raelina/Raemu-XL-V4',
|
| 35 |
'Raelina/Raemu-XL-V5',
|
|
@@ -45,6 +47,9 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 45 |
'Raelina/Raehoshi-illust-XL-7',
|
| 46 |
'Raelina/Raehoshi-illust-XL-7.1',
|
| 47 |
'Raelina/Raehoshi-illust-XL-8',
|
|
|
|
|
|
|
|
|
|
| 48 |
'camenduru/FLUX.1-dev-diffusers',
|
| 49 |
'black-forest-labs/FLUX.1-schnell',
|
| 50 |
'sayakpaul/FLUX.1-merged',
|
|
|
|
| 30 |
'BlueDancer/Artisanica_XL',
|
| 31 |
'neta-art/neta-noob-1.0',
|
| 32 |
'OnomaAIResearch/Illustrious-xl-early-release-v0',
|
| 33 |
+
'martineux/bismuth7-xl',
|
| 34 |
+
'martineux/perfectdeliberate8',
|
| 35 |
'Raelina/Rae-Diffusion-XL-V2',
|
| 36 |
'Raelina/Raemu-XL-V4',
|
| 37 |
'Raelina/Raemu-XL-V5',
|
|
|
|
| 47 |
'Raelina/Raehoshi-illust-XL-7',
|
| 48 |
'Raelina/Raehoshi-illust-XL-7.1',
|
| 49 |
'Raelina/Raehoshi-illust-XL-8',
|
| 50 |
+
'Raelina/Raehoshi-illust-XL-8.1',
|
| 51 |
+
'Raelina/Raehoshi-illust-XL-9',
|
| 52 |
+
'Raelina/Raehoshi-illust-vpred',
|
| 53 |
'camenduru/FLUX.1-dev-diffusers',
|
| 54 |
'black-forest-labs/FLUX.1-schnell',
|
| 55 |
'sayakpaul/FLUX.1-merged',
|
requirements.txt
CHANGED
|
@@ -3,7 +3,7 @@ diffusers
|
|
| 3 |
transformers
|
| 4 |
accelerate
|
| 5 |
huggingface_hub
|
| 6 |
-
hf_transfer
|
| 7 |
hf_xet
|
| 8 |
torch==2.5.1
|
| 9 |
torchvision
|
|
@@ -19,5 +19,6 @@ rapidfuzz
|
|
| 19 |
sentencepiece
|
| 20 |
unidecode
|
| 21 |
matplotlib-inline
|
|
|
|
| 22 |
https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.4.11/flash_attn-2.8.3+cu124torch2.5-cp310-cp310-linux_x86_64.whl
|
| 23 |
-
pydantic==2.10.6
|
|
|
|
| 3 |
transformers
|
| 4 |
accelerate
|
| 5 |
huggingface_hub
|
| 6 |
+
#hf_transfer
|
| 7 |
hf_xet
|
| 8 |
torch==2.5.1
|
| 9 |
torchvision
|
|
|
|
| 19 |
sentencepiece
|
| 20 |
unidecode
|
| 21 |
matplotlib-inline
|
| 22 |
+
mediapipe==0.10.5
|
| 23 |
https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.4.11/flash_attn-2.8.3+cu124torch2.5-cp310-cp310-linux_x86_64.whl
|
| 24 |
+
#pydantic==2.10.6
|