Update app.py
Browse files
app.py
CHANGED
|
@@ -34,12 +34,11 @@ login(token =HUGGINGFACE_TOKEN)
|
|
| 34 |
|
| 35 |
|
| 36 |
|
| 37 |
-
Load the processor and model
|
| 38 |
processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
|
| 39 |
model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large")
|
| 40 |
processor1 = BlipProcessor.from_pretrained("noamrot/FuseCap")
|
| 41 |
model2 = BlipForConditionalGeneration.from_pretrained("noamrot/FuseCap")
|
| 42 |
-
|
| 43 |
from diffusers import FluxPipeline
|
| 44 |
|
| 45 |
pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", torch_dtype=torch.bfloat16)
|
|
|
|
| 34 |
|
| 35 |
|
| 36 |
|
|
|
|
| 37 |
processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
|
| 38 |
model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large")
|
| 39 |
processor1 = BlipProcessor.from_pretrained("noamrot/FuseCap")
|
| 40 |
model2 = BlipForConditionalGeneration.from_pretrained("noamrot/FuseCap")
|
| 41 |
+
|
| 42 |
from diffusers import FluxPipeline
|
| 43 |
|
| 44 |
pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", torch_dtype=torch.bfloat16)
|