Upload folder using huggingface_hub
Browse files- src/.pipeline.py.swp +0 -0
- src/pipeline.py +3 -3
src/.pipeline.py.swp
CHANGED
|
Binary files a/src/.pipeline.py.swp and b/src/.pipeline.py.swp differ
|
|
|
src/pipeline.py
CHANGED
|
@@ -11,8 +11,7 @@ from torch import Generator
|
|
| 11 |
from torchao.quantization import quantize_, int8_weight_only
|
| 12 |
from transformers import T5EncoderModel, CLIPTextModel
|
| 13 |
|
| 14 |
-
|
| 15 |
-
# torch.backends.cuda.matmul.allow_tf32 = True
|
| 16 |
Pipeline: TypeAlias = FluxPipeline
|
| 17 |
|
| 18 |
CHECKPOINT = "jokerbit/flux.1-schnell-Robert-int8wo"
|
|
@@ -53,7 +52,8 @@ def load_pipeline() -> Pipeline:
|
|
| 53 |
torch_dtype=torch.bfloat16,
|
| 54 |
use_safetensors=False,
|
| 55 |
)
|
| 56 |
-
|
|
|
|
| 57 |
pipeline = FluxPipeline.from_pretrained(
|
| 58 |
CHECKPOINT,
|
| 59 |
revision=REVISION,
|
|
|
|
| 11 |
from torchao.quantization import quantize_, int8_weight_only
|
| 12 |
from transformers import T5EncoderModel, CLIPTextModel
|
| 13 |
|
| 14 |
+
|
|
|
|
| 15 |
Pipeline: TypeAlias = FluxPipeline
|
| 16 |
|
| 17 |
CHECKPOINT = "jokerbit/flux.1-schnell-Robert-int8wo"
|
|
|
|
| 52 |
torch_dtype=torch.bfloat16,
|
| 53 |
use_safetensors=False,
|
| 54 |
)
|
| 55 |
+
import pdb
|
| 56 |
+
pdb.set_trace()
|
| 57 |
pipeline = FluxPipeline.from_pretrained(
|
| 58 |
CHECKPOINT,
|
| 59 |
revision=REVISION,
|