chaechae7's picture
Upload transformer-only FP8 quantized runtime (scaled fp8)
a8eeee1 verified
raw
history blame contribute delete
431 Bytes
{
"_class_name": "Flux2Pipeline",
"_diffusers_version": "0.36.0.dev0",
"scheduler": [
"diffusers",
"FlowMatchEulerDiscreteScheduler"
],
"text_encoder": [
"transformers",
"Mistral3ForConditionalGeneration"
],
"tokenizer": [
"transformers",
"PixtralProcessor"
],
"transformer": [
"diffusers",
"Flux2Transformer2DModel"
],
"vae": [
"diffusers",
"AutoencoderKLFlux2"
]
}