Instructions to use tvandang/ai-boba with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Diffusers
How to use tvandang/ai-boba with Diffusers:
pip install -U diffusers transformers accelerate
import torch from diffusers import DiffusionPipeline # switch to "mps" for apple devices pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", dtype=torch.bfloat16, device_map="cuda") pipe.load_lora_weights("tvandang/ai-boba") prompt = "BOBA" image = pipe(prompt).images[0] - Inference
- Notebooks
- Google Colab
- Kaggle
- Local Apps
- Draw Things
- DiffusionBee
import torch
from diffusers import DiffusionPipeline
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", dtype=torch.bfloat16, device_map="cuda")
pipe.load_lora_weights("tvandang/ai-boba")
prompt = "BOBA"
image = pipe(prompt).images[0]Ai Boba
-->Trained on Replicate using:
https://replicate.com/ostris/flux-dev-lora-trainer/train
Trigger words
You should use BOBA to trigger the image generation.
Use it with the 🧨 diffusers library
from diffusers import AutoPipelineForText2Image
import torch
pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda')
pipeline.load_lora_weights('tvandang/ai-boba', weight_name='lora.safetensors')
image = pipeline('your prompt').images[0]
For more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers
- Downloads last month
- -
Model tree for tvandang/ai-boba
Base model
black-forest-labs/FLUX.1-dev