Text-to-Image
Diffusers
diffusers-training
lora
template:sd-lora
stable-diffusion-xl
stable-diffusion-xl-diffusers
import torch
from diffusers import DiffusionPipeline
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-base-1.0", dtype=torch.bfloat16, device_map="cuda")
pipe.load_lora_weights("PhilSad/outputs")
prompt = "interior design in sks style"
image = pipe(prompt).images[0]SDXL LoRA DreamBooth - PhilSad/outputs
Model description
These are PhilSad/outputs LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: None.
Trigger words
You should use interior design in sks style to trigger the image generation.
Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
Intended uses & limitations
How to use
# TODO: add an example code snippet for running this diffusion pipeline
Limitations and bias
[TODO: provide examples of latent issues and potential remediations]
Training details
[TODO: describe the data used to train the model]
- Downloads last month
- 5
Model tree for PhilSad/outputs
Base model
stabilityai/stable-diffusion-xl-base-1.0