-
Notifications
You must be signed in to change notification settings - Fork 125
Description
Hello,
When using the StableSiffusion3Pipeline, I've noticed that if the input prompt is too long, it gets truncated, which leads to suboptimal generation results. Is there any recommended way to address this issue? For example, is it possible to increase the prompt length limit via parameters, or are there suggested strategies for splitting or merging prompts?
I would appreciate any technical advice or suggestions for optimization. Thank you!
The following code is the exact script I ran when encountering this issue.
`import torch
from diffusers import StableDiffusion3Pipeline
pipe = StableDiffusion3Pipeline.from_pretrained("stabilityai/stable-diffusion-3.5-large", torch_dtype=torch.bfloat16)
pipe = pipe.to("cuda")
prompt = "A whimsical and creative image depicting a hybrid creature that is a mix of a waffle and a hippopotamus, basking in a river of melted butter amidst a breakfast-themed landscape. It features the distinctive, bulky body shape of a hippo. However, instead of the usual grey skin, the creature's body resembles a golden-brown, crispy waffle fresh off the griddle. The skin is textured with the familiar grid pattern of a waffle, each square filled with a glistening sheen of syrup. The environment combines the natural habitat of a hippo with elements of a breakfast table setting, a river of warm, melted butter, with oversized utensils or plates peeking out from the lush, pancake-like foliage in the background, a towering pepper mill standing in for a tree. As the sun rises in this fantastical world, it casts a warm, buttery glow over the scene. The creature, content in its butter river, lets out a yawn. Nearby, a flock of birds take flight"
image = pipe(
prompt=prompt,
num_inference_steps=28,
guidance_scale=3.5,
).images[0]
image.save("capybara.png")
`