You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
However, when I generate an image using the same model files in ComfyUI, everything works just fine (and quite fast). It DOES complain about VRAM but only during VAE decoding:
got prompt
Using scaled fp8: fp8 matrix mult: False, scale input: True
model weight dtype torch.float16, manual cast: None
model_type FLOW
Using split attention in VAE
Using split attention in VAE
VAE load device: cuda:0, offload device: cpu, dtype: torch.float32
Using scaled fp8: fp8 matrix mult: False, scale input: False
CLIP/text encoder model load device: cuda:0, offload device: cpu, current: cpu, dtype: torch.float16
Requested to load SD3ClipModel_
loaded completely 15122.8 6228.190093994141 True
/root/ComfyUI/comfy/ops.py:80: UserWarning: Attempting to use hipBLASLt on an unsupported architecture! Overriding blas backend to hipblas (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:331.)
return torch.nn.functional.linear(input, weight, bias)
Requested to load SD3
loaded completely 11615.620722656251 7683.561706542969 True
100%|██████████| 20/20 [01:30<00:00, 4.51s/it]
Requested to load AutoencodingEngine
loaded completely 3787.8148437500004 319.7467155456543 True
Warning: Ran out of memory when regular VAE decoding, retrying with tiled VAE decoding.
Prompt executed in 166.86 seconds
Now I'm wondering:
Maybe the model only narrowly fits into VRAM and the LocalAI stack has a little more overhead?
Maybe the model does not fit into VRAM but ComfyUI manages to offload some to the regular RAM which LocalAI doesn't?
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
-
Hello erverybody,
I wonder if somebody here may have an idea about a little problem.
I have an AMD Radeon RX 6800 XT with 16 GB of VRAM and I'm trying to set up image generation using SD 3.5. I am using this model:
https://huggingface.co/Comfy-Org/stable-diffusion-3.5-fp8/blob/main/sd3.5_large_fp8_scaled.safetensors
This is the model configuration:
LocalAI reports that it is out of VRAM:
However, when I generate an image using the same model files in ComfyUI, everything works just fine (and quite fast). It DOES complain about VRAM but only during VAE decoding:
Now I'm wondering:
Can somebody point me in the right direction?
Thank you very much in advance!
Beta Was this translation helpful? Give feedback.
All reactions