Using Flux.1-dev on EC2 instance #11230
Devank-Garg
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, I am trying to using flux.1-dev on an EC2 cluster (g5.12xlarge). Which has 4 A10s.
import torch
from diffusers import FluxPipeline
from accelerate import PartialState
model_path = "black-forest-labs/FLUX.1-dev"
pipe = FluxPipeline.from_pretrained(model_path, torch_dtype=torch.bfloat16, device_map="balanced", max_memory={0:"24GB", 1:"24GB", 2:"24GB", 3:"24GB"})
#pipe.enable_model_cpu_offload() #save some VRAM by offloading the model to CPU. Remove this if you have enough GPU power
#pipe.enable_sequential_cpu_offload()
print(pipe.hf_device_map)
prompt = " cat holding a sign that says hello world lovely, 8k"
image = pipe(
prompt,
height=1024,
width=1024,
guidance_scale=3.5,
output_type="pil",
num_inference_steps=20,
max_sequence_length=512,
).images[0]
image.save("flux-dev.png")
I am trying to use this snippet but I am getting
{'transformer': 'cpu', 'text_encoder_2': 0, 'text_encoder': 1, 'vae': 2}
from print(pipe.hf_device_map)
Can anyone tell me what else can I do to increase inference speed?
Beta Was this translation helpful? Give feedback.
All reactions