cartoon-control-lr_1e-4-wd_1e-4-gs_10.0-cd_0.1

These are Flux control weights trained on black-forest-labs/FLUX.1-dev with a new type of conditioning. instruction-tuning-sd/cartoonization dataset was used for training. You can find some example images below.

images_0)
images_1)
images_2)
prompt: Generate a cartoonized version of the image

License

Please adhere to the licensing terms as described here

Intended uses & limitations

How to use

from diffusers import FluxTransformer2DModel, FluxControlPipeline
from diffusers.utils import load_image
import torch 

path = "sayakpaul/cartoon-control-lr_1e-4-wd_1e-4-gs_10.0-cd_0.1"
transformer = FluxTransformer2DModel.from_pretrained(path, torch_dtype=torch.bfloat16)
pipe = FluxControlPipeline.from_pretrained(
  "black-forest-labs/FLUX.1-dev",  transformer=transformer, torch_dtype=torch.bfloat16
).to("cuda")

prompt = "Generate a cartoonized version of the image"
url = "https://huggingface.co/sayakpaul/cartoon-control-lr_1e-4-wd_1e-4-gs_10.0-cd_0.1/resolve/main/taj.jpg"

image = load_image(img).resize((1024, 1024))
gen_image = pipe(
    prompt=prompt,
    control_image=image,
    guidance_scale=10.,
    num_inference_steps=50,
    generator=torch.manual_seed(0),
    max_sequence_length=512,
).images[0]
gen_image.save("output.png")

Refer to the Flux Control docs here.

Limitations and bias

[TODO: provide examples of latent issues and potential remediations]

Training details

Refer to here. WandB logs are here.

Downloads last month
6
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for sayakpaul/cartoon-control-lr_1e-4-wd_1e-4-gs_10.0-cd_0.1

Finetuned
(294)
this model