Code for instruction-tuned cartoonization with Diffusion models.
Refer to the data_preparation
directory.
accelerate launch --mixed_precision="fp16" finetune_instruct_pix2pix.py \
--pretrained_model_name_or_path=timbrooks/instruct-pix2pix \
--dataset_name=sayakpaul/cartoonizer-dataset \
--use_ema \
--enable_xformers_memory_efficient_attention \
--resolution=256 --random_flip \
--train_batch_size=2 --gradient_accumulation_steps=4 --gradient_checkpointing \
--max_train_steps=15000 \
--checkpointing_steps=5000 --checkpoints_total_limit=1 \
--learning_rate=5e-05 --lr_warmup_steps=0 \
--mixed_precision=fp16 \
--val_image_url="https://hf.co/datasets/diffusers/diffusers-images-docs/resolve/main/mountain.png" \
--validation_prompt="Generate a cartoonized version of the natural image" \
--seed=42 \
--report_to=wandb
python cartoonize.py \
--image_path https://hf.co/datasets/diffusers/diffusers-images-docs/resolve/main/mountain.png \
--concept mountain
By default instruction-tuning-vision/instruction-tuned-cartoonizer
model will be used. You can also set --model_id
to be timbrooks/instruct-pix2pix
to use a pre-trained InstructPix2Pix model.
Refer to the validation
directory.