Skip to content

nuletizia/instruction-tuned-sd

Repository files navigation

Instruction-tuned Cartoonization

Code for instruction-tuned cartoonization with Diffusion models.

Data preparation

Refer to the data_preparation directory.

Command for launching training

accelerate launch --mixed_precision="fp16" finetune_instruct_pix2pix.py \
    --pretrained_model_name_or_path=timbrooks/instruct-pix2pix \
    --dataset_name=sayakpaul/cartoonizer-dataset \
    --use_ema \
    --enable_xformers_memory_efficient_attention \
    --resolution=256 --random_flip \
    --train_batch_size=2 --gradient_accumulation_steps=4 --gradient_checkpointing \
    --max_train_steps=15000 \
    --checkpointing_steps=5000 --checkpoints_total_limit=1 \
    --learning_rate=5e-05 --lr_warmup_steps=0 \
    --mixed_precision=fp16 \
    --val_image_url="https://hf.co/datasets/diffusers/diffusers-images-docs/resolve/main/mountain.png" \
    --validation_prompt="Generate a cartoonized version of the natural image" \
    --seed=42 \
    --report_to=wandb 

Inference

python cartoonize.py \
    --image_path https://hf.co/datasets/diffusers/diffusers-images-docs/resolve/main/mountain.png \
    --concept mountain

By default instruction-tuning-vision/instruction-tuned-cartoonizer model will be used. You can also set --model_id to be timbrooks/instruct-pix2pix to use a pre-trained InstructPix2Pix model.

Comparison across models

Refer to the validation directory.

Organization to keep track of the artifacts (datasets, models, etc.)

https://huggingface.co/instruction-tuning-vision

About

Code for instruction-tuning Stable Diffusion.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.9%
  • Makefile 0.1%