Skip to content

DyArtbank: Diverse artistic style transfer via pre-trained stable diffusion and dynamic style prompt Artbank (KBS 2025)

License

Notifications You must be signed in to change notification settings

Jamie-Cheung/DyArtbank

Repository files navigation

DyArtbank

This repo contains the code implementation of the paper:

DyArtbank: Diverse artistic style transfer via pre-trained stable diffusion and dynamic style prompt Artbank (KBS 2025)

[arXiv]

Abstract

Artistic style transfer aims to transfer the learned style onto an arbitrary content image. However, most existing style transfer methods can only render consistent artistic stylized images, making it difficult for users to get enough stylized images to enjoy. To solve this issue, we propose a novel artistic style transfer framework called DyArtbank, which can generate diverse and highly realistic artistic stylized images. Specifically, we introduce a Dynamic Style Prompt ArtBank (DSPA), a set of learnable parameters. It can learn and store the style information from the collection of artworks, dynamically guiding pre-trained stable diffusion to generate diverse and highly realistic artistic stylized images. DSPA can also generate random artistic image samples with the learned style information, providing a new idea for data augmentation. Besides, a Key Content Feature Prompt (KCFP) module is proposed to provide sufficient content prompts for pre-trained stable diffusion to preserve the detailed structure of the input content image. Extensive qualitative and quantitative experiments verify the effectiveness of our proposed method.

Usage

Preparation

pip install -r requirements.txt

Training

Place your reference images in a directory, for example Artworks/monet_water-lilies-1914, then run the following:

You also can download the collection of artworks (https://drive.google.com/drive/folders/1_2jykbjVCF6SqJisvIt5-4fAFzVAj-F0?usp=drive_link)

  accelerate launch train.py \
  --train_data_dir=Artworks/monet_water-lilies-1914 \
  --output_dir=output

such as: nohup accelerate launch train.py --train_data_dir=Artworks/monet_water-lilies-1914 --output_dir=output-monet

A more comprehensive list of command arguments is shown in train.sh

You also can download our pre-trained model from https://pan.quark.cn/s/71aa26c4d241

Generate random style images

Assume your checkpoint is saved at output/final-1000.pt.

python generate.py \
  --weights_path=output/final-1000.pt \
  --output_dir=output_images \
  --pretrained_model_name_or_path="stabilityai/stable-diffusion-2-1" \
  --n_images=50 \
  --bsz=4

Generate with scaled standard deviation:

python generate.py \
  --weights_path=output/final-1000.pt \
  --output_dir=output_images \
  --pretrained_model_name_or_path="stabilityai/stable-diffusion-2-1" \
  --n_images=50 \
  --bsz=4 \
  --std_scale=2.0

python generate.py --weights_path=output-monet/final-50000.pt --output_dir=output_images-monet --pretrained_model_name_or_path="stabilityai/stable-diffusion-2-1" --n_images=50 --bsz=4 --customize_prefix="a painting of" --customize_suffix="Monet style"

Style transfer

CUDA_VISIBLE_DEVICES=1 python generate-control-file.py --weights_path=output/final-50000.pt --output_dir=output_images-monet --pretrained_model_name_or_path="/home/zzj/.cache/huggingface/diffusers/models--runwayml--stable-diffusion-v1-5/snapshots/39593d5650112b4cc580433f6b0435385882d819" --n_images=5 --bsz=4

Citation

@article{zhang2025dyartbank,
title={DyArtbank: Diverse artistic style transfer via pre-trained stable diffusion and dynamic style prompt Artbank},
author={Zhang, Zhanjie and Zhang, Quanwei and Li, Guangyuan and Luan, Junsheng and Yang, Mengyuan and Wang, Yun and Zhao, Lei},
journal={Knowledge-Based Systems},
volume={310},
pages={112959},
year={2025},
publisher={Elsevier}
}

Contact

Please feel free to open an issue or contact us personally if you have questions, need help, or need explanations. Write to one of the following email addresses, and maybe put one other in the cc:

cszzj@zju.edu.cn

Acknowledgement

We borrow code from Hugging Face diffusers, Dreamdistribution and Learning to Prompt for Vision-Language Models (CoOp). We thank the authors and the open source contributors for their work and contribution.

About

DyArtbank: Diverse artistic style transfer via pre-trained stable diffusion and dynamic style prompt Artbank (KBS 2025)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published