--- license: apache-2.0 datasets: - yuvalkirstain/pickapic_v2 language: - en pipeline_tag: text-to-image --- **Self-Play Fine-Tuning of Diffusion Models for Text-to-Image Generation** (https://huggingface.co/papers/2402.10210) ![image/png](https://cdn-uploads.huggingface.co/production/uploads/657be24e8d360b690d5b665f/uzhoFO22ZdQ5XjBxxDA1a.png) # SPIN-Diffusion-iter1 This model is a self-play fine-tuned diffusion model at iteration 1 from [runwayml/stable-diffusion-v1-5](https://huggingface.co/runwayml/stable-diffusion-v1-5) using synthetic data based on the winner images of the [yuvalkirstain/pickapic_v2](https://huggingface.co/datasets/yuvalkirstain/pickapic_v2) dataset. We have also made a Gradio Demo at [UCLA-AGI/SPIN-Diffusion-demo-v1](https://huggingface.co/spaces/UCLA-AGI/SPIN-Diffusion-demo-v1). ## Model Details ### Model Description - Model type: An diffusion model with unet fine-tuned, based on the strucure of stable diffusion 1.5 - Language(s) (NLP): Primarily English - License: Apache-2.0 - Finetuned from model: runwayml/stable-diffusion-v1-5 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.0e-05 - train_batch_size: 8 - distributed_type: multi-GPU - num_devices: 8 - train_gradient_accumulation_steps: 32 - total_train_batch_size: 2048 - optimizer: AdamW - lr_scheduler: "linear" - lr_warmup_steps: 200 - num_training_steps: 500 ## Citation ``` @misc{yuan2024self, title={Self-Play Fine-Tuning of Diffusion Models for Text-to-Image Generation}, author={Yuan, Huizhuo and Chen, Zixiang and Ji, Kaixuan and Gu, Quanquan}, year={2024}, eprint={2402.10210}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```