Spaces:
Running
on
Zero
Running
on
Zero
title: Artistic Portrait Generation | |
emoji: 🎨 | |
colorFrom: yellow | |
colorTo: gray | |
sdk: gradio | |
sdk_version: 5.22.0 | |
app_file: app.py | |
pinned: true | |
license: apache-2.0 | |
models: | |
- AisingioroHao0/IP-Adapter-Art | |
- guozinan/PuLID | |
- stabilityai/stable-diffusion-xl-refiner-1.0 | |
- xinsir/controlnet-openpose-sdxl-1.0 | |
# IP Adapter Art: | |
<a href='https://huggingface.co/AisingioroHao0/IP-Adapter-Art'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Model-blue'></a><a href=''><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Dataset-blue'></a> [](https://colab.research.google.com/drive/1kV7q3Gzr8GPG9cChdDQ5ncCx84TYjuu3?usp=sharing) | |
 | |
------ | |
## Introduction | |
IP Adapter Art is a specialized version that uses a professional style encoder. Its goal is to achieve style control through reference images in the text-to-image diffusion model and solve the problems of instability and incomplete stylization of existing methods. This is a preprint version, and more models and training data coming soon. | |
## How to use | |
[](https://colab.research.google.com/drive/1kV7q3Gzr8GPG9cChdDQ5ncCx84TYjuu3?usp=sharing) can be used to conduct experiments directly. | |
For local experiments, please refer to a [demo](https://github.com/aihao2000/IP-Adapter-Art/blob/main/artistic_portrait_gen.ipynb). | |
Local experiments require a basic torch environment and dependencies: | |
``` | |
conda create -n artadapter python=3.10 | |
conda activate artadapter | |
pip install -r requirements.txt | |
pip install git+https://github.com/openai/CLIP.git | |
pip install -e . | |
``` | |
## Comparison with Existing Style Control Methods in Diffusion Models | |
Evaluation using [StyleBench](https://github.com/open-mmlab/StyleShot) style images. Image quality is evaluated using [improved aesthetic predictor](https://github.com/christophschuhmann/improved-aesthetic-predictor) | |
| | CLIP Style Similarity | CSD Style Similarity | CLIP Text Alignment | Image Quality | Average | | |
| --------------------- | --------------------- | -------------------- | ------------------- | ------------- | --------- | | |
| DEADiff | 61.99 | 43.54 | 20.82 | 60.76 | 46.78 | | |
| StyleShot | 63.01 | 52.40 | 18.93 | 55.54 | 47.47 | | |
| Instant Style | 65.39 | 58.39 | 21.09 | 60.62 | 51.37 | | |
| **Art-Adapter(ours)** | **67.03** | **65.02** | 20.25 | **62.23** | **53.63** | | |
 | |
## Examples of Text-guided Stylized Generation | |
 | |
## Artistic Portrait Generation | |
### Pipeline | |
We built an artistic portrait generation pipeline using Art-Adapter, PuLID, and ControlNet. The structure is shown in the figure below. | |
 | |
### Examples | |
 | |
## Stylize ControlNet Parameter Visualization | |
 | |
## Citation | |
``` | |
@misc{ipadapterart, | |
author = {Hao Ai, Xiaosai Zhang}, | |
title = {IP Adapter Art}, | |
year = {2024}, | |
publisher = {GitHub}, | |
journal = {GitHub repository}, | |
howpublished = {\url{https://github.com/aihao2000/IP-Adapter-Art}} | |
} | |
``` | |
## Acknowledgements | |