File size: 1,312 Bytes
653bf59 df5be5b 653bf59 4230602 df5be5b 4230602 df5be5b 6794c0f 645e5bd df5be5b 05b0c43 df5be5b 61e61fd 6794c0f 61e61fd |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 |
---
tags:
- text-to-image
- stable-diffusion
- dress
- try on
license: apache-2.0
language:
- en
library_name: diffusers
---
# IMAGDressing: Interactive Modular Apparel Generation for Dressing
## IMAGDressing-v1: Customizable Virtual Dressing
<div align="center">
[**Project Page**](https://imagdressing.github.io/) **|** [**Paper**](https://arxiv.org/pdf/2407.12705) **|** [**Code**](https://github.com/muzishen/IMAGDressing)**|** [**Data**](https://huggingface.co/datasets/IMAGDressing/IGPair)
</div>
---
## Introduction
To address the need for flexible and controllable customizations in virtual try-on systems, we propose IMAGDressing-v1. Specifically, we introduce a garment UNet that captures semantic features from CLIP and texture features from VAE. Our hybrid attention module includes a frozen self-attention and a trainable cross-attention, integrating these features into a frozen denoising UNet to ensure user-controlled editing. We will release a comprehensive dataset, IGv1, with over 200,000 pairs of clothing and dressed images, and establish a standard data assembly pipeline. Furthermore, IMAGDressing-v1 can be combined with extensions like ControlNet, IP-Adapter, T2I-Adapter, and AnimateDiff to enhance diversity and controllability.
![framework](assets/pipeline.png)
|