Generate shell image using FastGAN
Model description
FastGAN model is a Generative Adversarial Networks (GAN) training on a small amount of high-fidelity images with minimum computing cost. Using a skip-layer channel-wise excitation module and a self-supervised discriminator trained as a feature-encoder, the model was able to converge after some hours of training for either 100 high-quality images or 1000 images datasets.
This model was trained on a dataset of 64 high-quality images of Shells.
How to use
# Clone this model
git clone https://huggingface.co/huggan/fastgan-few-shot-shells/
def load_generator(model_name_or_path):
generator = Generator(in_channels=256, out_channels=3)
generator = generator.from_pretrained(model_name_or_path, in_channels=256, out_channels=3)
_ = generator.eval()
return generator
def _denormalize(input: torch.Tensor) -> torch.Tensor:
return (input * 127.5) + 127.5
# Load generator
generator = load_generator("huggan/fastgan-few-shot-shells")
# Generate a random noise image
noise = torch.zeros(1, 256, 1, 1, device=device).normal_(0.0, 1.0)
with torch.no_grad():
gan_images, _ = generator(noise)
gan_images = _denormalize(gan_images.detach())
save_image(gan_images, "sample.png", nrow=1, normalize=True)
Limitations and bias
- Converge faster and better with small datasets (less than 1000 samples)
Training data
Generated Images
BibTeX entry and citation info
@article{FastGAN,
title={Towards Faster and Stabilized GAN Training for High-fidelity Few-shot Image Synthesis},
author={Bingchen Liu, Yizhe Zhu, Kunpeng Song, Ahmed Elgammal},
journal={ICLR},
year={2021}
}
Unable to determine this model's library. Check the
docs
.