PyTorch code for BigGAN image synthesis research paper
Top 16.8% on sourcepulse
This repository provides an unofficial PyTorch implementation of BigGAN, designed for researchers and practitioners aiming to train high-fidelity generative models for natural image synthesis. It offers a flexible and extensible codebase for experimenting with BigGAN architectures and training strategies, including gradient accumulation for simulating larger batch sizes on multi-GPU setups.
How It Works
The implementation leverages gradient accumulation to simulate large effective batch sizes, enabling training of BigGAN models with batch sizes up to 2048 on multi-GPU systems without requiring TPU pods. It uses PyTorch's built-in inception network for calculating FID and IS metrics, with notes on differences from TensorFlow implementations and guidance on using TensorFlow for official scores. The code is structured for extensibility, allowing easy modification of components like loss functions, architectures, and attention blocks.
Quick Start & Requirements
sh scripts/utils/prepare_data.sh
.--load_in_mem
argument requires ~96GB+ RAM for loading HDF5.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The provided pre-trained models were not trained with orthogonal regularization, which may affect amenability to truncation. Synchronized BatchNorm and mixed-precision training (Tensor Cores) are noted as desired features that are not yet implemented, with suggestions for using Apex. The BigGAN-deep scripts are considered untested.
2 years ago
Inactive