Discover and explore top open-source AI tools and projects—updated daily.
orpatashnikText-driven StyleGAN imagery manipulation via CLIP models
Top 12.0% on SourcePulse
StyleCLIP enables text-driven manipulation of StyleGAN-generated imagery by leveraging CLIP's visual-language understanding. It offers three methods for users to edit images based on textual descriptions: latent vector optimization, a trained latent mapper, and global directions in StyleSpace, providing flexible control over image generation and modification.
How It Works
StyleCLIP integrates StyleGAN's generative capabilities with CLIP's text-image alignment. The latent vector optimization method uses a CLIP-based loss to adjust latent vectors according to text prompts. The latent mapper learns to infer text-guided latent vector residuals for faster, more stable edits. Global directions identify input-agnostic manipulations in StyleGAN's style space, allowing interactive, text-driven adjustments.
Quick Start & Requirements
pip install ftfy regex tqdm gdown git+https://github.com/openai/CLIP.gitcudatoolkit=<CUDA_VERSION>tensorflow-gpu==1.14)Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
2 years ago
Inactive
cloneofsimo
bloc97
nerdyrodent
Shengjia Zhao(Chief Scientist at Meta Superintelligence Lab),
openai