PyTorch implementation for text-guided image style transfer
Top 86.1% on sourcepulse
CLIPstyler provides the official PyTorch implementation for a CVPR 2022 paper on image style transfer using a single text condition. It enables users to apply artistic styles to images based on textual descriptions, offering a novel approach to text-guided image manipulation for researchers and artists.
How It Works
The core of CLIPstyler leverages CLIP (Contrastive Language–Image Pre-training) to bridge the gap between text and image domains. It uses a style transfer network that is conditioned on a text embedding, allowing for flexible and precise style application. This approach avoids the need for paired text-image data for training specific styles, relying instead on CLIP's general understanding of visual concepts and their textual representations.
Quick Start & Requirements
conda create -n CLIPstyler python=3.6
, conda install --yes -c pytorch pytorch=1.7.1 torchvision cudatoolkit=11.0
, pip install ftfy regex tqdm
, conda install -c anaconda git
, pip install git+https://github.com/openai/CLIP.git
Highlighted Details
Maintenance & Community
The project is associated with Gihyun Kwon and Jong Chul Ye. Links to the paper and citation details are provided. No specific community channels (Discord/Slack) or roadmap are mentioned in the README.
Licensing & Compatibility
The README does not explicitly state a license. The code is presented as the "Official Pytorch implementation," implying it may be tied to the research paper's terms. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The environment setup requires specific older versions of PyTorch (1.7.1) and Python (3.6), which may pose compatibility challenges with newer systems. Colab demonstrations are noted to have slow computation speeds. The fast style transfer requires downloading large datasets and pre-trained models.
3 years ago
Inactive