Discover and explore top open-source AI tools and projects—updated daily.
Anime life simulation with next game state prediction
Top 83.4% on SourcePulse
AnimeGamer enables infinite anime life simulations by predicting and generating consistent multi-turn game states, including dynamic animation sequences and character attribute updates. It targets users interested in interactive storytelling and character-driven experiences, allowing them to direct anime characters through natural language commands.
How It Works
The system leverages Multimodal Large Language Models (MLLMs) to generate game states. It employs a three-phase training process: first, an encoder-decoder model with a diffusion-based decoder reconstructs videos conditioned on action intensity; second, an MLLM predicts subsequent game state representations from historical inputs; and third, the decoder is fine-tuned using the MLLM's predictions to enhance animation quality. This approach allows for consistent, context-aware video generation and character state evolution.
Quick Start & Requirements
conda create -n animegamer python==3.10
), activate it (conda activate animegamer
), and install requirements (pip install -r requirements.txt
)../checkpoints
directory.python app.py
. Requires at least 24GB VRAM per GPU for a two-GPU setup, or 60GB VRAM for a single GPU (set LOW_VRAM_VERSION = False
).python inference_MLLM.py
for state generation and python inference_Decoder.py
for animation decoding.Highlighted Details
Maintenance & Community
The project is associated with Tencent ARC Lab and City University of Hong Kong. Key components are based on CogvideoX and SEED-X. TODOs include releasing data processing and training codes, and weights for models trained on mixed anime datasets.
Licensing & Compatibility
The repository does not explicitly state a license in the README. The code is presented for research purposes, and commercial use would require clarification.
Limitations & Caveats
The project is in its early stages, with training codes and mixed-anime datasets not yet released. The Gradio demo has significant VRAM requirements (24GB per GPU or 60GB for single GPU).
4 months ago
Inactive