Discover and explore top open-source AI tools and projects—updated daily.
nathan-barrySmall diffusion model for character-level text generation
Top 55.2% on SourcePulse
A character-level language diffusion model for text generation, tiny-diffusion offers a compact, 10.7 million parameter implementation. It makes diffusion models accessible for local experimentation by engineers and researchers, enabling exploration of text generation without substantial computational resources.
How It Works
The model is a modified version of the nanochat GPT architecture, adapted for character-level diffusion. It processes text sequences up to 256 characters long. This approach leverages diffusion principles, typically used in image generation, for generative tasks within a significantly reduced parameter footprint. This makes it distinct from larger, more resource-intensive language models and enables local experimentation.
Quick Start & Requirements
uv sync (requires Python 3.10+).uv run sample.py to generate text using pre-trained weights.uv run training.py.uv run animations/diffusion-process.py.Highlighted Details
weights/diffusion_model.pt).Maintenance & Community
No specific details regarding contributors, community channels, or roadmap were provided in the README snippet.
Licensing & Compatibility
Limitations & Caveats
The model operates strictly at a character level, which may impact the coherence, grammatical correctness, and linguistic nuance compared to token-based or word-based language models. It is trained exclusively on the "Tiny Shakespeare" dataset, inherently limiting its generative domain to the style and vocabulary present in that specific corpus. The current text generation context length is capped at 30 characters, potentially restricting the flow of longer outputs.
2 weeks ago
Inactive
dome272
sharonzhou
GuyTevet