Discover and explore top open-source AI tools and projects—updated daily.
Open-weights language model based on the Griffin architecture
Top 51.2% on SourcePulse
RecurrentGemma provides open-weights language models based on Google DeepMind's Griffin architecture, designed for efficient long-sequence generation through a hybrid attention-recurrence mechanism. It targets researchers and developers needing high-performance LLMs for tasks involving extended text, offering optimized Flax and reference PyTorch implementations.
How It Works
The Griffin architecture replaces global attention with a combination of local attention and linear recurrences. This approach significantly speeds up inference for long sequences by reducing the computational complexity associated with traditional self-attention mechanisms, making it more efficient for generating lengthy outputs.
Quick Start & Requirements
poetry install -E full
) or pip (pip install .[full]
). Library-specific installs are available (-E jax
, -E torch
, -E test
).python examples/sampling_jax.py --path_checkpoint=/path/to/weights --path_tokenizer=/path/to/tokenizer.model
Highlighted Details
Maintenance & Community
CONTRIBUTING.md
.Licensing & Compatibility
Limitations & Caveats
3 months ago
1 day