Discover and explore top open-source AI tools and projects—updated daily.
shaochenzeContinuous Autoregressive Language Models for efficient text generation
Top 49.7% on SourcePulse
CALM (Continuous Autoregressive Language Models) introduces a paradigm shift to overcome the token-by-token generation bottleneck in Large Language Models (LLMs). It enables predicting a single continuous vector representing an entire chunk of K tokens, significantly improving training and inference efficiency. This approach offers a novel scaling dimension for LLMs, termed "semantic bandwidth," benefiting researchers and practitioners seeking more efficient and scalable language models.
How It Works
CALM employs a two-stage process. First, a high-fidelity autoencoder compresses K tokens into a continuous vector and reconstructs them with near-perfect accuracy. Second, a continuous-domain language model performs autoregressive prediction in this vector space. This method reduces the number of autoregressive steps by a factor of K, leading to substantial efficiency gains and enabling scaling based on semantic bandwidth.
Quick Start & Requirements
git clone https://github.com/shaochenze/calm.git) and install dependencies (pip install -r requirements.txt).bash data/get_data.sh. Requires at least 2.5TB of free disk space.bash train/train_autoencoder.sh).bash train/train_energy.sh).train/train_diffusion.sh, train/train_flow.sh).train/train_ar.sh).bash train/eval_energy.sh to evaluate checkpoints.torchrun), sufficient disk space (2.5TB+), and multi-GPU setup (e.g., 8 GPUs per node). Scripts utilize bf16 for mixed-precision training.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 month ago
Inactive
karpathy
SafeAILab
dbiir
facebookresearch