Discover and explore top open-source AI tools and projects—updated daily.
facebookresearchResearch paper implementation for LLM reasoning in latent space
Top 30.4% on SourcePulse
This repository provides the official implementation for training Large Language Models (LLMs) to reason within a continuous latent space, addressing the challenge of structured reasoning in LLMs. It is intended for researchers and practitioners working on advancing LLM reasoning capabilities.
How It Works
The project introduces a novel training methodology that enables LLMs to learn and operate within a continuous latent space for reasoning. This approach aims to improve the structured and step-by-step reasoning abilities of LLMs by explicitly modeling intermediate reasoning steps. The core idea involves training stages that progressively refine the model's ability to generate and utilize continuous latent representations for problem-solving.
Quick Start & Requirements
pip install -r requirements.txt within a conda environment with Python 3.12.wandb for logging. Data should be in a specific JSON format.torchrun with multiple GPUs (e.g., 4x A100 80GB).preprocessing/gsm_icot.bashargs/ directoryHighlighted Details
Maintenance & Community
The project is from Meta AI (facebookresearch). No specific community links (Discord/Slack) or roadmap are provided in the README.
Licensing & Compatibility
Limitations & Caveats
The README implies significant computational resources (multiple high-end GPUs) are needed for reproducing experiments. The debug mode disables logging and model saving, potentially hindering analysis.
2 months ago
Inactive
txsun1997
jeffhj
openreasoner