marc  by ekinakyurek

Research paper implementation for abstract reasoning via test-time training

Created 11 months ago
328 stars

Top 83.2% on SourcePulse

GitHubView on GitHub
Project Summary

This repository provides the official implementation for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning," focusing on applying Test-Time Training (TTT) to large language models for abstract reasoning tasks. It is intended for researchers and practitioners interested in advancing AI capabilities in complex problem-solving.

How It Works

The project leverages a modified version of the torchtune library for its Test-Time Training pipeline. It fine-tunes large language models (specifically Llama-3 variants) and then applies TTT to adapt these models to specific abstract reasoning tasks during inference. This approach aims to improve performance by allowing the model to learn from the test data distribution without explicit retraining.

Quick Start & Requirements

  • Installation: Requires cloning the repository recursively, setting up a conda environment (Python 3.10), installing a specific fork of torchtune in editable mode, and then installing other dependencies via pip install torch torchao --pre --upgrade --index-url https://download.pytorch.org/whl/nightly/cu121 and pip install -r requirements.txt.
  • Data: Requires downloading the ARC dataset from Kaggle.
  • Models: Pre-trained and fine-tuned models, as well as TTT checkpoints and LoRA adapters, are available on Hugging Face.
  • Hardware: Requires a CUDA-enabled GPU (cu121 specified).
  • Docs: https://github.com/ekinakyurek/marc

Highlighted Details

  • Official implementation of a paper demonstrating TTT effectiveness for abstract reasoning.
  • Supports Llama-3 and Llama-3.1/3.2 models with specific vLLM compatibility instructions.
  • Provides LoRA adapters for efficient fine-tuning and TTT.
  • Includes scripts for both TTT training and inference.

Maintenance & Community

The repository is marked as "in progress" with a caution to report errors. No specific community channels or roadmap are explicitly mentioned in the README.

Licensing & Compatibility

The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

The repository is explicitly stated to be in progress and should be used with caution. Some functionalities, like lora_to_output, may not apply to all model versions. Separate vLLM environments are required for different Llama versions due to compatibility issues.

Health Check
Last Commit

10 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
1 stars in the last 30 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Pawel Garbacki Pawel Garbacki(Cofounder of Fireworks AI), and
4 more.

alpaca_farm by tatsu-lab

0%
825
RLHF simulation framework for accessible instruction-following/alignment research
Created 2 years ago
Updated 1 year ago
Starred by Théophile Gervet Théophile Gervet(Cofounder of Genesis AI), Jason Knight Jason Knight(Director AI Compilers at NVIDIA; Cofounder of OctoML), and
7 more.

lingua by facebookresearch

0.0%
5k
LLM research codebase for training and inference
Created 1 year ago
Updated 2 months ago
Feedback? Help us improve.