marc  by ekinakyurek

Research paper implementation for abstract reasoning via test-time training

created 9 months ago
324 stars

Top 83.8% on SourcePulse

GitHubView on GitHub
Project Summary

This repository provides the official implementation for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning," focusing on applying Test-Time Training (TTT) to large language models for abstract reasoning tasks. It is intended for researchers and practitioners interested in advancing AI capabilities in complex problem-solving.

How It Works

The project leverages a modified version of the torchtune library for its Test-Time Training pipeline. It fine-tunes large language models (specifically Llama-3 variants) and then applies TTT to adapt these models to specific abstract reasoning tasks during inference. This approach aims to improve performance by allowing the model to learn from the test data distribution without explicit retraining.

Quick Start & Requirements

  • Installation: Requires cloning the repository recursively, setting up a conda environment (Python 3.10), installing a specific fork of torchtune in editable mode, and then installing other dependencies via pip install torch torchao --pre --upgrade --index-url https://download.pytorch.org/whl/nightly/cu121 and pip install -r requirements.txt.
  • Data: Requires downloading the ARC dataset from Kaggle.
  • Models: Pre-trained and fine-tuned models, as well as TTT checkpoints and LoRA adapters, are available on Hugging Face.
  • Hardware: Requires a CUDA-enabled GPU (cu121 specified).
  • Docs: https://github.com/ekinakyurek/marc

Highlighted Details

  • Official implementation of a paper demonstrating TTT effectiveness for abstract reasoning.
  • Supports Llama-3 and Llama-3.1/3.2 models with specific vLLM compatibility instructions.
  • Provides LoRA adapters for efficient fine-tuning and TTT.
  • Includes scripts for both TTT training and inference.

Maintenance & Community

The repository is marked as "in progress" with a caution to report errors. No specific community channels or roadmap are explicitly mentioned in the README.

Licensing & Compatibility

The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

The repository is explicitly stated to be in progress and should be used with caution. Some functionalities, like lora_to_output, may not apply to all model versions. Separate vLLM environments are required for different Llama versions due to compatibility issues.

Health Check
Last commit

8 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
2
Star History
6 stars in the last 30 days

Explore Similar Projects

Starred by Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), Shizhe Diao Shizhe Diao(Research Scientist at NVIDIA; Author of LMFlow), and
4 more.

simpleRL-reason by hkust-nlp

0.4%
4k
RL recipe for reasoning ability in models
created 6 months ago
updated 1 week ago
Starred by Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), Ross Taylor Ross Taylor(Cofounder of General Reasoning; Creator of Papers with Code), and
9 more.

open-instruct by allenai

0.6%
3k
Training codebase for instruction-following language models
created 2 years ago
updated 19 hours ago
Starred by Junyang Lin Junyang Lin(Core Maintainer of Alibaba Qwen), Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), and
16 more.

alpaca-lora by tloen

0.1%
19k
LoRA fine-tuning for LLaMA
created 2 years ago
updated 1 year ago
Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera), and
16 more.

open-r1 by huggingface

0.3%
25k
SDK for reproducing DeepSeek-R1
created 6 months ago
updated 4 days ago
Feedback? Help us improve.