Discover and explore top open-source AI tools and projects—updated daily.
jerberLLM-powered asynchronous solver for Abstraction and Reasoning Corpus (ARC) puzzles
Top 86.5% on SourcePulse
This project provides an asynchronous pipeline for solving Abstraction and Reasoning Corpus (ARC) puzzles using large language models (LLMs). It targets researchers and power users seeking to automate complex reasoning tasks by iteratively generating, scoring, and refining LLM-generated instructions. The system aims to produce candidate solutions for ARC challenges, facilitating participation in competitions and research.
How It Works
The core of the system is an asynchronous pipeline orchestrated by src/run.py. It processes ARC challenge datasets in batches, employing a monitored semaphore to manage API concurrency. For each step, LLMs generate instructions based on training grids, which are then scored via cross-validation using another LLM call to execute the instructions. The system iteratively revises poorly performing instructions or synthesizes new plans from the best-performing ones, feeding back into the scoring loop. Finally, the strongest instructions are used to generate multiple candidate outputs for hidden test grids. This iterative refinement and asynchronous execution are key to its approach.
Quick Start & Requirements
uv sync or pip install.MAX_CONCURRENCY environment variable. Optional: NEON_DSN for PostgreSQL persistence.python src/run.py for a basic smoke test using default configurations and the 2025 evaluation challenges. Custom configurations and paths can be used via run_from_json.Highlighted Details
asyncio with concurrency control via MonitoredSemaphore.RunConfig presets.Maintenance & Community
No specific details regarding maintainers, community channels (e.g., Discord, Slack), or project roadmap were found in the provided README text.
Licensing & Compatibility
The license type is not specified in the provided README text, which may pose a compatibility concern for certain use cases, particularly commercial applications.
Limitations & Caveats
The system requires careful environment variable configuration, including API keys and MAX_CONCURRENCY. Visualization features (VIZ=1) may cause issues on headless servers. Repeated rate-limiting errors suggest adjusting concurrency levels. The lack of explicit licensing information is a notable caveat.
1 month ago
Inactive
allenai
lupantech
WecoAI