Discover and explore top open-source AI tools and projects—updated daily.
trevin-creatorAutonomous AI research loops for Apple Silicon
Top 30.3% on SourcePulse
This project ports Andrej Karpathy's autonomous AI research loops to Apple Silicon using the MLX framework, eliminating the need for PyTorch or CUDA. It targets researchers and power users seeking automated, hardware-optimized model configuration discovery within strict time constraints, enabling efficient on-device AI experimentation.
How It Works
The system preserves Karpathy's core design: fixed 5-minute experiment cycles, a single mutable train.py script, a primary val_bpb metric, and Git for experiment management (commit/revert). By leveraging MLX, it achieves native execution on Apple Silicon, utilizing unified memory for efficient computation without external dependencies like PyTorch or CUDA. This approach facilitates rapid, iterative AI research directly on Mac hardware.
Quick Start & Requirements
uv package manager.curl -LsSf https://astral.sh/uv/install.sh | sh
uv sync
uv run prepare.py # Data prep, tokenizer (one-time)
uv run train.py # Run a single experiment (~7 min cycle)
program.md.Highlighted Details
Maintenance & Community
The project acknowledges Andrej Karpathy for the core concept and references related MLX projects. No specific community channels (e.g., Discord, Slack) or detailed maintenance information are provided in the README.
Licensing & Compatibility
The project is released under the MIT License, preserving original copyright. This license generally permits commercial use and integration into closed-source projects.
Limitations & Caveats
Relies exclusively on MLX, which may have different performance characteristics than CUDA/PyTorch. MFU reporting is a placeholder; direct Apple Silicon FLOPs benchmarks are not equivalent to H100 standards. The evaluation token budget is reduced for faster iteration, potentially impacting full-scale evaluation. Muon optimizer's effectiveness is hardware-dependent.
1 month ago
Inactive