Fast CIFAR-10 training benchmarks
Top 95.2% on sourcepulse
This repository provides highly optimized PyTorch scripts for training neural networks on the CIFAR-10 dataset, achieving state-of-the-art speed benchmarks. It targets researchers and practitioners seeking to establish fast, reproducible baselines for image classification tasks, offering significant speedups over standard training methods.
How It Works
The project leverages custom optimizations, including the Muon optimizer and data filtering techniques, to drastically reduce training time. These methods are designed to accelerate convergence without sacrificing accuracy, making them suitable for rapid experimentation and baseline establishment. The core advantage lies in the aggressive optimization of the training loop and data loading pipeline for maximum GPU utilization.
Quick Start & Requirements
git clone https://github.com/KellerJordan/cifar10-airbench.git && cd airbench && python airbench94_muon.py
torch
), Torchvision (torchvision
).Highlighted Details
Maintenance & Community
The project appears to be a personal research effort by Keller Jordan. No specific community channels or roadmap are indicated in the README.
Licensing & Compatibility
The repository does not explicitly state a license. This is a significant omission for evaluating commercial use or integration into closed-source projects.
Limitations & Caveats
The primary limitation is the lack of a specified license, which hinders clear understanding of usage rights. The benchmarks are specific to NVIDIA A100 hardware, and achieving similar speeds on other GPUs may not be possible. The project is presented as a set of optimized scripts rather than a comprehensive library.
2 weeks ago
1 day