cifar10-airbench  by KellerJordan

Fast CIFAR-10 training benchmarks

Created 1 year ago
295 stars

Top 89.8% on SourcePulse

GitHubView on GitHub
Project Summary

This repository provides highly optimized PyTorch scripts for training neural networks on the CIFAR-10 dataset, achieving state-of-the-art speed benchmarks. It targets researchers and practitioners seeking to establish fast, reproducible baselines for image classification tasks, offering significant speedups over standard training methods.

How It Works

The project leverages custom optimizations, including the Muon optimizer and data filtering techniques, to drastically reduce training time. These methods are designed to accelerate convergence without sacrificing accuracy, making them suitable for rapid experimentation and baseline establishment. The core advantage lies in the aggressive optimization of the training loop and data loading pipeline for maximum GPU utilization.

Quick Start & Requirements

  • Install: git clone https://github.com/KellerJordan/cifar10-airbench.git && cd airbench && python airbench94_muon.py
  • Requirements: PyTorch (torch), Torchvision (torchvision).
  • Hardware: NVIDIA A100 GPU recommended for achieving stated benchmarks.
  • Documentation: Official Quickstart

Highlighted Details

  • Achieves 94% accuracy on CIFAR-10 in 2.6 seconds and 96% in 27 seconds on an NVIDIA A100.
  • Significantly faster than standard ResNet-18 training (7 minutes for 96%).
  • Includes a GPU-accelerated dataloader for custom experiments.
  • Demonstrates data-selection strategies for improved training signal.

Maintenance & Community

The project appears to be a personal research effort by Keller Jordan. No specific community channels or roadmap are indicated in the README.

Licensing & Compatibility

The repository does not explicitly state a license. This is a significant omission for evaluating commercial use or integration into closed-source projects.

Limitations & Caveats

The primary limitation is the lack of a specified license, which hinders clear understanding of usage rights. The benchmarks are specific to NVIDIA A100 hardware, and achieving similar speeds on other GPUs may not be possible. The project is presented as a set of optimized scripts rather than a comprehensive library.

Health Check
Last Commit

2 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
13 stars in the last 30 days

Explore Similar Projects

Starred by Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera) and Stas Bekman Stas Bekman(Author of "Machine Learning Engineering Open Book"; Research Engineer at Snowflake).

InternEvo by InternLM

0.2%
407
Lightweight training framework for model pre-training
Created 1 year ago
Updated 4 weeks ago
Starred by Victor Taelin Victor Taelin(Author of Bend, Kind, HVM), Sebastian Raschka Sebastian Raschka(Author of "Build a Large Language Model (From Scratch)"), and
2 more.

nanoT5 by PiotrNawrot

0.2%
1k
PyTorch code for T5 pre-training and fine-tuning on a single GPU
Created 2 years ago
Updated 1 year ago
Starred by Benjamin Bolte Benjamin Bolte(Cofounder of K-Scale Labs), Albert Gu Albert Gu(Cofounder of Cartesia; Professor at CMU), and
2 more.

Muon by KellerJordan

1.7%
2k
Optimizer for neural network hidden layers
Created 10 months ago
Updated 2 months ago
Feedback? Help us improve.