MLE-agent  by MLSysOps

Agent for AI engineering and research

Created 1 year ago
1,386 stars

Top 29.2% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

MLE-Agent is an AI-powered assistant designed to streamline AI engineering and research workflows for ML engineers and researchers. It automates baseline creation, code debugging, project organization, and even end-to-end Kaggle competition participation, integrating with academic resources like Arxiv and Papers with Code to leverage state-of-the-art methods.

How It Works

MLE-Agent functions as a multi-agent system, orchestrating various LLMs (OpenAI, Anthropic, Gemini, Ollama) to tackle complex ML tasks. Its core design emphasizes autonomous operation, allowing it to generate ML baselines, debug code through iterative coder-agent interaction, and manage project file structures. The integration with Arxiv and Papers with Code enables it to access and apply cutting-edge research and best practices directly within the workflow.

Quick Start & Requirements

  • Install: pip install mle-agent -U
  • Prerequisites: Python 3.8+, OpenAI API key (for certain models).
  • Usage: Navigate to your project directory and run mle start, mle chat, mle report, or mle kaggle.
  • Docs: https://mle-agent-site.vercel.app/

Highlighted Details

  • Autonomous Kaggle competition completion with mle kaggle --auto.
  • Weekly report generation via CLI or web UI using Git history.
  • Supports multiple LLM providers including OpenAI, Anthropic, Gemini, and Ollama.
  • Integrates Arxiv and Papers with Code for research-informed development.

Maintenance & Community

  • Active development with recent releases (v0.4.2 in Sept 2024).
  • Community support via Discord.
  • Roadmap: Includes planned integrations with Hugging Face, SkyPilot, Snowflake, AWS S3, Databricks, Wandb, MLflow, and DBT.

Licensing & Compatibility

  • MIT License. Permissive for commercial use and integration with closed-source projects.

Limitations & Caveats

  • While supporting multiple LLMs, optimal performance may depend on specific model configurations and API access.
  • Cloud platform integrations (AWS, Databricks, etc.) are listed as future roadmap items, indicating current limitations in cloud-native MLOps orchestration.
Health Check
Last Commit

1 month ago

Responsiveness

1 day

Pull Requests (30d)
1
Issues (30d)
1
Star History
33 stars in the last 30 days

Explore Similar Projects

Starred by Luis Capelo Luis Capelo(Cofounder of Lightning AI), Addy Osmani Addy Osmani(Head of Chrome Developer Experience at Google), and
23 more.

goose by block

0.8%
20k
Open-source AI agent for automating complex engineering tasks
Created 1 year ago
Updated 21 hours ago
Feedback? Help us improve.