MLE-agent  by MLSysOps

Agent for AI engineering and research

created 1 year ago
1,328 stars

Top 30.9% on sourcepulse

GitHubView on GitHub
Project Summary

MLE-Agent is an AI-powered assistant designed to streamline AI engineering and research workflows for ML engineers and researchers. It automates baseline creation, code debugging, project organization, and even end-to-end Kaggle competition participation, integrating with academic resources like Arxiv and Papers with Code to leverage state-of-the-art methods.

How It Works

MLE-Agent functions as a multi-agent system, orchestrating various LLMs (OpenAI, Anthropic, Gemini, Ollama) to tackle complex ML tasks. Its core design emphasizes autonomous operation, allowing it to generate ML baselines, debug code through iterative coder-agent interaction, and manage project file structures. The integration with Arxiv and Papers with Code enables it to access and apply cutting-edge research and best practices directly within the workflow.

Quick Start & Requirements

  • Install: pip install mle-agent -U
  • Prerequisites: Python 3.8+, OpenAI API key (for certain models).
  • Usage: Navigate to your project directory and run mle start, mle chat, mle report, or mle kaggle.
  • Docs: https://mle-agent-site.vercel.app/

Highlighted Details

  • Autonomous Kaggle competition completion with mle kaggle --auto.
  • Weekly report generation via CLI or web UI using Git history.
  • Supports multiple LLM providers including OpenAI, Anthropic, Gemini, and Ollama.
  • Integrates Arxiv and Papers with Code for research-informed development.

Maintenance & Community

  • Active development with recent releases (v0.4.2 in Sept 2024).
  • Community support via Discord.
  • Roadmap: Includes planned integrations with Hugging Face, SkyPilot, Snowflake, AWS S3, Databricks, Wandb, MLflow, and DBT.

Licensing & Compatibility

  • MIT License. Permissive for commercial use and integration with closed-source projects.

Limitations & Caveats

  • While supporting multiple LLMs, optimal performance may depend on specific model configurations and API access.
  • Cloud platform integrations (AWS, Databricks, etc.) are listed as future roadmap items, indicating current limitations in cloud-native MLOps orchestration.
Health Check
Last commit

6 days ago

Responsiveness

1 day

Pull Requests (30d)
16
Issues (30d)
10
Star History
62 stars in the last 90 days

Explore Similar Projects

Feedback? Help us improve.