{"id":14582876,"url":"https://github.com/MLSysOps/MLE-agent","last_synced_at":"2025-09-04T22:33:01.995Z","repository":{"id":242633172,"uuid":"787655686","full_name":"MLSysOps/MLE-agent","owner":"MLSysOps","description":"🤖 MLE-Agent: Your intelligent companion for seamless AI engineering and research. 🔍 Integrate with arxiv and paper with code to provide better code/research plans 🧰 OpenAI, Anthropic, Ollama, etc supported. :fireworks: Code RAG","archived":false,"fork":false,"pushed_at":"2024-10-23T17:19:58.000Z","size":2259,"stargazers_count":1075,"open_issues_count":10,"forks_count":45,"subscribers_count":8,"default_branch":"main","last_synced_at":"2024-10-24T00:58:53.332Z","etag":null,"topics":["agent","ai","llm","ml","mle","mlops"],"latest_commit_sha":null,"homepage":"https://mle-agent-site.vercel.app/","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/MLSysOps.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":"CONTRIBUTING.md","funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":".github/CODEOWNERS","security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2024-04-16T23:49:21.000Z","updated_at":"2024-10-23T17:20:02.000Z","dependencies_parsed_at":"2024-06-28T01:18:55.101Z","dependency_job_id":"8a915725-ca66-4a0f-9c30-c52cd953e5dc","html_url":"https://github.com/MLSysOps/MLE-agent","commit_stats":null,"previous_names":["mlsysops/mle-agent"],"tags_count":9,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/MLSysOps%2FMLE-agent","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/MLSysOps%2FMLE-agent/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/MLSysOps%2FMLE-agent/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/MLSysOps%2FMLE-agent/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/MLSysOps","download_url":"https://codeload.github.com/MLSysOps/MLE-agent/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":232003727,"owners_count":18458829,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["agent","ai","llm","ml","mle","mlops"],"created_at":"2024-09-08T01:01:16.740Z","updated_at":"2024-12-31T15:30:40.594Z","avatar_url":"https://github.com/MLSysOps.png","language":"Python","readme":"\u003cdiv align=\"center\"\u003e\n\u003ch1 align=\"center\"\u003eMLE-Agent: Your intelligent companion for seamless AI engineering and research.\u003c/h1\u003e\n\u003cimg alt=\"kaia-llama\" height=\"200px\" src=\"assets/kaia_llama.webp\"\u003e\n\u003ca href=\"https://trendshift.io/repositories/11658\" target=\"_blank\"\u003e\u003cimg src=\"https://trendshift.io/api/badge/repositories/11658\" alt=\"MLSysOps%2FMLE-agent | Trendshift\" style=\"width: 250px; height: 200px;\" width=\"250\" height=\"200px\"/\u003e\u003c/a\u003e\n\u003cp align=\"center\"\u003e:love_letter: Fathers' love for Kaia :love_letter:\u003c/p\u003e\n\n![](https://github.com/MLSysOps/MLE-agent/actions/workflows/lint.yml/badge.svg)\n![](https://github.com/MLSysOps/MLE-agent/actions/workflows/test.yml/badge.svg)\n![PyPI - Version](https://img.shields.io/pypi/v/mle-agent)\n[![Downloads](https://static.pepy.tech/badge/mle-agent)](https://pepy.tech/project/mle-agent)\n![GitHub License](https://img.shields.io/github/license/MLSysOps/MLE-agent)\n\u003ca href=\"https://discord.gg/d9vcY7PA8Z\"\u003e\u003cimg src=\"https://img.shields.io/badge/Discord-Join%20Us-purple?logo=discord\u0026logoColor=white\u0026style=flat\" alt=\"Join our Discord community\"\u003e\u003c/a\u003e\n\n[📚 Docs](https://mle-agent-site.vercel.app/) |\n[🐞 Report Issues](https://github.com/MLSysOps/MLE-agent/issues/new) |\n👋 Join us on \u003ca href=\"https://discord.gg/d9vcY7PA8Z\" target=\"_blank\"\u003eDiscord\u003c/a\u003e\n\n\u003c/div\u003e\n\n## Overview\n\nMLE-Agent is designed as a pairing LLM agent for machine learning engineers and researchers. It is featured by:\n\n- 🤖 Autonomous Baseline: Automatically builds ML/AI baselines and solutions based on your requirements.\n- 🏅End-to-end ML Task: Participates in Kaggle competitions and completes tasks independently.\n- 🔍 [Arxiv](https://arxiv.org/) and [Papers with Code](https://paperswithcode.com/) Integration: Access best practices\n  and state-of-the-art methods.\n- 🐛 Smart Debugging: Ensures high-quality code through automatic debugger-coder interactions.\n- 📂 File System Integration: Organizes your project structure efficiently.\n- 🧰 Comprehensive Tools Integration: Includes AI/ML functions and MLOps tools for a seamless workflow.\n- ☕ Interactive CLI Chat: Enhances your projects with an easy-to-use chat interface.\n- 🧠 Smart Advisor: Provides personalized suggestions and recommendations for your ML/AI project.\n- 📊 Weekly Report: Automatically generates detailed summaries of your weekly works.\n\nhttps://github.com/user-attachments/assets/dac7be90-c662-4d0d-8d3a-2bc4df9cffb9\n\n## Milestones\n\n- :rocket: 09/24/2024: Release the `0.4.2` with enhanced `Auto-Kaggle` mode to complete an end-to-end competition with minimal effort.\n- :rocket: 09/10/2024: Release the `0.4.0` with new CLIs like `MLE report`, `MLE kaggle`, `MLE integration` and many new\n  models like `Mistral`.\n- :rocket: 07/25/2024: Release the `0.3.0` with huge refactoring, many integrations, etc. (v0.3.0)\n- :rocket: 07/11/2024: Release the `0.2.0` with multiple agents interaction (v0.2.0)\n- 👨‍🍼 **07/03/2024: Kaia is born**\n- :rocket: 06/01/2024: Release the first rule-based version of MLE agent (v0.1.0)\n\n## Get started\n\n### Installation\n\n```bash\npip install mle-agent -U\n# or from source\ngit clone git@github.com:MLSysOps/MLE-agent.git\npip install -e .\n```\n\n### Usage\n\n```bash\nmle new \u003cproject name\u003e\n```\n\nAnd a project directory will be created under the current path, you need to start the project under the project\ndirectory.\n\n```bash\ncd \u003cproject name\u003e\nmle start\n```\n\nYou can also start an interactive chat in the terminal under the project directory:\n\n```bash\nmle chat\n```\n\n## Use cases\n\n### 🧪 Prototype an ML Baseline\n\nMLE agent can help you prototype an ML baseline with the given requirements, and test the model on the local machine.\nThe requirements can be vague, such as \"I want to predict the stock price based on the historical data\".\n\n```bash\ncd \u003cproject name\u003e\nmle start\n```\n\n### :bar_chart: Generate Work Report\n\nMLE agent can help you summarize your weekly report, including development progress, communication notes, reference, and\nto-do lists.\n\n#### Mode 1: Web Application to Generate Report from GitHub\n\n```bash\ncd \u003cproject name\u003e\nmle report\n```\n\nThen, you can visit http://localhost:3000/ to generate your report locally.\n\n#### Mode 2: CLI Tool to Generate Report from Local Git Repository\n```bash\ncd \u003cproject name\u003e\nmle report-local --email=\u003cgit email\u003e --start-date=YYYY-MM-DD --end-date=YYYY-MM-DD \u003cpath_to_git_repo\u003e\n```\n\n- `--start-date` and `--end-date` are optional parameters. If omitted, the command will generate a report for the default date range of the last 7 days.\n- Replace `\u003cgit email\u003e` with your Git email and `\u003cpath_to_git_repo\u003e` with the path to your local Git repository.\n\n### :trophy: Start with Kaggle Competition\n\nMLE agent can participate in Kaggle competitions and finish coding and debugging from data preparation to model training\nindependently. Here is the basic command to start a Kaggle competition:\n\n```bash\ncd \u003cproject name\u003e\nmle kaggle\n```\n\nOr you can let the agents finish the Kaggle task without human interaction if you have the dataset and submission file\nready:\n\n```bash\ncd \u003cproject name\u003e\nmle kaggle --auto \\\n--datasets \"\u003cpath_to_dataset1\u003e,\u003cpath_to_dataset2\u003e,...\" \\\n--description \"\u003cdescription_file_path_or_text\u003e\" \\\n--submission \"\u003csubmission_file_path\u003e\" \\\n--sub_example \"\u003csubmission_example_file_path\u003e\" \\ \n--comp_id \"\u003ccompetition_id\u003e\"\n```\n\nPlease make sure you have joined the competition before running the command. For more details, see the [MLE-Agent Tutorials](https://mle-agent-site.vercel.app/tutorial/Start_a_kaggle_task).\n\n## Roadmap\n\nThe following is a list of the tasks we plan to do, welcome to propose something new!\n\n\u003cdetails\u003e\n  \u003csummary\u003e\u003cb\u003e :hammer: General Features\u003c/b\u003e\u003c/summary\u003e\n\n- [x] Understand users' requirements to create an end-to-end AI project\n- [x] Suggest the SOTA data science solutions by using the web search\n- [x] Plan the ML engineering tasks with human interaction\n- [x] Execute the code on the local machine/cloud, debug and fix the errors\n- [x] Leverage the built-in functions to complete ML engineering tasks\n- [x] Interactive chat: A human-in-the-loop mode to help improve the existing ML projects\n- [x] Kaggle mode: to finish a Kaggle task without humans\n- [x] Summary and reflect the whole ML/AI pipeline\n- [ ] Integration with Cloud data and testing and debugging platforms\n- [x] Local RAG support to make personal ML/AI coding assistant\n- [ ] Function zoo: generate AI/ML functions and save them for future usage\n\n\u003c/details\u003e\n\n\u003cdetails\u003e\n  \u003csummary\u003e\u003cb\u003e:star: More LLMs and Serving Tools\u003c/b\u003e\u003c/summary\u003e\n\n- [x] Ollama LLama3\n- [x] OpenAI GPTs\n- [x] Anthropic Claude 3.5 Sonnet\n\n\u003c/details\u003e\n\n\u003cdetails\u003e\n  \u003csummary\u003e\u003cb\u003e:sparkling_heart: Better user experience\u003c/b\u003e\u003c/summary\u003e\n\n- [x] CLI Application\n- [x] Web UI\n- [x] Discord\n\n\u003c/details\u003e\n\n\u003cdetails\u003e\n  \u003csummary\u003e\u003cb\u003e:jigsaw: Functions and Integrations\u003c/b\u003e\u003c/summary\u003e\n\n- [x] Local file system\n- [x] Local code exectutor\n- [x] Arxiv.org search\n- [x] Papers with Code search\n- [x] General keyword search\n- [ ] Hugging Face\n- [ ] SkyPilot cloud deployment\n- [ ] Snowflake data\n- [ ] AWS S3 data\n- [ ] Databricks data catalog\n- [ ] Wandb experiment monitoring\n- [ ] MLflow management\n- [ ] DBT data transform\n\n\u003c/details\u003e\n\n## Contributing\n\nWe welcome contributions from the community. We are looking for contributors to help us with the following tasks:\n\n- Benchmark and Evaluate the agent\n- Add more features to the agent\n- Improve the documentation\n- Write tests\n\nPlease check the [CONTRIBUTING.md](CONTRIBUTING.md) file if you want to contribute.\n\n## Support and Community\n\n- [Discord community](https://discord.gg/SgxBpENGRG). If you have any questions, please ask in the Discord community.\n\n## Star History\n\n[![Star History Chart](https://api.star-history.com/svg?repos=MLSysOps/MLE-agent\u0026type=Date)](https://star-history.com/#MLSysOps/MLE-agent\u0026Date)\n\n## License\n\nCheck [MIT License](LICENSE) file for more information.\n","funding_links":[],"categories":["Python","Applications","A01_文本生成_文本对话","Repos","🔧 Projects"],"sub_categories":["Autonomous Agent Task Solver Projects","大语言对话模型及数据","Research Assessment"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2FMLSysOps%2FMLE-agent","html_url":"https://awesome.ecosyste.ms/projects/github.com%2FMLSysOps%2FMLE-agent","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2FMLSysOps%2FMLE-agent/lists"}