Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/innixma/autogluon-benchmarking
Benchmarking Utilities for AutoGluon
https://github.com/innixma/autogluon-benchmarking
Last synced: about 1 month ago
JSON representation
Benchmarking Utilities for AutoGluon
- Host: GitHub
- URL: https://github.com/innixma/autogluon-benchmarking
- Owner: Innixma
- License: apache-2.0
- Created: 2020-02-13T23:34:44.000Z (almost 5 years ago)
- Default Branch: master
- Last Pushed: 2022-01-11T23:38:20.000Z (almost 3 years ago)
- Last Synced: 2024-11-30T01:35:06.457Z (about 1 month ago)
- Language: Python
- Size: 664 KB
- Stars: 20
- Watchers: 6
- Forks: 5
- Open Issues: 2
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# WARNING: THIS CODE IS OUTDATED. IT IS NOT RECOMMENDED TO TRY RUNNING THIS CODE.
This code is from early 2020 and due to dependency version mismatches such as sklearn between the AutoML frameworks in the old versions, it is unlikely to work properly in 2022+.
Instead, please refer to the AutoMLBenchmark repo to run AutoGluon and other AutoML frameworks on OpenML: https://github.com/openml/automlbenchmark
A setup guide for running AutoGluon is available here: https://github.com/Innixma/autogluon-benchmark/blob/master/examples/automlbenchmark/README_automlbenchmark.md
We are working on an updated repo to run the Kaggle competitions, as it unfortunately is quite involved due to a lack of Kaggle API's for fetching leaderboard scores and the lack of other AutoML frameworks compatibility with raw data in Kaggle requiring special preprocessing.
# Original Description
autogluon_utils is a package containing code for benchmarking [AutoGluon](https://autogluon.mxnet.io/) and other AutoML frameworks, to accompany the following paper:
[AutoGluon-Tabular: Robust and Accurate AutoML for Structured Data](https://arxiv.org/abs/2003.06505)
To get started, first install this autogluon_utils package using:
**python setup.py develop**
To run the Kaggle benchmark, see instructions in:
**autogluon_utils/benchmarking/kaggle/README-kagglebenchmark.txt**
To run the AutoML benchmark, see instructions in:
**autogluon_utils/benchmarking/openml/README-automlbenchmark.txt**
Code to run the other AutoML frameworks is in the directory:
**autogluon_utils/benchmarking/baselines/**
Files containing our raw benchmark results are available in the **data/results/input/** subdirectory. These results can be formatted as in the paper via code from the directory (**run_full.py** formats all results):
**autogluon_utils/benchmarking/evaluation/runners/**
The formatted results will be stored in: **data/results/output/**