Project Name | Stars | Downloads | Repos Using This | Packages Using This | Most Recent Commit | Total Releases | Latest Release | Open Issues | License | Language |
---|---|---|---|---|---|---|---|---|---|---|
Ray | 24,783 | 80 | 199 | 20 hours ago | 76 | June 09, 2022 | 2,897 | apache-2.0 | Python | |
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a toolkit of libraries (Ray AIR) for accelerating ML workloads. | ||||||||||
Best Of Ml Python | 13,088 | 6 days ago | 15 | cc-by-sa-4.0 | ||||||
🏆 A ranked list of awesome machine learning Python libraries. Updated weekly. | ||||||||||
Nni | 12,653 | 8 | 22 | a day ago | 51 | June 22, 2022 | 285 | mit | Python | |
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning. | ||||||||||
Tpot | 8,989 | 40 | 18 | 5 days ago | 60 | January 06, 2021 | 284 | lgpl-3.0 | Python | |
A Python Automated Machine Learning tool that optimizes machine learning pipelines using genetic programming. | ||||||||||
Featuretools | 6,574 | 35 | 29 | 20 hours ago | 89 | July 05, 2022 | 171 | bsd-3-clause | Python | |
An open source python library for automated feature engineering | ||||||||||
H2o 3 | 6,194 | 18 | 30 | 21 hours ago | 232 | September 19, 2022 | 211 | apache-2.0 | Jupyter Notebook | |
H2O is an Open Source, Distributed, Fast & Scalable Machine Learning Platform: Deep Learning, Gradient Boosting (GBM) & XGBoost, Random Forest, Generalized Linear Modeling (GLM with Elastic Net), K-Means, PCA, Generalized Additive Models (GAM), RuleFit, Support Vector Machine (SVM), Stacked Ensembles, Automatic Machine Learning (AutoML), etc. | ||||||||||
Autogluon | 5,517 | a day ago | 219 | apache-2.0 | Python | |||||
AutoGluon: AutoML for Image, Text, Time Series, and Tabular Data | ||||||||||
Igel | 2,964 | a year ago | 34 | November 19, 2021 | 5 | mit | Python | |||
a delightful machine learning tool that allows you to train, test, and use models without writing code | ||||||||||
Zenml | 2,760 | 21 hours ago | 23 | apache-2.0 | Python | |||||
ZenML 🙏: Build portable, production-ready MLOps pipelines. https://zenml.io. | ||||||||||
Mljar Supervised | 2,390 | 2 | 3 months ago | 77 | March 02, 2022 | 118 | mit | Python | ||
Python package for AutoML on Tabular Data with Feature Engineering, Hyper-Parameters Tuning, Explanations and Automatic Documentation |
Install Instructions | Documentation (Stable | Latest)
AutoGluon automates machine learning tasks enabling you to easily achieve strong predictive performance in your applications. With just a few lines of code, you can train and deploy high-accuracy machine learning and deep learning models on image, text, time series, and tabular data.
# First install package from terminal:
# pip install -U pip
# pip install -U setuptools wheel
# pip install autogluon # autogluon==0.7.0
from autogluon.tabular import TabularDataset, TabularPredictor
train_data = TabularDataset('https://autogluon.s3.amazonaws.com/datasets/Inc/train.csv')
test_data = TabularDataset('https://autogluon.s3.amazonaws.com/datasets/Inc/test.csv')
predictor = TabularPredictor(label='class').fit(train_data, time_limit=120) # Fit models for 120s
leaderboard = predictor.leaderboard(test_data)
AutoGluon Task | Quickstart | API |
---|---|---|
TabularPredictor | ||
MultiModalPredictor | ||
TimeSeriesPredictor |
See the AutoGluon Website for documentation and instructions on:
fit()
with argument presets='best_quality'
).Refer to the AutoGluon Roadmap for details on upcoming features and releases.
We are actively accepting code contributions to the AutoGluon project. If you are interested in contributing to AutoGluon, please read the Contributing Guide to get started.
If you use AutoGluon in a scientific publication, please cite the following paper:
Erickson, Nick, et al. "AutoGluon-Tabular: Robust and Accurate AutoML for Structured Data." arXiv preprint arXiv:2003.06505 (2020).
BibTeX entry:
@article{agtabular,
title={AutoGluon-Tabular: Robust and Accurate AutoML for Structured Data},
author={Erickson, Nick and Mueller, Jonas and Shirkov, Alexander and Zhang, Hang and Larroy, Pedro and Li, Mu and Smola, Alexander},
journal={arXiv preprint arXiv:2003.06505},
year={2020}
}
If you are using AutoGluon Tabular's model distillation functionality, please cite the following paper:
Fakoor, Rasool, et al. "Fast, Accurate, and Simple Models for Tabular Data via Augmented Distillation." Advances in Neural Information Processing Systems 33 (2020).
BibTeX entry:
@article{agtabulardistill,
title={Fast, Accurate, and Simple Models for Tabular Data via Augmented Distillation},
author={Fakoor, Rasool and Mueller, Jonas W and Erickson, Nick and Chaudhari, Pratik and Smola, Alexander J},
journal={Advances in Neural Information Processing Systems},
volume={33},
year={2020}
}
If you use AutoGluon's multimodal text+tabular functionality in a scientific publication, please cite the following paper:
Shi, Xingjian, et al. "Multimodal AutoML on Structured Tables with Text Fields." 8th ICML Workshop on Automated Machine Learning (AutoML). 2021.
BibTeX entry:
@inproceedings{agmultimodaltext,
title={Multimodal AutoML on Structured Tables with Text Fields},
author={Shi, Xingjian and Mueller, Jonas and Erickson, Nick and Li, Mu and Smola, Alex},
booktitle={8th ICML Workshop on Automated Machine Learning (AutoML)},
year={2021}
}
AutoGluon's state-of-the-art tools for hyperparameter optimization, such as ASHA, Hyperband, Bayesian Optimization and BOHB have moved to the stand-alone package syne-tune.
To learn more, checkout our paper "Model-based Asynchronous Hyperparameter and Neural Architecture Search" arXiv preprint arXiv:2003.10865 (2020).
@article{abohb,
title={Model-based Asynchronous Hyperparameter and Neural Architecture Search},
author={Klein, Aaron and Tiao, Louis and Lienart, Thibaut and Archambeau, Cedric and Seeger, Matthias},
journal={arXiv preprint arXiv:2003.10865},
year={2020}
}
This library is licensed under the Apache 2.0 License.