Provide an input CSV and a target field to predict, generate a model + code to run it.

Overview

automl-gs

console gif

Give an input CSV file and a target field you want to predict to automl-gs, and get a trained high-performing machine learning or deep learning model plus native Python code pipelines allowing you to integrate that model into any prediction workflow. No black box: you can see exactly how the data is processed, how the model is constructed, and you can make tweaks as necessary.

demo output

automl-gs is an AutoML tool which, unlike Microsoft's NNI, Uber's Ludwig, and TPOT, offers a zero code/model definition interface to getting an optimized model and data transformation pipeline in multiple popular ML/DL frameworks, with minimal Python dependencies (pandas + scikit-learn + your framework of choice). automl-gs is designed for citizen data scientists and engineers without a deep statistical background under the philosophy that you don't need to know any modern data preprocessing and machine learning engineering techniques to create a powerful prediction workflow.

Nowadays, the cost of computing many different models and hyperparameters is much lower than the opportunity cost of an data scientist's time. automl-gs is a Python 3 module designed to abstract away the common approaches to transforming tabular data, architecting machine learning/deep learning models, and performing random hyperparameter searches to identify the best-performing model. This allows data scientists and researchers to better utilize their time on model performance optimization.

  • Generates native Python code; no platform lock-in, and no need to use automl-gs after the model script is created.
  • Train model configurations super-fast for free using a TPU and TensorFlow in Google Colaboratory. (in Beta: you can access the Colaboratory notebook here).
  • Handles messy datasets that normally require manual intervention, such as datetime/categorical encoding and spaced/parenthesized column names.
  • Each part of the generated model pipeline is its own function w/ docstrings, making it much easier to integrate into production workflows.
  • Extremely detailed metrics reporting for every trial stored in a tidy CSV, allowing you to identify and visualize model strengths and weaknesses.
  • Correct serialization of data pipeline encoders on disk (i.e. no pickled Python objects!)
  • Retrain the generated model on new data without making any code/pipeline changes.
  • Quit the hyperparameter search at any time, as the results are saved after each trial.
  • Training progress bars with ETAs for both the overall experiment and per-epoch during the experiment.

The models generated by automl-gs are intended to give a very strong baseline for solving a given problem; they're not the end-all-be-all that often accompanies the AutoML hype, but the resulting code is easily tweakable to improve from the baseline.

You can view the hyperparameters and their values here, and the metrics that can be optimized here. Some of the more controversial design decisions for the generated models are noted in DESIGN.md.

Framework Support

Currently automl-gs supports the generation of models for regression and classification problems using the following Python frameworks:

  • TensorFlow (via tf.keras) | tensorflow
  • XGBoost (w/ histogram binning) | xgboost

To be implemented:

  • Catboost | catboost
  • LightGBM | lightgbm

How to Use

automl-gs can be installed via pip:

pip3 install automl_gs

You will also need to install the corresponding ML/DL framework (e.g. tensorflow/tensorflow-gpu for TensorFlow, xgboost for xgboost, etc.)

After that, you can run it directly from the command line. For example, with the famous Titanic dataset:

automl_gs titanic.csv Survived

If you want to use a different framework or configure the training, you can do it with flags:

automl_gs titanic.csv Survived --framework xgboost --num_trials 1000

You may also invoke automl-gs directly from Python. (e.g. via a Jupyter Notebook)

from automl_gs import automl_grid_search

automl_grid_search('titanic.csv', 'Survived')

The output of the automl-gs training is:

  • A timestamped folder (e.g. automl_tensorflow_20190317_020434) with contains:
    • model.py: The generated model file.
    • pipeline.py: The generated pipeline file.
    • requirements.txt: The generated requirements file.
    • /encoders: A folder containing JSON-serialized encoder files
    • /metadata: A folder containing training statistics + other cool stuff not yet implemented!
    • The model itself (format depends on framework)
  • automl_results.csv: A CSV containing the training results after each epoch and the hyperparameters used to train at that time.

Once the training is done, you can run the generated files from the command line within the generated folder above.

To predict:

python3 model.py -d data.csv -m predict

To retrain the model on new data:

python3 model.py -d data.csv -m train

CLI Arguments/Function Parameters

You can view these at any time by running automl_gs -h in the command line.

  • csv_path: Path to the CSV file (must be in the current directory) [Required]
  • target_field: Target field to predict [Required]
  • target_metric: Target metric to optimize [Default: Automatically determined depending on problem type]
  • framework: Machine learning framework to use [Default: 'tensorflow']
  • model_name: Name of the model (if you want to train models with different names) [Default: 'automl']
  • num_trials: Number of trials / different hyperparameter combos to test. [Default: 100]
  • split: Train-validation split when training the models [Default: 0.7]
  • num_epochs: Number of epochs / passes through the data when training the models. [Default: 20]
  • col_types: Dictionary of fields:data types to use to override automl-gs's guesses. (only when using in Python) [Default: {}]
  • gpu: For non-Tensorflow frameworks and Pascal-or-later GPUs, boolean to determine whether to use GPU-optimized training methods (TensorFlow can detect it automatically) [Default: False]
  • tpu_address: For TensorFlow, hardware address of the TPU on the system. [Default: None]

Examples

For a quick Hello World on how to use automl-gs, see this Jupyter Notebook.

Due to the size of some examples w/ generated code and accompanying data visualizations, they are maintained in a separate repository. (and also explain why there are two distinct "levels" in the example viz above!)

How automl-gs Works

TL;DR: auto-ml gs generates raw Python code using Jinja templates and trains a model using the generated code in a subprocess: repeat using different hyperparameters until done and save the best model.

automl-gs loads a given CSV and infers the data type of each column to be fed into the model. Then it tries a ETL strategy for each column field as determined by the hyperparameters; for example, a Datetime field has its hour and dayofweek binary-encoded by default, but hyperparameters may dictate the encoding of month and year as additional model fields. ETL strategies are optimized for frameworks; TensorFlow for example will use text embeddings, while other frameworks will use CountVectorizers to encode text (when training, TensorFlow will also used a shared text encoder via Keras's functional API). automl-gs then creates a statistical model with the specified framework. Both the model ETL functions and model construction functions are saved as a generated Python script.

automl-gs then runs the generated training script as if it was a typical user. Once the model is trained, automl-gs saves the training results in its own CSV, along with all the hyperparameters used to train the model. automl-gs then repeats the task with another set of hyperparameters, until the specified number of trials is hit or the user kills the script.

The best model Python script is kept after each trial, which can then easily be integrated into other scripts, or run directly to get the prediction results on a new dataset.

Helpful Notes

  • It is the user's responsibility to ensure the input dataset is high-quality. No model hyperparameter search will provide good research on flawed/unbalanced datasets. Relatedly, hyperparameter optimization may provide optimistic predictions on the validation set, which may not necessarily match the model performance in the real world.
  • A neural network approach alone may not necessarily be the best approach. Try using xgboost. The results may surprise you!
  • automl-gs is only attempting to solve tabular data problems. If you have a more complicated problem to solve (e.g. predicting a sequence of outputs), I recommend using Microsoft's NNI and Uber's Ludwig as noted in the introduction.

Known Issues

  • Issues when using Anaconda (#8). Use an installed Python is possible.
  • Issues when using Windows (#13)
  • Issues when a field name in the input dataset starts with a number (#18)

Future Work

Feature development will continue on automl-gs as long as there is interest in the package.

Top Priority

  • Add more frameworks
  • Results visualization (via plotnine)
  • Holiday support for datetimes
  • Remove redundant generated code
  • Native distributed/high level automation support (Polyaxon/Kubernetes, Airflow)
  • Image field support (both as a CSV column field, and a special flow mode to take advantage of hyperparameter tuning)
  • PyTorch model code generation.

Elsework

  • Generate script given an explicit set of hyperparameters
  • More hyperparameters.
  • Bayesian hyperparameter search for standalone version.
  • Support for generating model code for R/Julia
  • Tool for generating a Flask/Starlette REST API from a trained model script
  • Allow passing an explicit, pre-defined test set CSV.

Maintainer/Creator

Max Woolf (@minimaxir)

Max's open-source projects are supported by his Patreon. If you found this project helpful, any monetary contributions to the Patreon are appreciated and will be put to good creative use.

License

MIT

The code generated by automl-gs is unlicensed; the owner of the generated code can decide the license.

Comments
  • YAMLLoadWarning disrupting progress bar

    YAMLLoadWarning disrupting progress bar

    Trying out the example titanic dataset in a conda environment and encountered the following error very frequently such that it disrupts the tqdm progress bar.

    /anaconda3/envs/automl-gs/lib/python3.6/site-packages/automl_gs/utils_automl.py:270:
    YAMLLoadWarning: calling yaml.load() without Loader=... is deprecated, as the default 
    Loader is unsafe. Please read https://msg.pyyaml.org/load for full details.
      metrics = yaml.load(f)
    
    opened by remykarem 3
  • fix #13

    fix #13

    This initial commit enables automl-gs to work on win10. I can test on win8 and linux, but don't have access to a Mac to make sure everything still works normally there.

    opened by evan-burke 3
  •  FileNotFoundError

    FileNotFoundError

    Hi,

    Just trying to work through your example colab notebook. I work through the cells, upload the titanic.csv, and get

    ---------------------------------------------------------------------------
    
    FileNotFoundError                         Traceback (most recent call last)
    
    <ipython-input-3-9f452c025bdd> in <module>()
          2                    target_field='origin',
          3                    model_name='tpu',
    ----> 4                    tpu_address = tpu_address)
    
    5 frames
    
    /usr/local/lib/python3.6/dist-packages/pandas/io/parsers.py in __init__(self, src, **kwds)
       1889         kwds["usecols"] = self.usecols
       1890 
    -> 1891         self._reader = parsers.TextReader(src, **kwds)
       1892         self.unnamed_cols = self._reader.unnamed_cols
       1893 
    
    pandas/_libs/parsers.pyx in pandas._libs.parsers.TextReader.__cinit__()
    
    pandas/_libs/parsers.pyx in pandas._libs.parsers.TextReader._setup_parser_source()
    
    FileNotFoundError: [Errno 2] File tpu_train/metadata/results.csv does not exist: 'tpu_train/metadata/results.csv'
    
    
    opened by shawngraham 1
  • xgboost: GPU support

    xgboost: GPU support

    xgboost supports GPUs by setting gpu_hist instead of hist, and the code is prepared for that. Two problems:

    1. Unlike TensorFlow, xgboost does not have a way to automatically determine if a GPU is present.
    2. GPU hist training requires a Pascal minimum; the GPUs in Colaboratory notebooks are K80s (Kepler) which does not qualify.

    Will keep at CPU support for now but there has to be a better solution.

    opened by minimaxir 1
  • Fixed some typos and grammar

    Fixed some typos and grammar

    Changes:

    1. parathesized ==> parenthesized
    2. JSON-seralized ==> JSON-serialized
    3. necessairly ==> necessarily (2)
    4. hyperameter ==> hyperparameter
    opened by mikeshatch 0
  • Ensure col_types works as expected

    Ensure col_types works as expected

    I started playing with a slightly messy file I had available, and noticed that the col_types argument didn't seem to work to allow me to ignore/override some columns. Found what seems to be the issue and it works for me locally.

    Thanks!

    opened by drien 0
  • SyntaxError: invalid decimal literal produced by automl

    SyntaxError: invalid decimal literal produced by automl

    Wanted to try automl_gs, but I get this error and can't figure out why.

    File "C:\Users\XXX\automl_train\model.py", line 3, in from pipeline import * File "C:\Users\XXX\automl_train\pipeline.py", line 29 0_enc = df['0'] ^ SyntaxError: invalid decimal literal

    Any ideas about that?

    opened by Haifischfutter 0
  • Google Colab - automl_train/metadata/results.csv does not exist

    Google Colab - automl_train/metadata/results.csv does not exist

    Input:

    from automl_gs import automl_grid_search
    
    automl_grid_search("data.csv", "diagnosis")
    

    Output:

    Solving a binary_classification problem, maximizing accuracy using tensorflow.
    
    Modeling with field specifications:
    id: ignore
    radius_mean: numeric
    texture_mean: numeric
    perimeter_mean: numeric
    area_mean: numeric
    smoothness_mean: numeric
    compactness_mean: numeric
    concavity_mean: numeric
    concave points_mean: numeric
    symmetry_mean: numeric
    fractal_dimension_mean: numeric
    radius_se: numeric
    texture_se: numeric
    perimeter_se: numeric
    area_se: numeric
    smoothness_se: numeric
    compactness_se: numeric
    concavity_se: numeric
    concave points_se: numeric
    symmetry_se: numeric
    fractal_dimension_se: numeric
    radius_worst: numeric
    texture_worst: numeric
    perimeter_worst: numeric
    area_worst: numeric
    smoothness_worst: numeric
    compactness_worst: numeric
    concavity_worst: numeric
    concave points_worst: numeric
    symmetry_worst: numeric
    fractal_dimension_worst: numeric
    Unnamed: 32: numeric
    0%
    0/100 [00:04<?, ?trial/s]
    0%
    0/20 [00:00<?, ?epoch/s]
    ---------------------------------------------------------------------------
    FileNotFoundError                         Traceback (most recent call last)
    <ipython-input-37-308e97508c91> in <module>()
          1 from automl_gs import automl_grid_search
          2 
    ----> 3 automl_grid_search("data.csv", "diagnosis")
    
    5 frames
    /usr/local/lib/python3.6/dist-packages/pandas/io/parsers.py in __init__(self, src, **kwds)
       1889         kwds["usecols"] = self.usecols
       1890 
    -> 1891         self._reader = parsers.TextReader(src, **kwds)
       1892         self.unnamed_cols = self._reader.unnamed_cols
       1893 
    
    pandas/_libs/parsers.pyx in pandas._libs.parsers.TextReader.__cinit__()
    
    pandas/_libs/parsers.pyx in pandas._libs.parsers.TextReader._setup_parser_source()
    
    FileNotFoundError: [Errno 2] File automl_train/metadata/results.csv does not exist: 'automl_train/metadata/results.csv'
    
    opened by batmanscode 3
  • Fix duplicates kwarg in pd.cut

    Fix duplicates kwarg in pd.cut

    From pandas version >= 0.23 you have to specify 'duplicates' kwarg to drop duplicates in 'bins' array. Otherwise, it will throw an error.

    https://pandas.pydata.org/pandas-docs/version/0.23.4/generated/pandas.cut.html

    opened by kitkatk 0
  • Using automl-gs for bin packing

    Using automl-gs for bin packing

    Is there any way to use a tool like automl-gs for bin-packing problems? I've seen a way to model it as linear optimization where you do a cross join of all objects and all potential bins, set a constraint of each object being selected once, and then optimizing the bins however desired. Cross joins can end up being needlessly heavy though, so I'm wondering if there is a way to model that sort of problem such that you could use pre-optimized and continually developed tools like this one.

    opened by jdotjdot 0
  • AttributeError: 'float' object has no attribute 'lower'

    AttributeError: 'float' object has no attribute 'lower'

    While trying out automl_gs with Jupiter, I got file not found error:

    FileNotFoundError: [Errno 2] File b'automl_train/metadata/results.csv' does not exist: b'automl_train/metadata/results.csv'

    Trying do it with terminal before the file missing error it returns:

    AttributeError: 'float' object has no attribute 'lower'

    Searching in StackOverflow, I found that the problem is how pandas converts inputs to python datatypes.

    https://stackoverflow.com/questions/34724246/attributeerror-float-object-has-no-attribute-lower/34724771

    Is it possible to prevent this behaviour using automl_gs ?

    opened by AlexandraMakarova 0
Releases(v0.2.1)
Owner
Max Woolf
Data Scientist @buzzfeed. Plotter of pretty charts.
Max Woolf
YouTube Spam Detection with python

YouTube Spam Detection This code deletes spam comment on youtube videos based on two characteristics (currently) If the author of the comment has a se

MohamadReza Taalebi 5 Sep 27, 2022
A repository of PyBullet utility functions for robotic motion planning, manipulation planning, and task and motion planning

pybullet-planning (previously ss-pybullet) A repository of PyBullet utility functions for robotic motion planning, manipulation planning, and task and

Caelan Garrett 260 Dec 27, 2022
Code for the TCAV ML interpretability project

Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) Been Kim, Martin Wattenberg, Justin Gilmer, C

552 Dec 27, 2022
Project to deploy a machine learning model based on Titanic dataset from Kaggle

kaggle_titanic_deploy Project to deploy a machine learning model based on Titanic dataset from Kaggle In this project we used the Titanic dataset from

Vivian Yamassaki 8 May 23, 2022
Machine Learning Study 혼자 해보기

Machine Learning Study 혼자 해보기 기여자 (Contributors) ✨ Teddy Lee 🏠 HongJaeKwon 🏠 Seungwoo Han 🏠 Tae Heon Kim 🏠 Steve Kwon 🏠 SW Song 🏠 K1A2 🏠 Wooil

Teddy Lee 1.7k Jan 01, 2023
Simple, fast, and parallelized symbolic regression in Python/Julia via regularized evolution and simulated annealing

Parallelized symbolic regression built on Julia, and interfaced by Python. Uses regularized evolution, simulated annealing, and gradient-free optimization.

Miles Cranmer 924 Jan 03, 2023
Auto updating website that tracks closed & open issues/PRs on scikit-learn/scikit-learn.

Repository Status for Scikit-learn Live webpage Auto updating website that tracks closed & open issues/PRs on scikit-learn/scikit-learn. Running local

Thomas J. Fan 6 Dec 27, 2022
Microsoft contributing libraries, tools, recipes, sample codes and workshop contents for machine learning & deep learning.

Microsoft contributing libraries, tools, recipes, sample codes and workshop contents for machine learning & deep learning.

Microsoft 366 Jan 03, 2023
Spark development environment for k8s

Local Spark Dev Env with Docker Development environment for k8s. Using the spark-operator image to ensure it will be the same environment. Start conta

Otacilio Filho 18 Jan 04, 2022
Automatically create Faiss knn indices with the most optimal similarity search parameters.

It selects the best indexing parameters to achieve the highest recalls given memory and query speed constraints.

Criteo 419 Jan 01, 2023
A unified framework for machine learning with time series

Welcome to sktime A unified framework for machine learning with time series We provide specialized time series algorithms and scikit-learn compatible

The Alan Turing Institute 6k Jan 06, 2023
A machine learning toolkit dedicated to time-series data

tslearn The machine learning toolkit for time series analysis in Python Section Description Installation Installing the dependencies and tslearn Getti

2.3k Jan 05, 2023
A modular active learning framework for Python

Modular Active Learning framework for Python3 Page contents Introduction Active learning from bird's-eye view modAL in action From zero to one in a fe

modAL 1.9k Dec 31, 2022
Distributed scikit-learn meta-estimators in PySpark

sk-dist: Distributed scikit-learn meta-estimators in PySpark What is it? sk-dist is a Python package for machine learning built on top of scikit-learn

Ibotta 282 Dec 09, 2022
A python library for Bayesian time series modeling

PyDLM Welcome to pydlm, a flexible time series modeling library for python. This library is based on the Bayesian dynamic linear model (Harrison and W

Sam 438 Dec 17, 2022
Automated machine learning: Review of the state-of-the-art and opportunities for healthcare

Automated machine learning: Review of the state-of-the-art and opportunities for healthcare

42 Dec 23, 2022
A GitHub action that suggests type annotations for Python using machine learning.

Typilus: Suggest Python Type Annotations A GitHub action that suggests type annotations for Python using machine learning. This action makes suggestio

40 Sep 18, 2022
Using Logistic Regression and classifiers of the dataset to produce an accurate recall, f-1 and precision score

Using Logistic Regression and classifiers of the dataset to produce an accurate recall, f-1 and precision score

Thines Kumar 1 Jan 31, 2022
A simple python program that draws a tree for incrementing values using the Collatz Conjecture.

Collatz Conjecture A simple python program that draws a tree for incrementing values using the Collatz Conjecture. Values which can be edited: Length

davidgasinski 1 Oct 28, 2021
cuML - RAPIDS Machine Learning Library

cuML - GPU Machine Learning Algorithms cuML is a suite of libraries that implement machine learning algorithms and mathematical primitives functions t

RAPIDS 3.1k Dec 28, 2022