AutoML Benchmark in Production
Comparison and Analysis of Different AutoML Systems in the Production Domain.
AutoML Systems are tools that propose to automate the machine learning (ML) pipeline:
integration, preparation, modeling and model deployment.
Although all AutoML systems aim to facilitate the usage of ML in production,
they may differ on how to accomplish this objective, approaching the ML pipeline in different levels.
The purpose of this benchmark is to, using currently available AutoML systems in the market,
evaluate how each system approaches the ML pipeline and help a user to choose which system to pick.
25 AutoML systems are presented in this benchmark, listed in chapter AutoML Systems.
For a more objective evaluation approach,
this research includes evaluations concerning different criteria of each system:
Requirements and Boundaries, Performance,
Data Integration, Data Preparation,
Modeling, Deployment and Results.
Ending the benchmark, chapter Practical Examples presents what systems are recommended to four exemplified users.
1. Introduction to AutoML in Production
In the pursuit of leveraging the potential of data, companies usually rely on Machine Learning (ML) technologies. Applications of ML range from detecting anomalies in the operation and non-specialized machines designed to learn by themselves, to the optimization of routing and demand forecasting. Its most common use cases are for knowledge extraction operations and as the core technology underlying the control of processes and products. Projects that rely on the application of ML are usually referred to as ML projects.
However, the development of an ML project is not an automated task - much of it still relies on the expertise of data scientists and knowledge of the manufacturing process. Seeking a solution to this problem, AutoML systems are currently being developed, enabling a wide range of users to benefit from the valuable potential of data and facilitating the use of ML in real-world problems.
In short, AutoML systems propose to automate the ML pipeline. Although this naming is useful to get a straightforward view of the main idea, it conceals several tasks, as depicted in Fig. 1.

2. Methodology of the AutoML Benchmark
This benchmark is divided into 3 steps: AutoML Systems, Evaluation and Usage.
-
AutoML Systems
This step lists AutoML Systems currently available in the market and specifies how deeply each one of them is evaluated in step Evaluation.
Evaluation
This step aims to evaluate the models generated by the systems listed in step AutoML Systems and compare them. Each of the benchmarked systems was analysed in 7 different criteria:
- Requirements and Boundaries
- Performance
- Data Integration
- Data Preparation
- Modeling
- Deployment
- Results
These criteria are spread out from chapters Characteristics Requirements and Boundaries to Results.
Usage
This step guides the user on how to use the information provided in step Evaluation to choose systems that fulfill the user's requirements.
The criteria can be divided into two groups: user-oriented and system-oriented criteria.
To visualize how personal information may affect the choice of system, take a look at Requirements and Boundaries, the user-oriented criteria. Here, the required user's programming skills, hardware, software, price budget and other metrics are presented.
To compare models, created by the systems, between themselves, the system-oriented criteria can be used. Spread out from chapters Characteristics Requirements and Boundaries to Results, information ranging from input datatypes to deployment methods is found and can be compared to the use case of interest to find a system that fits the best the user's needs.To demonstrate how the procedure of choosing a system would occur, four exemplified users are presented in chapter Practical Examples, and AutoML systems are recommended for each of them.
3. AutoML Systems
AutoML systems are evaluated based on publicly available data from the production domain. These are the systems included in this benchmark:
Some were included in the analysis, but were not evaluated in every criteria:
- Uber Ludwig 16, not tested regarding its Performance.
- TransmogrifAI 17, systems that require no programming knowledge or Python skills were tested. Since TransmogrifAI requisites expertise in the programming language Scala, its Performance was not tested.
- The Automatic Statistician 18, since no version is available for the public, its Performance was not tested.
The following ones were not included in the analysis:
- Auto-WEKA19, last updated in 2017.
- Darwin20, a demo must be requested for use and the free trial only allows the user to use the dataset provided by the system.
- DataRobot21, a demo must be requested for use.
- Devol22, is not being actively developed, and most information required for this research is not provided by the system.
- ExploreKit23, most information required for this research is not provided by the system.
- AutoML Zero 24, most information required for this research is not provided by the system.
- Auto-PyTorch 25, it's currently in a very early pre-alpha version and most information required for this research is not provided by the system.
4. Characteristics, Requirements and Boundaries
The chapter Characteristics presents general information about each system and Requirements and Boundaries focuses on demonstrating how personal information of the user affects the choice of system.
4.1 Characteristics
Each benchmarked system was tested in a specific version, as can be seen in the "Tested at" column of Table 1.
Even though this research is focused on AutoML systems, not every evaluated system covers the whole AutoML Pipeline. This distinction is presented in the "AutoML" column of Table 1, where systems marked with a Yes cover a large portion of the Pipeline and the ones which do not are defined in plain text.
System |
Tested at | AutoML |
---|---|---|
Hyperopt-sklearn | Commit #149- February 2020 | Yes |
Auto-sklearn | Version 0.6.0 | Yes |
TPOT | Version 0.10.2 | Yes |
H2O AutoML | Version 3.28.0.3 | Yes |
SAS | February 2020 | Yes |
MLBox | Version 0.8.1 | Yes |
Google AutoML | February 2020 | Yes |
Azure Machine Learning | February 2020 | Yes |
MLJar | Version 2016 - 2017 | Yes |
ATM | Version 2 Pre-Alpha | Yes |
Auto_ml | Version 2.9.10 | Yes |
Amazon SageMaker | Version 1.56.3 | Yes |
Uber Ludwig | Version 2.0 | Yes |
TransmogrifAI | Version 0.6.1 | Yes |
The Automatic Statistician | May 2020 | Automated model discovery. |
AutoKeras | Version 1.0.1 | Yes |
Feature Tools | Version 0.14.0 | Only automated feature engineering. |
tsfresh | Version 0.15.1 | Only time series features extraction tool. |
There may be more algorithms used by these systems which are not explicitly stated.
Feature Tools and tsfresh automate only the Data Preparation step of the AutoML Pipeline. Nevertheless, they were kept in the analysis since they propose to automate the most time-consuming step of it.
4.2 Requirements and Boundaries
The systems differ in user knowledge, hardware and software requirements and price, which may imply limitations for different users and use cases. Table 2 exhibits these limitations for each system.
System |
|||||||||
---|---|---|---|---|---|---|---|---|---|
Required Skills | Programming Language(s) | Payment Information | Barriers | Cloud Computing | OS | ||||
Programming | Data Science | Linux | MacOS | Windows | |||||
Hyperopt-sklearn | Basic | Python, R, Java | None | None | No | Yes | - | Yes | |
Auto-sklearn | Basic | Python | None | None | No | Yes | No | No | |
TPOT | Basic | Python | None | None | No | Yes | Yes | Yes | |
H2O AutoML | Basic | Basic | Python, R | None | None | Yes | Yes | Yes | Yes |
SAS | None | None | None | None | Yes | - | - | - | |
MLBox | Basic | Basic | Python | None | None | No | Yes | Yes | Yes |
Google AutoML | None | None | None | Register a credit card before use. 19$ per training hour. | None | Yes | - | - | - |
Azure Machine Learning | Basic | Python | Price depends on the chosen setup, which includes the virtual machine configuration and reservation period. | Free trial doesn't allow the user to train a model. | Yes | - | - | - | |
MLJar | None | None | Python, R | Charges in a credit-based approach, where training spends credits. Free trial provides the user with an initial amount of credits. | - | Yes | - | - | - |
ATM | Basic | Basic | Python | None | None | No | Yes | - | - |
Auto_ml | Basic | Basic | Python | None | None | No | - | - | Yes |
Amazon SageMaker | None | None | None | Register a credit card before use. 2 months free trial. | Testing phase is not automated, this can be done by sending inference requests to the deployed endpoint. | Yes | - | - | - |
Uber Ludwig | None | Basic | Python | None | None | No | Yes | Yes | - |
TransmogrifAI | Scala | None | Uncommon programming language. | No | - | - | - | ||
The Automatic Statistician | - | - | No version available for the public. | - | - | - | - | ||
AutoKeras | Basic | Basic | Python | None | None | No | Yes | - | - |
Feature Tools | Basic | Basic | Python | None | None | No | Yes | Yes | Yes |
tsfresh | Basic | Basic | Python | None | None | No | Yes | - | Yes |
There may be more algorithms used by these systems which are not explicitly stated.
Most of the systems are free to use through a
Python-based API and require little knowledge in the programming language.
When the user has no experience in programming,
cloud-based paid systems should be chosen, since these offer an interface easy to use.
A deep knowledge in Data Science is usually not required, but may impact on results depending on the use case.
5. Performance of the AutoML Systems
In order to compare performances between models created by the AutoML systems, they were tested on an ML use case from production, where the following data from a CNC mill was used: CNC Mill Tool Wear data set . This is a classification problem, where the objective here is to predict the success of a test. The dataset was pre-processed before being tested by each system, having in the end 7586 instances and 50 dimensions. The results are presented in Table 3.
A future version of this benchmark will further include the results generated with the SECOM data set.
An overview of other publicly available data sets for production can be found at Fraunhofer IPT Application Fields and Free-Access Data Records .
System |
||||||
---|---|---|---|---|---|---|
Classification | ||||||
Accuracy | Logarithm Loss | F1 Score | AUC | Confusion Matrix Provided | Runtime | |
Hyperopt-sklearn | 0.9883 | 0.4 | 0.9889 | - | No | 47 min |
Auto-sklearn | 0.9964 | 0.1229 | 0.9966 | - | No | 120 min |
TPOT | 0.9928 | -0.0287 | 0.9928 | 0.9993 | No | 68.93 min |
H2O AutoML | 0.999 | 0.01207 | - | 0.999 | No | 53.1 min |
SAS | 0.994 | - | 0.995 | - | No | |
MLBox | 0.9937 | - | - | - | No | |
Google AutoML | 0.996 | 0.014 | 0.9955 | 0.999 | Yes | 60 min |
Azure Machine Learning | 0.9917 | - | 0.9917 | 0.9996 | Yes | 50 min |
MLJar | - | - | - | 0.9997 | No | 77 min |
ATM | 0.9998 | 0.0045 | 0.9998 | 0.9998 | Yes | 115 min |
Auto_ml | 0.9914 | 0.2959 | 0.9918 | 0.9913 | Yes | |
AutoKeras | 0.9997 | 0.004 | 0.9964 | 0.9961 | No | 60.7 min |
There may be more algorithms used by these systems which are not explicitly stated.
With state of the art results, high Accuracy, high F1 score and low Loss,
all systems presented a more than acceptable model as a solution for this particular problem
and a distinction between better and worse system is hard to establish here.
Performance also depends on the runtime, since running a system for more time can output better results.
6. Functionalities of the AutoML systems
The systems presented here propose to automate the AutoML Pipeline, or parts of it. Having that in mind, this chapter of the benchmark will evaluate to which degree every step of the Pipeline is covered by each system.
6.1 Data Integration
Data Integration aims to integrate data residing in different sources. In Table 4, data types accepted by each system are displayed, as well as if these automate the data integration process.
System |
|||||||
---|---|---|---|---|---|---|---|
Automatic | Accepted Input Datatypes | ||||||
Numerical | Time Series | Images | Videos | Audio | Text | ||
Hyperopt-sklearn | No | Yes | - | - | - | - | - |
Auto-sklearn | No | Yes | Yes | - | - | - | Yes |
TPOT | No | Yes | No | Yes | - | - | Yes |
H2O AutoML | No | Yes | Yes | Yes | - | - | Yes |
SAS | No | Yes | Yes | - | No | No | Yes |
MLBox | No | Yes | - | - | No | No | Yes |
Google AutoML | No | Yes | Yes | Yes | - | - | Yes |
Azure Machine Learning | No | Yes | Yes | Yes | - | Yes | Yes |
MLJar | No | Yes | Yes | - | - | - | Yes |
ATM | No | Yes | Yes | Yes | No | No | No |
Auto_ml | No | Yes | Yes | No | No | No | No |
Amazon SageMaker | No | Yes | - | - | - | - | |
Uber Ludwig | No | Yes | Yes | Yes | - | Yes | Yes |
TransmogrifAI | No | Yes | - | Yes | - | - | - |
The Automatic Statistician | No | Yes | Yes | - | - | - | - |
AutoKeras | No | Yes | - | Yes | - | - | Yes |
Feature Tools | No | Yes | Yes | - | - | - | - |
tsfresh | No | No | Yes | No | No | No | No |
There may be more algorithms used by these systems which are not explicitly stated.
No systems proposes to automate the Data Integration step,
since adapting to the vast quantity of different kinds of data sources is not trivial.
Regarding the data type, some systems are more limited and others are more inclusive,
for example, Auto_ml and Google AutoML respectively.
Regarding a specific use case, when looking for what system to use,
some systems can already be filtered out based on Table 4.
For example, a user facing an audio classification problem and having audio files as input data,
can pick Uber Ludwig, but not ATM.
6.2 Data Preparation
After assembling the dataset in a way it can be utilized by AutoML Systems, it's time to prepare it for the Modeling phase. That means increasing the quality of the data (Data Preprocessing) and restructuring it in order to facilitate the extraction of knowledge by an ML algorithm (Feature Engineering). As can be seen in Table 5, each system was evaluated with respect to how deep this preparation step is explored.
System |
|||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Input Required | Data Preprocessing | Feature Engineering | |||||||||
Handling Missing Values | Outliers Detection and Deletion | Data Standardization | Data Encoding | Feature Scaling | Data Reduction | Data Balancing | Data Augmentation | Feature Generation | Integrated with Modeling | ||
Hyperopt-sklearn | Train and test datasets. | No | No | Yes | Yes | Yes | Yes | Yes | - | - | Yes |
Auto-sklearn | Train and test datasets. | Yes | - | Yes | Yes | Yes | Yes | - | - | - | Yes |
TPOT | Yes | No | No | Yes | Yes | Yes | No | - | - | Yes | |
H2O AutoML | Yes | No | Yes | Yes | - | - | Yes | - | - | No | |
SAS | Whole dataset. | Yes | No | No | Yes | - | Yes | No | Yes | - | No |
MLBox | Train and test datasets. Testset must not contain a target. | Yes | No | No | Yes | - | No | - | - | - | No |
Google AutoML | Whole dataset. | Yes | - | - | Yes | - | - | - | - | - | No |
Azure Machine Learning | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | ||
MLJar | Train and test datasets. | Yes | - | Yes | Yes | Yes | - | - | - | Yes | No |
ATM | Train and test datasets. | - | - | - | Yes | - | - | - | - | - | Yes |
Auto_ml | Train and test datasets. | - | - | - | Yes | Yes | - | - | - | Yes | Yes |
Amazon SageMaker | Train and test datasets. | Yes | Yes | Yes | Yes | Yes | - | - | Yes | Yes | Yes |
Uber Ludwig | Train and test datasets. | Yes | - | Yes | Yes | Yes | - | - | No | - | Yes |
TransmogrifAI | Whole dataset. | Yes | - | Yes | Yes | - | Yes | Yes | - | Yes | Yes |
The Automatic Statistician | - | Yes | - | - | Yes | - | Yes | Yes | - | - | - |
AutoKeras | Train and test datasets. | - | - | Yes | Yes | - | - | Yes | - | No | |
Feature Tools | Datasets and relations between them. | - | - | - | Yes | - | - | - | - | Yes | No |
tsfresh | Time series dataset and target dataset. | - | - | - | - | - | - | - | - | Yes | No |
There may be more algorithms used by these systems which are not explicitly stated.
The exact approach taken for each of these methods is not specified in this benchmark, e.g. System X uses PCA for Data Reduction, therefore a Yes means some approach is used for the method and a No means no approach is taken. Nevertheless, the concept might be helpful when deciding which system to use for a certain use case. H2O AutoML might grant better results when dealing with unbalanced data sets, whereas other systems may be more effective for data sets with many dimensions, for example.
6.3 Modeling
The Modeling phase requires a prepared dataset to output effective results. To that end, the systems might
iterate through Data Preparation and Modeling
multiple times before reaching the final result.
Some of the algorithms used for each system, as well as how they are selected,
are shown in Table 6.
System |
|||||||||
---|---|---|---|---|---|---|---|---|---|
Training | ML Algorithms Included | Anti-overfitting approach | Algorithm Selection | Hyperparameter Selection | Diagnosis | ||||
Evaluation Approach | Search | GUI | Metrics | Charts | |||||
Hyperopt-sklearn | Yes | Decision Tree, Gaussian Processes, Annealing, Tree of Parzen Estimators, Light GBM, Random Search. | - | Full | Model | No | Yes | Yes | No |
Auto-sklearn | Yes | AdaBoost, Decision Tree, Gradient Boosting, K-NN, Random Forest, SVM. | Cross-validation | Full | Model | No | Yes | Yes | No |
TPOT | Yes | AdaBoost, Decision Tree, K-NN, Logistic Regression, Random Forest, SVM, XGBoost. | Cross-validation | Full | Population | No | Yes | Yes | No |
H2O AutoML | Yes | Neural Network, Gradient Boosting, Random Forest, XGBoost. | Cross-validation | Random search | No | Yes | Yes | No | |
SAS | Yes | Neural Network, Gradient Boosting, Regression, Random Forest, SVM. | Cross-validation | Full | Model | No | Yes | Yes | Yes |
MLBox | Yes | Random Forest, XGBoost (GB), K-nearest Neighbors, Logistic Regression, Random Forest, Light GBM, Regularized Greedy Forest, Extra Trees. | - | Full | Model | No | Yes | Yes | No |
Google AutoML | Yes | AdaBoost, Gradient Boosting. | - | - | - | No | Yes | Yes | Yes |
Azure Machine Learning | Yes | Decision Tree, Logistic Regression, SVM. | - | Full | Model | Yes | Yes | Yes | Yes |
MLJar | Yes | Deep Neural Network, Gradient Boosting, K-nearest Neighbors, Logistic Regression, Random Forest, Light GBM, Regularized Greedy Forest, Extra Trees. | Cross-validation | Yes | Yes | Yes | Yes | ||
ATM | Yes | - | Cross-validation | No | Yes | Yes | No | ||
Auto_ml | Yes | - | - | No | Yes | Yes | No | ||
Amazon SageMaker | Yes | Cross-validation | No | Yes | Yes | No | |||
Uber Ludwig | Yes | - | Cross-validation | No | Yes | Yes | Yes | ||
TransmogrifAI | Yes | Logistic Regression, Random Forest. | Cross-validation | No | Yes | Yes | No | ||
The Automatic Statistician | Yes | - | - | - | Yes | - | - | ||
AutoKeras | Yes | - | - | - | - | No | Yes | Yes | No |
Feature Tools | No | None | None | None | None | No | No | None | None |
tsfresh | No | None | None | None | None | No | No | No | No |
There may be more algorithms used by these systems which are not explicitly stated.
Have in mind that some systems allow exporting specific algorithms, others just output the best model and the results. See Table 8 for more information. More visual results can be obtained by a system which provides graphs and other metrics. See column Diagnosis of table Table 6.
6.4 Deployment
Ending the AutoML Pipeline, deployment aims to make the model available to users. For that, Table 7 answers the following questions:
- How can the model be deployed?
- How can the model be accessed by the user?
- Is training during production possible?
System |
|||
---|---|---|---|
Productionize Model | Web Service | Online Model Training | |
Hyperopt-sklearn | - | No | - |
Auto-sklearn | Export model as a .pickle file. | No | - |
TPOT | - | No | - |
H2O AutoML | Export model as POJO or MOJO format. | No | - |
SAS | Yes | - | |
MLBox | - | No | - |
Google AutoML | Classification models are stored in the platform and can be loaded from there. | Yes | Yes, batch prediction. |
Azure Machine Learning | Export model as a .pickle file. | Yes | - |
MLJar | Export model as .model files. | Yes | - |
ATM | - | Yes | - |
Auto_ml | Export model as a .dill file. | No | - |
Amazon SageMaker | Yes | - | |
Uber Ludwig | Model is saved as a group of files, which can be loaded using a ludwig method. | No | - |
TransmogrifAI | - | No | - |
AutoKeras | Export model as a tf.keras.Model instance. | No | - |
Feature Tools | Export features as .csv file. | No | No |
tsfresh | - | No | No |
There may be more algorithms used by these systems which are not explicitly stated.
There is not a standard way of exporting the generated model,
as can be seen in the Productionize Model column of Table 7.
Regarding the Web Service, usually a REST API is provided,
where the model can be accessed through an endpoint.
7: Results
Table 8
aims to show how the trained models results of each system are presented to the user.
Some systems also provide intermediary results - in case Modeling takes a long time,
it can be interesting to have access to a model that has already fully trained, even if that is not the best one.
System |
||
---|---|---|
Preliminary Results | Final Results | |
Hyperopt-sklearn | Yes | Best model + evaluation |
Auto-sklearn | No | Ensamble models + evaluation |
TPOT | Yes | Best model + evaluation |
H2O AutoML | Yes | Best model + evaluation + model characteristics + leaderboard of algorithms tested |
SAS | No | Best model + evaluation + model characteristics + leaderboard of algorithms tested |
MLBox | Yes | Best model + evaluation + model characteristics |
Google AutoML | Yes | Best model + evaluation + model characteristics |
Azure Machine Learning | Yes | Best model + evaluation + model characteristics |
MLJar | Yes | Ensamble models + evaluation |
ATM | No | Evaluation + model characteristics |
Auto_ml | No | Evaluation + model characteristics |
Amazon SageMaker | Yes | Ensemble models + model characteristics |
Uber Ludwig | No | Best model + evaluation + model characteristics |
TransmogrifAI | No | Best model + evaluation + model characteristics |
The Automatic Statistician | - | Reports written in natural language |
AutoKeras | Yes | Best model + evaluation |
Feature Tools | No | Feature matrix |
tsfresh | No | Feature matrix |
There may be more algorithms used by these systems which are not explicitly stated.
As can be seen in the table, the results vary from outputting sometimes a list of models and their characteristics,
and sometimes just the best model and evaluations.
Consequently, the user is not always allowed to decide which model can be exported, and therefore run in production.
Looking at the non-AutoML System, since Feature Tools and tsfresh propose to automate only the
Data Preparation phase,
their outputs are structured as a new dataset, containing new features and transformed old ones.
9: Practical Examples
In order to simulate how the process of finding systems
that fulfill a person's restrictions and requirements occurs,
the Personas Cards were created.
Each card represents a persona - an exemplified user.
At the end of each card, the chosen systems are displayed,
followed by the reasons they were designated for the specific persona and use case.

Sofia
Programming Experience
Basic
Programming Languages
R
Data Science Experience
Basic
OS
Windows
Hardware Available
Basic PC
(8GB of RAM and no GPU)
Available Time
2 days
Cost Budget
0
Use Case

Compare a model manually
built with one automatically
built.
Recommendations
System
H2O.
Reason
With no cost Budget,
a person with some experience in programming and data science
could use this system, since it is the only one that supports
the R language and can be run in a simple PC.

Johannes
Programming Experience
None
Programming Languages
None
Data Science Experience
None
OS
Windows
Hardware Available
Basic PC
(8GB of RAM and no GPU)
Available Time
3h
Cost Budget
1000€
Use Case

Predict rejection
rate of drills fabricated on a
production line.
Recommendations
Systems
SAS, Google, Azure, MLJar.
Reason
Having a budget, but a short time available,
a person with no experience in the area can use these systems,
to get a quick and cloud-based solution.

Klaus
Programming Experience
Full
Programming Languages
Python, R
Data Science Experience
Full
OS:
Linux
Hardware Available
Powerful PC
(32GB of RAM and GPU)
Available Time
2 days
Cost Budget
0
Use Case

Detect scratches
on a surface with
image recognition.
Recommendations
Systems
H2O, ATM, Uber Ludwig, Autokeras.
Reason
With no cost Budget, an experienced programmer and
data scientist could use these systems for image recognition
in the time available.

Hannah
Programming Experience
Full
Programming Languages
Python, R
Data Science Experience
Full
OS:
Windows
Hardware Available
Basic PC
(8GB of RAM and no GPU)
Available Time
1 days
Cost Budget
0
Use Case

Clean and improve data
collected from a CNC
Laser Cutting Machine.
Recommendations
Systems
Feature tools, tsfresh
Reason
Aiming at getting features for the given dataset,
systems that automate only this part of the pipeline are
recommended.
Paid systems, such as Google AutoML and Azure,
can be used by the personas with no budget since these systems provide free trials,
but it would not be a long term solution.
Recommended systems are not ordered from most recommended to less recommended.
To test the systems, in case of security issues or non-availability of data, it is possible to use publicly available data sets from production. See Fraunhofer IPT Application Fields and Free-Access Data Records .
10. Wrap Up
The past developments in the area of AutoML indicate that progress towards improved automation of specific steps within the AutoML Pipeline can be expected. Overall, a full automation of the whole AutoML Pipeline from Data Integration to Deployment is a concept that requires more research. In the near future, it can be expected that tasks such as Modeling will be automated enough so that ML models will be created with little to no ML knowledge. Semi-AutoML systems that support data scientists in other activities can be expected in the future as well.
11. Bibliography
-
Komer B, Bergstra J, Eliasmith C (2019) Hyperopt-Sklearn. in Hutter F, Kotthoff L, Vanschoren J, (Eds.). Automated Machine Learning. Springer International Publishing. Cham, pp. 97–111. ↩
-
Feurer M, Klein A, Eggensperger K, Springenberg JT, Blum M, Hutter F (2019) Auto-sklearn: Efficient and Robust Automated Machine Learning. in Hutter F, Kotthoff L, Vanschoren J, (Eds.). Automated Machine Learning. Springer International Publishing. Cham, pp. 113–134. ↩
-
Olson RS, Moore JH (2019) TPOT: A Tree-Based Pipeline Optimization Tool for Automating Machine Learning. in Hutter F, Kotthoff L, Vanschoren J, (Eds.). Automated Machine Learning. Springer International Publishing. Cham, pp. 151–160. ↩
-
H2O.ai. AutoML: Automatic Machine Learning — H2O 3.26.0.10 documentation. http://docs.h2o.ai/h2o/latest-stable/h2o-docs/automl.html. Accessed on 20.11.2019. ↩
-
SAS Institute Inc. SAS Visual Data Mining and Machine Learning. https://www.sas.com/en_us/software/visual-data-mining-machine-learning.html. Accessed on 26.01.2020. ↩
-
ARONIO DE ROMBLAY A. MLBox Documentation. https://mlbox.readthedocs.io/en/latest/index.html. Accessed on 20.11.2019. ↩
-
Google Cloud. Best practices for creating training data | AutoML Tables Documentation | Google Cloud. https://cloud.google.com/automl-tables/docs/data-best-practices#tables-does. Accessed on 20.11.2019. ↩
-
Microsoft Azure. Azure Machine Learning documentation. https://docs.microsoft.com/en-us/azure/machine-learning/. Accessed on 20.11.2019. ↩
-
MLJAR. mljar-docs. https://docs.mljar.com/. Accessed on 06.02.2020. ↩
-
Swearingen T, Drevo W, Cyphers B, Cuesta-Infante A, Ross A, Veeramachaneni K (2017 - 2017) ATM: A distributed, collaborative, scalable system for automated machine learning. 2017 IEEE International Conference on Big Data (Big Data). IEEE, pp. 151–162. ↩
-
Parry P. auto_ml 0.1.0 documentation. https://auto-ml.readthedocs.io/en/latest/index.html. Accessed on 20.11.2019. ↩
-
AWS. Amazon SageMaker - Developer Guide. https://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-dg.pdf . Accessed on 20.11.2019. ↩
-
Feurer M, Klein A, Eggensperger K, Springenberg JT, Blum M, Hutter F (2019) Auto-sklearn: Efficient and Robust Automated Machine Learning. in Hutter F, Kotthoff L, Vanschoren J, (Eds.). Automated Machine Learning. Springer International Publishing. Cham, pp. 113–134. ↩
-
Feature Labs. Featuretools 0.12.0 documentation. https://docs.featuretools.com/en/stable/index.html. Accessed on 20.11.2019. ↩
-
Christ M, Braun N, Neuffer J. tsfresh — tsfresh 0.12.0 documentation. https://tsfresh.readthedocs.io/en/latest/index.html. Accessed on 06.02.2020. ↩
-
Uber Ludwig documentation. https://uber.github.io/ludwig/. Accessed on 27.05.2020. ↩
-
Salesforce.com, Inc. AutoML library for building modular, reusable, strongly typed machine learning workflows on Spark from Salesforce Engineering. https://transmogrif.ai/. Accessed on 26.01.2020. ↩
-
The Automatic Statistician. https://link.springer.com/chapter/10.1007/978-3-030-05318-5_9. Accessed on 20.05.2020. ↩
-
Kotthoff L, Thornton C, Hoos HH, Hutter F, Leyton-Brown K (2019) Auto-WEKA: Automatic Model Selection and Hyperparameter Optimization in WEKA. in Hutter F, Kotthoff L, Vanschoren J, (Eds.). Automated Machine Learning. Springer International Publishing. Cham, pp. 81–95. ↩
-
SparkCognition (2019) From Data to Application: DARWINS UNIQUE APPROACH TO AUTOML. ↩
-
DataRobot. https://www.datarobot.com/. Accessed on 12.05.2020. ↩
-
Devol. https://github.com/joeddav/devol. Accessed on 20.05.2020. ↩
-
ExploreKit. http://people.eecs.berkeley.edu/~dawnsong/papers/icdm-2016.pdf. Accessed on 20.05.2020. ↩
-
AutoML Zero. https://arxiv.org/pdf/2003.03384.pdf. Accessed on 12.05.2020.↩
-
Hector Mendoza, Aaron Klein, Matthias Feurer, Jost Tobias Springenberg, Matthias Urban, Michael Burkart, Max Dippel, Marius Lindauer, Frank Hutter (2018) Towards Automatically-Tuned Deep Neural Networks: 7. in Hutter F, Kotthoff L, Vanschoren J, (Eds.). AutoML: Methods, Sytems, Challenges. Springer, pp. 141–156. ↩