Saltar para o conteúdo principal

Deploy Your Own MLflow Workspace On-Premise with Docker

Deploy Your Own MLflow Workspace On-Premise with Docker

MLflow (Abre numa nova janela) is an open source platform to manage the lifecycle of ML models end to end. It tracks the code, data and results for each ML experiment, which means you have a history of all experiments at any time. A dream for every Data Scientist. Moreover, MLflow is library-agnostic, which means you can use all ML libraries like TensorFlow, PyTorch or scikit-learn. All MLflow functions are available via a REST API (Abre numa nova janela), CLI (Abre numa nova janela), Python API (Abre numa nova janela), R API (Abre numa nova janela) and Java API (Abre numa nova janela).

As a Data Scientist, you spend a lot of time optimizing ML models. The best models often depend on an optimal hyperparameter or feature selection, and it is challenging to find an optimal combination. Also, you have to remember all experiments, which is very time-consuming. MLflow is an efficient platform to address these challenges.

In this post, we briefly introduce the basics of MLflow and show how to set up an MLflow workspace on-premise. We set up the MLflow environment in a Docker stack so that we can run it on all systems. In this context, we have the services Postgres database, SFTP server, JupyterLab and MLflow tracking server UI. Let’s start.

To read this post you'll need to become a member. Members help us fund our work to ensure we can stick around long-term.

See our plans (Abre numa nova janela)

Tópico Data Science

0 comentários

Gostaria de ser o primeiro a escrever um comentário?
Torne-se membro de Tinz Twins Hub e comece a conversa.
Torne-se membro