Skip to content

Introduction

What is mlinfra?

mlinfra is a python package designed to streamline the deployment of various MLOps tools within a MLOps stack, quickly and with best practices. The core philosophy behind mlinfra is to simplify and expedite the deployment of MLOps infrastructure. This approach enables ML Engineers or ML Platform engineers to concentrate on delivering business value, by significantly reducing the time and resources typically required for deploying MLOps tools on the cloud.

Why mlinfra?

mlinfra was conceived from a personal challenge I faced while deploying various MLOps tools in cloud environments. The absence of a standardized method for deploying MLOps tools and infrastructure on the cloud was not only noticeable but also a source of frustration. As my discussions with industry peers expanded, it became evident that knowledge regarding the deployment of ML Infrastructure is fragmented across various tool-specific documentation sources. This fragmentation hinders rapid and best practice-compliant deployment of diverse MLOps tools within a stack, posing challenges for platform teams who aim to:

  • Experiment with different tools in MLOps stacks to customize solutions according to their specific needs.
  • Swiftly evaluate new tools without delving into extensive deployment documentation.
  • Deploy tools within the MLOps stack efficiently and in accordance with best practices, thereby circumventing the need for prolonged development cycles and complex planning.

The fundamental concept behind mlinfra is to establish a universal Infrastructure as Code (IaC) framework that expedites the creation and deployment of MLOps stacks. This package empowers MLOps Engineers and Platform Engineers to deploy a variety of MLOps tools across different stages of the machine learning lifecycle. The essence of mlinfra lies in its Python layer, which interprets the deployment configuration and utilizes Terraform modules to deploy these tools. It further enhances the process by dynamically generating a suite of inputs, roles, and permissions, thereby simplifying and streamlining the deployment process.

How does it work?

mlinfra deploys infrastructure using a declarative approach. The minimal spec for aws cloud as infra with custom applications deployed is as follows:

The following is just a sample configuration.

The following sample yaml serves as a reference for the configuration of a MLOps stack which can be deployed using mlinfra. Some of the stacks and their toolings are currently under active development and may not be available to use right away.

name: aws-mlops-stack
provider:
  name: aws
  account-id: xxxxxxxxx
deployment:
  type: kubernetes
stack:
  data_versioning:
    - lakefs # can also be pachyderm or neptune and so on
  experiment_tracker:
    - mlflow # can be weights and biases or determined, or neptune or clearml and so on...
  orchestrator:
    - zenml # can also be argo, or luigi, or aws-batch or airflow, or dagster, or prefect  or kubeflow or flyte
  artifact_tracker:
    - mlflow # can also be neptune or clearml or lakefs or pachyderm or determined or wandb and so on...
  model_registry:
    - bentoml # can also be  mlflow or neptune or determined and so on...
  model_serving:
    - nvidia_triton # can also be bentoml or fastapi or cog or ray or seldoncore or tf serving
  monitoring:
    - nannyML # can be grafana or alibi or evidently or neptune or mlflow or prometheus or weaveworks and so on...
  alerting:
    - mlflow # can be mlflow or neptune or determined or weaveworks or prometheus or grafana and so on...

The above configuration can be simply deployed using the following command:

mlinfra terraform --action=apply --stack-config-path=aws-mlops-stack.yaml

For more details, refer to the User Guide section.