Kubeflow is a cloud-native, open source machine learning operations (MLOps) platform designed for developing and deploying ML models on Kubernetes. Kubeflow helps data scientists and machine learning engineers run the entire ML lifecycle within one tool.
Charmed Kubeflow is Canonical’s official distribution of Kubeflow. The key benefits of Charmed Kubeflow include security maintenance of container images, enterprise support, and further tooling integration with Spark, Feast, MLFlow, and others. Learn more about the differences between Charmed Kubeflow and the upstream project.
This blog explains the environments Charmed Kubeflow can run in and how to deploy it. By the time you’ve finished reading, you’ll have an overview of what you need to consider before deploying Kubeflow on Azure. You’ll also learn how to approach deployment based on your specific use case, existing infrastructure, long-term strategy, and level of expertise.
Kubeflow is the upstream project. It’s fully open source, and became an incubating project in the Cloud Native Computing Foundation (CNCF) in October 2022.
Most organizations looking for an MLOps platform to experiment with begin using Kubeflow directly from the official website. However, introducing Kubeflow into production environments like this can create problems:
Charmed Kubeflow was created to give organizations a secure path to move from experimentation to production.
As we’ll discuss in this blog, one example of how organizations can benefit from using Charmed Kubeflow is by using it to run their ML workloads on Azure. Depending on the organization’s level of expertise, Charmed Kubeflow can be operated with enterprise support, firefighting, or managed services.
Like the upstream project, Charmed Kubeflow is fully open source, and can run on any CNCF-conformant Kubernetes. This flexibility means Charmed Kubeflow can run on major public clouds like Azure, AWS and Google, or on-premises. Our documentation describes some common deployment scenarios used by enterprises, such as deployment in an airgapped environment (a system that doesn’t have access to the public internet), or behind a proxy.
Users often choose to run Charmed Kubeflow on major public clouds – like Azure – because they provide the readily available computing power required. When developing ML models, ML engineers and data scientists need access to multiple General Processing Units (GPUs), which are easier to access when deploying on the public clouds. Public clouds enable them to quickly experiment and validate projects and to run their ML workloads in production.
However, whether running Charmed Kubeflow on Azure is the best decision for you depends on multiple considerations:
[Banner to join the preview]
There are lots of ways you can runCharmed Kubeflow on Azure, depending on some of the key considerations previously mentioned. For example:
Here’s the guide to deploy Elastic Stack on Ubuntu VPS, with secure access, HTTPS proxying,…
This guide walks through deploying Nagios Core on an Ubuntu VPS, from system prep to…
After a decade under Pablo Cantero's stewardship, Shoryuken has a new maintainer - me. I'm…
MAAS 3.7 has been officially released and it includes a bunch of cool new features.…
Update: This article originally concluded that Eisel-Lemire wasn't worth it for Ruby. I was wrong.…
Recently, the team at MinIO moved the open source project into maintenance mode and will…