In many cases you can survive without MLOps. In my opinion though, any ML-focused, successful company must employ the principles and approaches advocated by MLOps. In this article,Adrian demonstrates when you need MLOps and when you don’t.
TLDR: immediately: no; at some point: yes
Slack has developed a unified framework called the Recommend API, which allows the user to bootstrap new recommendation use cases behind an API, which is easily accessible to engineers at Slack.
Behind the scenes, these recommenders reuse a common set of infrastructure for every part of the recommendation engine, such as data processing, model training, candidate generation, and monitoring. This has allowed to deliver a number of different recommendation models across the product, driving improved customer experience in a variety of contexts.
In this article, William focuses on the data processing abilities in Snowflake through the use of dbt, from development to deployment and monitoring using the example of their Data Platform. The platform is structured into four main pillars: infrastructure, ingestions, processing and governance.
An ML Model loses its predictive performance when probability distribution of any *relevant* input feature or output or both changes (which means that system variables undergo change, i.e., IID assumption violation) -or- relationship between relevant input features and output changes (i.e., the underlying system itself changes).
Takeaways:
Netflix developed a new machine learning algorithm based on reinforcement learning to create an optimal list of recommendations considering a finite time budget for the user. In a recommendation use case, often the factor of finite time to make a decision is ignored; Netflix added this dimension to its recommendation system and in general in decision problems, in addition to the relevance of the recommendations.
Companies frequently deploy their models to virtual machines (Google Compute Engine or even on-prem machines). This is something that should be avoided. Google Cloud provides a dedicated service called Vertex AI Endpoints to deploy your models.
Vertex AI Endpoints provides great flexibility paired with easy usage. You can keep it simple or get full in and customize it to your needs using custom containers.
This article covers how to put your models into production and serve requests at a large scale. Also a few workarounds surrounding the limitations of the service.
Kedro is an open-source Python framework for creating reproducible, maintainable and modular data science code. By using Kedro we address the following challenges in ML projects:
It’s about plugins for Kedro, that allow running Kedro pipelines at scale on various serverless/managed cloud services from ALL of the major cloud providers - Google Cloud Platform, AWS and Azure as well as on existing Kubernetes-based infrastructures - Kubeflow Pipelines or Apache Airflow.
Since this edition of DATA Pill is focused on MLOps, this ebook could not be missing from the list
Jon talks with Erik Bernhardsson, who invented Spotify’s original music recommendation system. A few of the topics they covered:
Machine Learning models can add value and insight into many projects, but they can also be challenging to put into production due to problems such as a lack of reproducibility, difficulty in maintaining integrations and sneaky data quality issues. Kedro, a framework for creating reproducible, maintainable and modular data science code, and Great Expectations, a framework for data validations are two great open-source Python tools that can address some of these problems. Both integrate seamlessly with Airflow for flexible and powerful ML pipeline orchestration.
Last reminder. Topics: