What’s New in v0.5.7 | ZenML Blog

Last updated: November 3, 2022.

ZenML 0.5.7 is out now and we bring you not one but two brand new integrations to cover common use cases. ZenML now supports MLFlow for tracking pipelines as experiments and Evidently for detecting drift in your ML pipelines in production!

There are a bunch of other smaller changes, including preventing Kubeflow Pipelines from timing out during the local installation process and a bunch of improvements to the feedback that the ZenML CLI tool gives you. For a detailed look at what’s changed, give our full release notes a glance.

MLFlow Tracking

You voted, we integrated! This new addition of MLFlow Tracking as part of ZenML’s integrations means that you can track your training runs using MLFlow. This gives you a handy web UI that you can use to log and query your experiments. ZenML is tracking all of your parameters and metrics already, but you might prefer to visualize those changes in the MLFlow web UI.

Tracking machine learning training runs with MLFlow

We’ve written an example that showcases the integration that you can check out by using the zenml example pull mlflow CLI command. The README file offers full instructions for how to set this up manually, but if you just want to try out the UI and let ZenML handle the local setup for you, just type zenml example run mlflow.

Watch this space for more MLFlow goodness coming your way!

Evidently for Drift Detection

Evidently is an open-source tool for detecting drift among your data inputs. Machine learning pipelines are built on top of those data inputs, so it is worth checking for drift if you have a model that was trained on a certain distribution of data.

Detecting and visualizing data drift with Evidently

The ZenML integration with Evidently implements this functionality in the form of several standardized steps. You select which of the profile sections you want to use in your step by passing a string into the EvidentlyProfileConfig. Possible options supported by Evidently are:

For example, you could define a step to detect drift using our standard interface in the following way:

drift_detector = EvidentlyProfileStep(
    EvidentlyProfileConfig(
        column_mapping=None,
        profile_section="datadrift",
    )
)

Here you can see that defining the step is extremely simple using our class-based interface and then you just have to pass in the two dataframes when defining the pipeline for the comparison to take place.

@pipeline
def drift_detection_pipeline(data_loader, full_data, partial_data, drift_detector):
    data_loader = data_loader()
    full_data = full_data(data_loader)
    partial_data = partial_data(data_loader)
    drift_detector(reference_dataset=full_data, comparison_dataset=partial_data)

As with MLFlow, we’ve written an example for you to see how it all works in code. You can check it out by using the zenml example pull drift_detection CLI command. The README file offers full instructions for how to set this up manually, but if you just want to try out the UI and let ZenML handle the local setup for you, just type zenml example run drift_detection.

Contribute to ZenML!

Join our Slack to let us know what you think we should build next!

Keep your eyes open for future releases and make sure to vote on your favorite feature of our roadmap to make sure it gets implemented as soon as possible.

[Image credit: Photo by Yoksel 🌿 Zok on Unsplash]


More from us: