Last updated: November 3, 2022.
Our 0.11.0 release contains our new annotation workflow and stack component. We’ve been blogging about this for a few weeks, and even started maintaining our own repository of open-source annotation tools. With ZenML 0.11.0 you can bring data labeling into your MLOps pipelines and workflows as a first-class citizen. We’ve started our first iteration of this functionality by integrating with Label Studio, a leader in the open-source annotation tool space.
This release also includes a ton of updates to our documentation. (Seriously, go check them out! We added tens of thousands of words since the last release.) We continued the work on our data validation story from the previous release: Deepchecks is the newest data validator we support, and we updated our Evidently and whylogs integrations to include all the latest and greatest from those tools.
Beyond this, as usual we included a number of smaller bugfixes and documentation changes to cumulatively improve experience of using ZenML as a user. For a detailed look at what’s changed, give our full release notes a glance.
We’ve been hard at work on our new stack component and integration with Label Studio, the first of our data annotation tools. 🥳
Annotators are a stack component that enables the use of data annotation as part of your ZenML stack and pipelines. You can use the associated CLI command to launch annotation, configure your datasets and get stats on how many labeled tasks you have ready for use.
Data annotation/labeling is a core part of MLOps that is frequently left out of the conversation. With this release, ZenML now supports annotation as an integrated and first-class citizen as part of the MLOps lifecycle.
Got feedback or just want to let us know how you feel? Connect with us or join us for a Community Meetup 👋 which happens every Wednesday!
We continued the work started in the previous release, adding a new integration with Deepchecks. Deepchecks is a feature-rich data validation open-source library to painlessly do data validation. Deepchecks can do a variety of data validation tasks, from data integrity checks that work with a single dataset to data+model evaluation to data drift analyses. All this can be done with minimal configuration input from the user, or customized with specialized conditions that the validation checks should perform. Check out our example if you want to see it in action!
We also updated our integrations with Great Expectations, Evidently and whylogs to support their latest releases. These tools have added lots of great new features in recent weeks and we now support most of them, all under the new data validator stack component and abstraction.
We made a significant overhaul of our documentation since the last release:
We fixed a bug that was preventing users who upgraded to 0.10.0 from pulling new examples. This now works without any problem.
The latest release includes several smaller features and updates to existing functionality:
@enable_mlflow
decorator) which will allow the creation of nested runs for each step of a ZenML pipeline in MLflow.requirements.txt
file if the file ended with a newline.The 0.11.0 release remodels the Evidently and whylogs integrations as Data Validator stack components, in an effort to converge all data profiling and validation libraries around the same abstraction. As a consequence, you now need to configure and add a Data Validator stack component to your stack if you wish to use Evidently or whylogs in your pipelines:
for Evidently:
zenml data-validator register evidently -f evidently
zenml stack update -dv evidently
for whylogs:
zenml data-validator register whylogs -f whylogs
zenml stack update -dv whylogs
In this release, we have also upgraded the Evidently and whylogs libraries to their latest and greatest versions (whylogs 1.0.6 and evidently 0.1.52). These versions introduce non-backwards compatible changes that are also reflected in the ZenML integrations:
Evidently profiles are now materialized using their original evidently.model_profile.Profile
data type and the builtin EvidentlyProfileStep
step now also returns a Profile
instance instead of the previous dictionary representation. This may impact your existing pipelines as you may have to update your steps to take in Profile
artifact instances instead of dictionaries.
the whylogs whylogs.DatasetProfile
data type was replaced by whylogs.core.DatasetProfileView
in the builtin whylogs materializer and steps. This may impact your existing pipelines as you may have to update your steps to return and take in whylogs.core.DatasetProfileView
artifact instances instead of whylogs.DatasetProfile
objects.
the whylogs library has gone through a major transformation that completely removed the session concept. As a result, the enable_whylogs
step decorator was replaced by an enable_whylabs
step decorator. You only need to use the step decorator if you wish to log your profiles to the Whylabs platform.
Please refer to the examples provided for Evidently and whylogs to learn more about how to use the new integration versions:
We received several new community contributions during this release cycle. Here’s everybody who contributed towards this release:
Join our Slack to let us know if you have an idea for a feature or something you’d like to contribute to the framework.
We have a new home for our roadmap where you can vote on your favorite upcoming feature or propose new ideas for what the core team should work on. You can vote without needing to log in, so please do let us know what you want us to build!
[Photo by Gaelle Marcel on Unsplash]