Concept | Metrics & checks (pre-12.6)#

Watch the video


This article applies to Dataiku versions 12.5 and below. For versions 12.6 and above, checks on datasets have been replaced by data quality rules.

In this lesson, we introduce the:

  • Challenges of automation

  • The nature of metrics and checks

  • Why they are crucial for automation

  • How to leverage them in Dataiku

Automation challenges#

The lifecycle of a data or machine learning project doesn’t end once a Flow is complete. To maintain our workflows and improve our models, we must continuously feed them new data. Automation allows us to do this more efficiently by reducing the amount of manual supervision.

However, as we automate workflows, we are exposed to some risks, such as ingesting poor quality data without knowing it, which could affect output datasets, models, and dashboards.

For instance, our workflow could become broken as an extra column is added to an input dataset. Or, our model could stop capturing the pattern in the data and become obsolete.

While automation promises to save time, it also creates the need to implement key metrics and checks, so that our project doesn’t break and remains relevant.

Slide depicting common automation challenges, such as bad quality data, broken workflows, and obsolete models.

Defining metrics#

Metrics are metadata used to take measurements on the following Flow items:

  • Datasets

  • Managed folders

  • Saved models

They allow us to monitor the evolution of a Dataiku item. For example, we could compute:

  • The number of missing values of a column

  • The size of a folder

  • The accuracy of a model

Metrics can also be set on partitioned objects and be computed on a per-partition basis.

Defining checks#

Metrics are often used in combination with checks to verify their evolution.

For instance, we could check that:

  • There are no missing values for a given column.

  • The size of a folder does not exceed 3GB.

  • The model accuracy does not fall below 0.8.

Slide depicting example metrics and checks in Dataiku.

Checks return one of the four following status updates after each run:

  • EMPTY if the metrics value hasn’t been computed.

  • ERROR if the check condition has not been respected.

  • WARNING if the check fails a soft condition but not a hard one.

  • OK if the check should not raise any concern.

Slide depicting an example of outcomes for a check called Positive Purchase Amount.

Metrics and checks can be set to be computed automatically after a rebuild, if desired.

Dataiku screenshot of the Edit Metrics page of a dataset showing how metrics can be auto-computed or not.

Depending on the outcome of the checks, some actions could be triggered, such as, for example:

  • Retraining a model.

  • Sending a report.

  • Refreshing a dashboard, etc.

Slide depicting how metrics are conditioned by checks which in turn trigger actions in scenarios.

This can be achieved through scenarios. The scope of the scenario will be presented in the next lesson.

Viewing metrics#

Now, let’s see in practice how we can leverage metrics and checks in Dataiku. We’ll work on a dataset containing credit card transaction data.

Metrics and checks can be accessed from the Status tab on a dataset, a managed folder, or a machine learning model.

When you enter the Status tab, you are presented with the default “Last Value” view of the metrics, which is a “tile view” displaying the latest value of each selected metric.

Dataiku screenshot of a sample metrics page of a dataset.

Clicking on the value of a metric will bring up a modal box with the history of this value.

We can also switch to other views of the metrics, such as the History view, which is a “ribbon” view displaying the history of all selected metrics.

For partitioned datasets, we will have more views to visualize metrics on a per-partition basis.

Since there can be a lot of available metrics on an item, we must select the metrics we want to add to the displayed screen tiles on the Metrics page.

Adding metrics#

Within the Status tab of a dataset, there is an “Edit” subtab. From here, the list of available metrics can be customized and extended using:

  • Built-in probes

  • Custom code

  • Plugins

For now, we will focus on built-in probes. We will explore building probes with custom code or via plugins in another lesson.

A probe is a component that can compute several metrics on an item. Each probe has a configuration that indicates what should be computed for this probe.

For instance, the Column statistics probe can compute the minimum and maximum value of a column, such as purchase_amount.

The Column statistics probe within the Edit Metrics page of a dataset.

In addition to column statistics, you can also find probes to compute the most frequent values of a specific column, the percentiles of a specific column, or statistics about data validity.

Data validity metrics are to be used in association with user-defined meanings which complement the description of the column and optionally provide a list or a pattern for the valid values.

You can also use a cell value probe to retrieve the values of one or more cells from the dataset as metrics, with historization. This allows you to filter the rows on specific conditions and display dataset values for selected columns.

More customized metrics can be defined using Python code or via a plugin. The use of these custom metrics will be detailed in a later lesson on custom metric, checks, and scenarios.

The engine computation settings used to compute the metrics can also be edited.

Adding checks#

Now that we have defined our metric, we can set checks to validate that our latest computed value does not raise any concerns.

The Checks page is similar to the Metrics page. We can choose between different views of the checks status, and select which checks we want to display.

Dataiku screenshot of the Checks page of a dataset.

Similar to metrics, checks can be built from the Edit tab, from built-in visual probes, custom code, or plugins.

In our case, we want to check that the minimum value of purchase_amount is positive, as this was our initial assumption. We do this with a numerical range check on the metric Min of purchase_amount.

Dataiku screenshot of the Edit Checks page of a dataset.

We can set a soft minimum to return a warning if not respected or a hard minimum to trigger an error.

As for a metric, we can set the option to automatically run the check after each rebuild, if desired.

We can now save our settings and test the check. We can see that it’s working, but it returns an error. We had incorrectly assumed that the purchase_amount value is always positive, while in fact the minimum of purchase_amount is negative.

This could mean that we have unconsciously ingested invalid data, or, in our case, simply that we have made an incorrect assumption about the nature of the data. Either way, the failed check brings attention to a potential inconsistency in our data.

Metrics & checks on other objects#

Although not covered in this lesson, metrics and checks can also be used on managed folders and saved models. The nature of metrics and checks in those use cases will differ from the one we’ve seen with datasets. However, the rationale remains similar.

Slide depicting that metrics and checks can be tracked on other Dataiku items like managed folders and saved models.

What’s next?#

To learn more about metrics and checks, including through hands-on exercises, please register for the free Academy course on this subject found in the Advanced Designer learning path.

You can also learn more about metrics and checks in the reference documentation.